var/home/core/zuul-output/0000755000175000017500000000000015116475663014543 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015116506770015501 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000005254540315116506762017716 0ustar rootrootDec 11 08:11:06 crc systemd[1]: Starting Kubernetes Kubelet... Dec 11 08:11:06 crc restorecon[4685]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:06 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 11 08:11:07 crc restorecon[4685]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 11 08:11:07 crc restorecon[4685]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Dec 11 08:11:07 crc kubenswrapper[4860]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 11 08:11:07 crc kubenswrapper[4860]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Dec 11 08:11:07 crc kubenswrapper[4860]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 11 08:11:07 crc kubenswrapper[4860]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 11 08:11:07 crc kubenswrapper[4860]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Dec 11 08:11:07 crc kubenswrapper[4860]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.398909 4860 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.401782 4860 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.401800 4860 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.401805 4860 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.401810 4860 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.401814 4860 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.401818 4860 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.401822 4860 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.401826 4860 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.401830 4860 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.401835 4860 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.401840 4860 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.401845 4860 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.401851 4860 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.401856 4860 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.401860 4860 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.401864 4860 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.401868 4860 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.401873 4860 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.401877 4860 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.401882 4860 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.401886 4860 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.401890 4860 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.401893 4860 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.401897 4860 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.401901 4860 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.401906 4860 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.401910 4860 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.401913 4860 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.401917 4860 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.401926 4860 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.401931 4860 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.401935 4860 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.401939 4860 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.401944 4860 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.401950 4860 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.401955 4860 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.401959 4860 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.401963 4860 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.401968 4860 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.401973 4860 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.401978 4860 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.401982 4860 feature_gate.go:330] unrecognized feature gate: Example Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.401986 4860 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.401990 4860 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.401995 4860 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.401999 4860 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.402003 4860 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.402007 4860 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.402010 4860 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.402014 4860 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.402017 4860 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.402021 4860 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.402024 4860 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.402028 4860 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.402032 4860 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.402037 4860 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.402040 4860 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.402046 4860 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.402051 4860 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.402054 4860 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.402059 4860 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.402063 4860 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.402066 4860 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.402070 4860 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.402074 4860 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.402078 4860 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.402081 4860 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.402085 4860 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.402088 4860 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.402092 4860 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.402098 4860 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402369 4860 flags.go:64] FLAG: --address="0.0.0.0" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402378 4860 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402387 4860 flags.go:64] FLAG: --anonymous-auth="true" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402393 4860 flags.go:64] FLAG: --application-metrics-count-limit="100" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402398 4860 flags.go:64] FLAG: --authentication-token-webhook="false" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402403 4860 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402409 4860 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402414 4860 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402419 4860 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402423 4860 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402427 4860 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402432 4860 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402436 4860 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402440 4860 flags.go:64] FLAG: --cgroup-root="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402443 4860 flags.go:64] FLAG: --cgroups-per-qos="true" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402448 4860 flags.go:64] FLAG: --client-ca-file="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402452 4860 flags.go:64] FLAG: --cloud-config="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402455 4860 flags.go:64] FLAG: --cloud-provider="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402460 4860 flags.go:64] FLAG: --cluster-dns="[]" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402464 4860 flags.go:64] FLAG: --cluster-domain="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402468 4860 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402472 4860 flags.go:64] FLAG: --config-dir="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402476 4860 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402480 4860 flags.go:64] FLAG: --container-log-max-files="5" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402485 4860 flags.go:64] FLAG: --container-log-max-size="10Mi" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402490 4860 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402494 4860 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402499 4860 flags.go:64] FLAG: --containerd-namespace="k8s.io" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402503 4860 flags.go:64] FLAG: --contention-profiling="false" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402508 4860 flags.go:64] FLAG: --cpu-cfs-quota="true" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402513 4860 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402517 4860 flags.go:64] FLAG: --cpu-manager-policy="none" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402522 4860 flags.go:64] FLAG: --cpu-manager-policy-options="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402527 4860 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402531 4860 flags.go:64] FLAG: --enable-controller-attach-detach="true" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402536 4860 flags.go:64] FLAG: --enable-debugging-handlers="true" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402540 4860 flags.go:64] FLAG: --enable-load-reader="false" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402545 4860 flags.go:64] FLAG: --enable-server="true" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402549 4860 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402554 4860 flags.go:64] FLAG: --event-burst="100" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402558 4860 flags.go:64] FLAG: --event-qps="50" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402562 4860 flags.go:64] FLAG: --event-storage-age-limit="default=0" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402566 4860 flags.go:64] FLAG: --event-storage-event-limit="default=0" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402570 4860 flags.go:64] FLAG: --eviction-hard="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402575 4860 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402579 4860 flags.go:64] FLAG: --eviction-minimum-reclaim="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402583 4860 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402587 4860 flags.go:64] FLAG: --eviction-soft="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402591 4860 flags.go:64] FLAG: --eviction-soft-grace-period="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402595 4860 flags.go:64] FLAG: --exit-on-lock-contention="false" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402599 4860 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402603 4860 flags.go:64] FLAG: --experimental-mounter-path="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402607 4860 flags.go:64] FLAG: --fail-cgroupv1="false" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402611 4860 flags.go:64] FLAG: --fail-swap-on="true" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402615 4860 flags.go:64] FLAG: --feature-gates="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402620 4860 flags.go:64] FLAG: --file-check-frequency="20s" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402624 4860 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402628 4860 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402632 4860 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402649 4860 flags.go:64] FLAG: --healthz-port="10248" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402653 4860 flags.go:64] FLAG: --help="false" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402657 4860 flags.go:64] FLAG: --hostname-override="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402661 4860 flags.go:64] FLAG: --housekeeping-interval="10s" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402666 4860 flags.go:64] FLAG: --http-check-frequency="20s" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402670 4860 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402674 4860 flags.go:64] FLAG: --image-credential-provider-config="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402678 4860 flags.go:64] FLAG: --image-gc-high-threshold="85" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402682 4860 flags.go:64] FLAG: --image-gc-low-threshold="80" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402687 4860 flags.go:64] FLAG: --image-service-endpoint="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402691 4860 flags.go:64] FLAG: --kernel-memcg-notification="false" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402695 4860 flags.go:64] FLAG: --kube-api-burst="100" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402700 4860 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402704 4860 flags.go:64] FLAG: --kube-api-qps="50" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402708 4860 flags.go:64] FLAG: --kube-reserved="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402712 4860 flags.go:64] FLAG: --kube-reserved-cgroup="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402716 4860 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402721 4860 flags.go:64] FLAG: --kubelet-cgroups="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402725 4860 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402729 4860 flags.go:64] FLAG: --lock-file="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402733 4860 flags.go:64] FLAG: --log-cadvisor-usage="false" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402737 4860 flags.go:64] FLAG: --log-flush-frequency="5s" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402741 4860 flags.go:64] FLAG: --log-json-info-buffer-size="0" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402747 4860 flags.go:64] FLAG: --log-json-split-stream="false" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402751 4860 flags.go:64] FLAG: --log-text-info-buffer-size="0" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402755 4860 flags.go:64] FLAG: --log-text-split-stream="false" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402759 4860 flags.go:64] FLAG: --logging-format="text" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402763 4860 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402767 4860 flags.go:64] FLAG: --make-iptables-util-chains="true" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402771 4860 flags.go:64] FLAG: --manifest-url="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402775 4860 flags.go:64] FLAG: --manifest-url-header="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402780 4860 flags.go:64] FLAG: --max-housekeeping-interval="15s" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402784 4860 flags.go:64] FLAG: --max-open-files="1000000" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402789 4860 flags.go:64] FLAG: --max-pods="110" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402794 4860 flags.go:64] FLAG: --maximum-dead-containers="-1" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402798 4860 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402802 4860 flags.go:64] FLAG: --memory-manager-policy="None" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402806 4860 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402810 4860 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402814 4860 flags.go:64] FLAG: --node-ip="192.168.126.11" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402818 4860 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402827 4860 flags.go:64] FLAG: --node-status-max-images="50" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402831 4860 flags.go:64] FLAG: --node-status-update-frequency="10s" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402835 4860 flags.go:64] FLAG: --oom-score-adj="-999" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402839 4860 flags.go:64] FLAG: --pod-cidr="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402844 4860 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402850 4860 flags.go:64] FLAG: --pod-manifest-path="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402854 4860 flags.go:64] FLAG: --pod-max-pids="-1" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402858 4860 flags.go:64] FLAG: --pods-per-core="0" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402863 4860 flags.go:64] FLAG: --port="10250" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402867 4860 flags.go:64] FLAG: --protect-kernel-defaults="false" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402871 4860 flags.go:64] FLAG: --provider-id="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402875 4860 flags.go:64] FLAG: --qos-reserved="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402879 4860 flags.go:64] FLAG: --read-only-port="10255" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402884 4860 flags.go:64] FLAG: --register-node="true" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402888 4860 flags.go:64] FLAG: --register-schedulable="true" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402892 4860 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402899 4860 flags.go:64] FLAG: --registry-burst="10" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402903 4860 flags.go:64] FLAG: --registry-qps="5" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402907 4860 flags.go:64] FLAG: --reserved-cpus="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402911 4860 flags.go:64] FLAG: --reserved-memory="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402916 4860 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402920 4860 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402924 4860 flags.go:64] FLAG: --rotate-certificates="false" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402928 4860 flags.go:64] FLAG: --rotate-server-certificates="false" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402932 4860 flags.go:64] FLAG: --runonce="false" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402936 4860 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402940 4860 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402945 4860 flags.go:64] FLAG: --seccomp-default="false" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402949 4860 flags.go:64] FLAG: --serialize-image-pulls="true" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402953 4860 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402957 4860 flags.go:64] FLAG: --storage-driver-db="cadvisor" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402961 4860 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402965 4860 flags.go:64] FLAG: --storage-driver-password="root" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402969 4860 flags.go:64] FLAG: --storage-driver-secure="false" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402979 4860 flags.go:64] FLAG: --storage-driver-table="stats" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402983 4860 flags.go:64] FLAG: --storage-driver-user="root" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402986 4860 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402991 4860 flags.go:64] FLAG: --sync-frequency="1m0s" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402995 4860 flags.go:64] FLAG: --system-cgroups="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.402999 4860 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.403006 4860 flags.go:64] FLAG: --system-reserved-cgroup="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.403010 4860 flags.go:64] FLAG: --tls-cert-file="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.403013 4860 flags.go:64] FLAG: --tls-cipher-suites="[]" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.403019 4860 flags.go:64] FLAG: --tls-min-version="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.403022 4860 flags.go:64] FLAG: --tls-private-key-file="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.403026 4860 flags.go:64] FLAG: --topology-manager-policy="none" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.403030 4860 flags.go:64] FLAG: --topology-manager-policy-options="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.403034 4860 flags.go:64] FLAG: --topology-manager-scope="container" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.403038 4860 flags.go:64] FLAG: --v="2" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.403043 4860 flags.go:64] FLAG: --version="false" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.403049 4860 flags.go:64] FLAG: --vmodule="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.403053 4860 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.403058 4860 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.403151 4860 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.403156 4860 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.403160 4860 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.403164 4860 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.403168 4860 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.403172 4860 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.403176 4860 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.403180 4860 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.403184 4860 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.403187 4860 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.403191 4860 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.403195 4860 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.403198 4860 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.403203 4860 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.403207 4860 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.403210 4860 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.403215 4860 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.403219 4860 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.403224 4860 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.403227 4860 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.403231 4860 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.403235 4860 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.403239 4860 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.403243 4860 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.403246 4860 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.403250 4860 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.403254 4860 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.403257 4860 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.403262 4860 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.403265 4860 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.403269 4860 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.403272 4860 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.403275 4860 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.403279 4860 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.403282 4860 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.403286 4860 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.403289 4860 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.403292 4860 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.403296 4860 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.403299 4860 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.403304 4860 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.403308 4860 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.403312 4860 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.403315 4860 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.403319 4860 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.403324 4860 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.403328 4860 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.403331 4860 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.403334 4860 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.403338 4860 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.403342 4860 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.403345 4860 feature_gate.go:330] unrecognized feature gate: Example Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.403348 4860 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.403352 4860 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.403358 4860 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.403362 4860 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.403365 4860 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.403368 4860 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.403372 4860 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.403375 4860 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.403380 4860 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.403383 4860 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.403387 4860 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.403392 4860 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.403396 4860 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.403401 4860 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.403404 4860 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.403409 4860 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.403413 4860 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.403417 4860 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.403421 4860 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.403427 4860 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.415500 4860 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.415554 4860 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.415758 4860 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.415777 4860 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.415790 4860 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.415801 4860 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.415811 4860 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.415821 4860 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.415835 4860 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.415848 4860 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.415858 4860 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.415869 4860 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.415880 4860 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.415890 4860 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.415903 4860 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.415914 4860 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.415924 4860 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.415934 4860 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.415943 4860 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.415953 4860 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.415963 4860 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.415972 4860 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.415982 4860 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.415992 4860 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.416002 4860 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.416011 4860 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.416028 4860 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.416045 4860 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.416057 4860 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.416067 4860 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.416078 4860 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.416089 4860 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.416102 4860 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.416115 4860 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.416126 4860 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.416136 4860 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.416146 4860 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.416156 4860 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.416166 4860 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.416175 4860 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.416185 4860 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.416194 4860 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.416203 4860 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.416213 4860 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.416223 4860 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.416233 4860 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.416243 4860 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.416254 4860 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.416263 4860 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.416273 4860 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.416283 4860 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.416293 4860 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.416302 4860 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.416312 4860 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.416322 4860 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.416332 4860 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.416341 4860 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.416351 4860 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.416362 4860 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.416371 4860 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.416379 4860 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.416387 4860 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.416396 4860 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.416404 4860 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.416412 4860 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.416420 4860 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.416428 4860 feature_gate.go:330] unrecognized feature gate: Example Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.416436 4860 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.416443 4860 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.416451 4860 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.416459 4860 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.416466 4860 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.416474 4860 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.416488 4860 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.416739 4860 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.416774 4860 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.416786 4860 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.416796 4860 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.416803 4860 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.416811 4860 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.416819 4860 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.416826 4860 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.416834 4860 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.416842 4860 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.416849 4860 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.416857 4860 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.416864 4860 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.416872 4860 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.416880 4860 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.416887 4860 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.416896 4860 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.416907 4860 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.416918 4860 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.416926 4860 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.416934 4860 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.416942 4860 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.416951 4860 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.416961 4860 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.416970 4860 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.416977 4860 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.416985 4860 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.416992 4860 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.417000 4860 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.417009 4860 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.417017 4860 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.417024 4860 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.417034 4860 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.417043 4860 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.417051 4860 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.417060 4860 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.417068 4860 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.417076 4860 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.417085 4860 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.417093 4860 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.417101 4860 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.417109 4860 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.417117 4860 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.417125 4860 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.417133 4860 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.417140 4860 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.417148 4860 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.417158 4860 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.417167 4860 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.417175 4860 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.417184 4860 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.417193 4860 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.417201 4860 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.417208 4860 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.417216 4860 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.417224 4860 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.417232 4860 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.417242 4860 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.417251 4860 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.417261 4860 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.417273 4860 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.417282 4860 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.418337 4860 feature_gate.go:330] unrecognized feature gate: Example Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.418865 4860 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.418878 4860 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.418890 4860 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.419111 4860 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.419230 4860 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.419475 4860 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.419498 4860 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.419512 4860 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.419527 4860 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.420104 4860 server.go:940] "Client rotation is on, will bootstrap in background" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.424526 4860 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.424705 4860 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.425600 4860 server.go:997] "Starting client certificate rotation" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.425671 4860 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.425900 4860 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-11-27 18:05:06.07629534 +0000 UTC Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.426009 4860 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.433005 4860 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 11 08:11:07 crc kubenswrapper[4860]: E1211 08:11:07.434510 4860 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.169:6443: connect: connection refused" logger="UnhandledError" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.435297 4860 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.443437 4860 log.go:25] "Validated CRI v1 runtime API" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.459744 4860 log.go:25] "Validated CRI v1 image API" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.460782 4860 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.464213 4860 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-12-11-08-06-17-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.464266 4860 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.492582 4860 manager.go:217] Machine: {Timestamp:2025-12-11 08:11:07.489915533 +0000 UTC m=+0.218434668 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:299d3cd2-e1ad-4f5c-8af4-f12c0a5d76e7 BootID:52631a1a-bc27-4dd5-8309-5dc39b5dcc41 Filesystems:[{Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:29:76:3d Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:29:76:3d Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:26:88:1f Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:41:86:a3 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:cd:7c:0e Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:95:e4:56 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:f2:e9:df:d0:5c:0c Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:e6:cb:bc:89:f8:3f Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.493065 4860 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.493291 4860 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.494155 4860 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.494482 4860 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.494530 4860 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.495010 4860 topology_manager.go:138] "Creating topology manager with none policy" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.495032 4860 container_manager_linux.go:303] "Creating device plugin manager" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.495373 4860 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.495416 4860 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.495871 4860 state_mem.go:36] "Initialized new in-memory state store" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.496207 4860 server.go:1245] "Using root directory" path="/var/lib/kubelet" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.497359 4860 kubelet.go:418] "Attempting to sync node with API server" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.497394 4860 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.497435 4860 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.497466 4860 kubelet.go:324] "Adding apiserver pod source" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.497486 4860 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.499899 4860 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.169:6443: connect: connection refused Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.500005 4860 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Dec 11 08:11:07 crc kubenswrapper[4860]: E1211 08:11:07.500115 4860 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.169:6443: connect: connection refused" logger="UnhandledError" Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.500346 4860 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.169:6443: connect: connection refused Dec 11 08:11:07 crc kubenswrapper[4860]: E1211 08:11:07.500490 4860 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.169:6443: connect: connection refused" logger="UnhandledError" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.500571 4860 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.502094 4860 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.502861 4860 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.502903 4860 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.502919 4860 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.502933 4860 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.502956 4860 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.502970 4860 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.502984 4860 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.503007 4860 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.503023 4860 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.503039 4860 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.503061 4860 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.503076 4860 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.503319 4860 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.504037 4860 server.go:1280] "Started kubelet" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.504558 4860 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.504576 4860 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.504728 4860 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.169:6443: connect: connection refused Dec 11 08:11:07 crc systemd[1]: Started Kubernetes Kubelet. Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.505853 4860 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.507494 4860 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.507558 4860 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.507778 4860 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-07 07:21:42.002012249 +0000 UTC Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.507857 4860 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 647h10m34.49416097s for next certificate rotation Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.509305 4860 volume_manager.go:287] "The desired_state_of_world populator starts" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.509335 4860 volume_manager.go:289] "Starting Kubelet Volume Manager" Dec 11 08:11:07 crc kubenswrapper[4860]: E1211 08:11:07.508088 4860 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.169:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.18801aeef29a87e8 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 08:11:07.503986664 +0000 UTC m=+0.232505759,LastTimestamp:2025-12-11 08:11:07.503986664 +0000 UTC m=+0.232505759,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.509531 4860 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Dec 11 08:11:07 crc kubenswrapper[4860]: E1211 08:11:07.509523 4860 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.510242 4860 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.169:6443: connect: connection refused Dec 11 08:11:07 crc kubenswrapper[4860]: E1211 08:11:07.510331 4860 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.169:6443: connect: connection refused" logger="UnhandledError" Dec 11 08:11:07 crc kubenswrapper[4860]: E1211 08:11:07.510951 4860 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.169:6443: connect: connection refused" interval="200ms" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.511448 4860 server.go:460] "Adding debug handlers to kubelet server" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.513044 4860 factory.go:153] Registering CRI-O factory Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.513078 4860 factory.go:221] Registration of the crio container factory successfully Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.513166 4860 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.513180 4860 factory.go:55] Registering systemd factory Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.513192 4860 factory.go:221] Registration of the systemd container factory successfully Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.513221 4860 factory.go:103] Registering Raw factory Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.513247 4860 manager.go:1196] Started watching for new ooms in manager Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.514691 4860 manager.go:319] Starting recovery of all containers Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.528357 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.528503 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.528536 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.528563 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.528588 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.528612 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.528635 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.528697 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.528726 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.528751 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.528774 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.528797 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.528819 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.528845 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.528869 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.528892 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.528921 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.528943 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.529011 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.529072 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.529117 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.529137 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.529156 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.529174 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.529194 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.529211 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.529235 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.529255 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.529276 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.529293 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.529312 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.529331 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.529351 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.529370 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.529388 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.529406 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.530264 4860 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.530316 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.530339 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.530361 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.530378 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.530399 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.530416 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.530433 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.530451 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.530469 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.530489 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.530508 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.530528 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.530547 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.530565 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.530583 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.530600 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.530623 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.530677 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.530708 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.530731 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.530749 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.530769 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.530788 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.530804 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.530821 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.530839 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.530860 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.530880 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.530930 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.530948 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.530969 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.531039 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.531060 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.531081 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.531098 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.531115 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.531132 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.531149 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.531166 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.531187 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.531204 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.531221 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.531239 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.531261 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.531278 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.531296 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.531314 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.531331 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.531349 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.531366 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.531384 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.531401 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.531419 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.531439 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.531458 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.531475 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.531494 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.531512 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.531530 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.531551 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.531567 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.531591 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.531609 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.531628 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.531713 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.531744 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.531766 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.531788 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.531817 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.531835 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.531855 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.531873 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.531892 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.531912 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.531931 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.531951 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.531969 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.531988 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.532007 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.532029 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.532049 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.532066 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.532084 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.532101 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.532119 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.532138 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.532156 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.532174 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.532191 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.532208 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.532226 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.532244 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.532262 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.532284 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.532303 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.532321 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.532340 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.532359 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.532377 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.532394 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.532413 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.532430 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.532448 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.532467 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.532485 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.532502 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.532547 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.532567 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.532585 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.532602 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.532620 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.532668 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.532697 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.532716 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.532733 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.532752 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.532771 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.532789 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.532806 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.532826 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.532844 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.532862 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.532880 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.532899 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.532916 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.532937 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.532954 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.532974 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.532991 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.533011 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.533027 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.533044 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.533061 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.533078 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.533097 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.533115 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.533133 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.533150 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.533167 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.533188 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.533204 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.533222 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.533240 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.533259 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.533319 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.533339 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.533358 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.533375 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.533393 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.533412 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.533430 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.533449 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.533466 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.533484 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.533504 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.533522 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.533547 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.533567 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.533585 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.533602 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.533619 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.533636 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.533691 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.533714 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.533735 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.533759 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.533782 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.533803 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.533824 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.533851 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.533870 4860 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.533888 4860 reconstruct.go:97] "Volume reconstruction finished" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.533902 4860 reconciler.go:26] "Reconciler: start to sync state" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.550089 4860 manager.go:324] Recovery completed Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.566211 4860 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.568216 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.568267 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.568279 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.569135 4860 cpu_manager.go:225] "Starting CPU manager" policy="none" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.569168 4860 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.569289 4860 state_mem.go:36] "Initialized new in-memory state store" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.576101 4860 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.577480 4860 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.577525 4860 status_manager.go:217] "Starting to sync pod status with apiserver" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.577554 4860 kubelet.go:2335] "Starting kubelet main sync loop" Dec 11 08:11:07 crc kubenswrapper[4860]: E1211 08:11:07.577602 4860 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Dec 11 08:11:07 crc kubenswrapper[4860]: E1211 08:11:07.611080 4860 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 08:11:07 crc kubenswrapper[4860]: W1211 08:11:07.621741 4860 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.169:6443: connect: connection refused Dec 11 08:11:07 crc kubenswrapper[4860]: E1211 08:11:07.621815 4860 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.169:6443: connect: connection refused" logger="UnhandledError" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.629928 4860 policy_none.go:49] "None policy: Start" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.630887 4860 memory_manager.go:170] "Starting memorymanager" policy="None" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.630908 4860 state_mem.go:35] "Initializing new in-memory state store" Dec 11 08:11:07 crc kubenswrapper[4860]: E1211 08:11:07.677713 4860 kubelet.go:2359] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.700944 4860 manager.go:334] "Starting Device Plugin manager" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.701385 4860 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.702057 4860 server.go:79] "Starting device plugin registration server" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.702890 4860 eviction_manager.go:189] "Eviction manager: starting control loop" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.702924 4860 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.703519 4860 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.703789 4860 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.703826 4860 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Dec 11 08:11:07 crc kubenswrapper[4860]: E1211 08:11:07.712037 4860 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.169:6443: connect: connection refused" interval="400ms" Dec 11 08:11:07 crc kubenswrapper[4860]: E1211 08:11:07.718534 4860 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.805807 4860 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.807508 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.807549 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.807564 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.807594 4860 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 11 08:11:07 crc kubenswrapper[4860]: E1211 08:11:07.808155 4860 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.169:6443: connect: connection refused" node="crc" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.878272 4860 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.878429 4860 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.879734 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.879803 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.879817 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.879998 4860 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.880355 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.880462 4860 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.880988 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.881032 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.881048 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.881217 4860 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.881370 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.881431 4860 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.882002 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.882029 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.882039 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.882291 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.882330 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.882342 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.882473 4860 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.882586 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.882613 4860 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.883308 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.883369 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.883393 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.883421 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.883453 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.883464 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.883564 4860 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.883719 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.883764 4860 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.884022 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.884038 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.884049 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.884568 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.884627 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.884685 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.884735 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.884751 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.884761 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.885003 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.885063 4860 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.886341 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.886390 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.886400 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.938675 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.938708 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.938725 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.938742 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.938757 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.938770 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.938947 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.939046 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.939117 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.939196 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.939255 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.939282 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.939307 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.939327 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 11 08:11:07 crc kubenswrapper[4860]: I1211 08:11:07.939350 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 11 08:11:08 crc kubenswrapper[4860]: I1211 08:11:08.009324 4860 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 08:11:08 crc kubenswrapper[4860]: I1211 08:11:08.011061 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:08 crc kubenswrapper[4860]: I1211 08:11:08.011120 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:08 crc kubenswrapper[4860]: I1211 08:11:08.011141 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:08 crc kubenswrapper[4860]: I1211 08:11:08.011210 4860 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 11 08:11:08 crc kubenswrapper[4860]: E1211 08:11:08.011915 4860 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.169:6443: connect: connection refused" node="crc" Dec 11 08:11:08 crc kubenswrapper[4860]: I1211 08:11:08.040166 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 11 08:11:08 crc kubenswrapper[4860]: I1211 08:11:08.040230 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 11 08:11:08 crc kubenswrapper[4860]: I1211 08:11:08.040267 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 11 08:11:08 crc kubenswrapper[4860]: I1211 08:11:08.040302 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 08:11:08 crc kubenswrapper[4860]: I1211 08:11:08.040333 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 08:11:08 crc kubenswrapper[4860]: I1211 08:11:08.040362 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 11 08:11:08 crc kubenswrapper[4860]: I1211 08:11:08.040429 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 11 08:11:08 crc kubenswrapper[4860]: I1211 08:11:08.040389 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 08:11:08 crc kubenswrapper[4860]: I1211 08:11:08.040462 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 11 08:11:08 crc kubenswrapper[4860]: I1211 08:11:08.040496 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 11 08:11:08 crc kubenswrapper[4860]: I1211 08:11:08.040531 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 11 08:11:08 crc kubenswrapper[4860]: I1211 08:11:08.040543 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 08:11:08 crc kubenswrapper[4860]: I1211 08:11:08.040554 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 11 08:11:08 crc kubenswrapper[4860]: I1211 08:11:08.040573 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 08:11:08 crc kubenswrapper[4860]: I1211 08:11:08.040637 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 11 08:11:08 crc kubenswrapper[4860]: I1211 08:11:08.040563 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 11 08:11:08 crc kubenswrapper[4860]: I1211 08:11:08.040603 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 11 08:11:08 crc kubenswrapper[4860]: I1211 08:11:08.040616 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 11 08:11:08 crc kubenswrapper[4860]: I1211 08:11:08.040759 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 11 08:11:08 crc kubenswrapper[4860]: I1211 08:11:08.040697 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 11 08:11:08 crc kubenswrapper[4860]: I1211 08:11:08.040614 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 08:11:08 crc kubenswrapper[4860]: I1211 08:11:08.040796 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 11 08:11:08 crc kubenswrapper[4860]: I1211 08:11:08.040830 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 11 08:11:08 crc kubenswrapper[4860]: I1211 08:11:08.040850 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 11 08:11:08 crc kubenswrapper[4860]: I1211 08:11:08.040859 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 11 08:11:08 crc kubenswrapper[4860]: I1211 08:11:08.040898 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 11 08:11:08 crc kubenswrapper[4860]: I1211 08:11:08.040892 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 11 08:11:08 crc kubenswrapper[4860]: I1211 08:11:08.040930 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 11 08:11:08 crc kubenswrapper[4860]: I1211 08:11:08.040834 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 11 08:11:08 crc kubenswrapper[4860]: I1211 08:11:08.040996 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 11 08:11:08 crc kubenswrapper[4860]: E1211 08:11:08.113846 4860 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.169:6443: connect: connection refused" interval="800ms" Dec 11 08:11:08 crc kubenswrapper[4860]: I1211 08:11:08.234471 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 11 08:11:08 crc kubenswrapper[4860]: I1211 08:11:08.258983 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 11 08:11:08 crc kubenswrapper[4860]: W1211 08:11:08.268930 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-cb16bfef591e923187690b30fbbf173953b75032c65618f1b7c52d957724b7ed WatchSource:0}: Error finding container cb16bfef591e923187690b30fbbf173953b75032c65618f1b7c52d957724b7ed: Status 404 returned error can't find the container with id cb16bfef591e923187690b30fbbf173953b75032c65618f1b7c52d957724b7ed Dec 11 08:11:08 crc kubenswrapper[4860]: I1211 08:11:08.271217 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 08:11:08 crc kubenswrapper[4860]: W1211 08:11:08.284003 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-a04d1b7847cdbc2785686d448f613fdfb028a11937dcefe80fc657dade72d11a WatchSource:0}: Error finding container a04d1b7847cdbc2785686d448f613fdfb028a11937dcefe80fc657dade72d11a: Status 404 returned error can't find the container with id a04d1b7847cdbc2785686d448f613fdfb028a11937dcefe80fc657dade72d11a Dec 11 08:11:08 crc kubenswrapper[4860]: I1211 08:11:08.297629 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 11 08:11:08 crc kubenswrapper[4860]: I1211 08:11:08.305564 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 11 08:11:08 crc kubenswrapper[4860]: W1211 08:11:08.320252 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-670f254db17fac1308175c52e54df16da346cc58dc7d75c0ac6c273ae8c18329 WatchSource:0}: Error finding container 670f254db17fac1308175c52e54df16da346cc58dc7d75c0ac6c273ae8c18329: Status 404 returned error can't find the container with id 670f254db17fac1308175c52e54df16da346cc58dc7d75c0ac6c273ae8c18329 Dec 11 08:11:08 crc kubenswrapper[4860]: W1211 08:11:08.320768 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-c9fafe1158a4ecb7ae719ed21cd9ac88e17e93b6929b7dd52b050348f002127a WatchSource:0}: Error finding container c9fafe1158a4ecb7ae719ed21cd9ac88e17e93b6929b7dd52b050348f002127a: Status 404 returned error can't find the container with id c9fafe1158a4ecb7ae719ed21cd9ac88e17e93b6929b7dd52b050348f002127a Dec 11 08:11:08 crc kubenswrapper[4860]: I1211 08:11:08.412329 4860 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 08:11:08 crc kubenswrapper[4860]: I1211 08:11:08.414117 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:08 crc kubenswrapper[4860]: I1211 08:11:08.414158 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:08 crc kubenswrapper[4860]: I1211 08:11:08.414166 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:08 crc kubenswrapper[4860]: I1211 08:11:08.414189 4860 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 11 08:11:08 crc kubenswrapper[4860]: E1211 08:11:08.414682 4860 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.169:6443: connect: connection refused" node="crc" Dec 11 08:11:08 crc kubenswrapper[4860]: I1211 08:11:08.506158 4860 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.169:6443: connect: connection refused Dec 11 08:11:08 crc kubenswrapper[4860]: W1211 08:11:08.544813 4860 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.169:6443: connect: connection refused Dec 11 08:11:08 crc kubenswrapper[4860]: E1211 08:11:08.544889 4860 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.169:6443: connect: connection refused" logger="UnhandledError" Dec 11 08:11:08 crc kubenswrapper[4860]: I1211 08:11:08.581369 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"d5a5418d863fbc2b9a000e5d2d8c24da360a097410f6afcbd8e488b66ca75b79"} Dec 11 08:11:08 crc kubenswrapper[4860]: I1211 08:11:08.582196 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"a04d1b7847cdbc2785686d448f613fdfb028a11937dcefe80fc657dade72d11a"} Dec 11 08:11:08 crc kubenswrapper[4860]: I1211 08:11:08.583185 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"cb16bfef591e923187690b30fbbf173953b75032c65618f1b7c52d957724b7ed"} Dec 11 08:11:08 crc kubenswrapper[4860]: I1211 08:11:08.584011 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"c9fafe1158a4ecb7ae719ed21cd9ac88e17e93b6929b7dd52b050348f002127a"} Dec 11 08:11:08 crc kubenswrapper[4860]: I1211 08:11:08.584999 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"670f254db17fac1308175c52e54df16da346cc58dc7d75c0ac6c273ae8c18329"} Dec 11 08:11:08 crc kubenswrapper[4860]: W1211 08:11:08.707305 4860 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.169:6443: connect: connection refused Dec 11 08:11:08 crc kubenswrapper[4860]: E1211 08:11:08.707682 4860 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.169:6443: connect: connection refused" logger="UnhandledError" Dec 11 08:11:08 crc kubenswrapper[4860]: E1211 08:11:08.915326 4860 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.169:6443: connect: connection refused" interval="1.6s" Dec 11 08:11:09 crc kubenswrapper[4860]: W1211 08:11:09.056068 4860 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.169:6443: connect: connection refused Dec 11 08:11:09 crc kubenswrapper[4860]: E1211 08:11:09.056161 4860 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.169:6443: connect: connection refused" logger="UnhandledError" Dec 11 08:11:09 crc kubenswrapper[4860]: W1211 08:11:09.077228 4860 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.169:6443: connect: connection refused Dec 11 08:11:09 crc kubenswrapper[4860]: E1211 08:11:09.077354 4860 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.169:6443: connect: connection refused" logger="UnhandledError" Dec 11 08:11:09 crc kubenswrapper[4860]: I1211 08:11:09.214786 4860 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 08:11:09 crc kubenswrapper[4860]: I1211 08:11:09.216224 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:09 crc kubenswrapper[4860]: I1211 08:11:09.216286 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:09 crc kubenswrapper[4860]: I1211 08:11:09.216311 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:09 crc kubenswrapper[4860]: I1211 08:11:09.216358 4860 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 11 08:11:09 crc kubenswrapper[4860]: E1211 08:11:09.217004 4860 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.169:6443: connect: connection refused" node="crc" Dec 11 08:11:09 crc kubenswrapper[4860]: I1211 08:11:09.483088 4860 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Dec 11 08:11:09 crc kubenswrapper[4860]: E1211 08:11:09.484265 4860 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.169:6443: connect: connection refused" logger="UnhandledError" Dec 11 08:11:09 crc kubenswrapper[4860]: I1211 08:11:09.506458 4860 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.169:6443: connect: connection refused Dec 11 08:11:09 crc kubenswrapper[4860]: I1211 08:11:09.589679 4860 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="d5de794187456b3576b969069a3b839d75c7918418a97fa35dff8bb40bb76e39" exitCode=0 Dec 11 08:11:09 crc kubenswrapper[4860]: I1211 08:11:09.589757 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"d5de794187456b3576b969069a3b839d75c7918418a97fa35dff8bb40bb76e39"} Dec 11 08:11:09 crc kubenswrapper[4860]: I1211 08:11:09.589830 4860 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 08:11:09 crc kubenswrapper[4860]: I1211 08:11:09.591220 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:09 crc kubenswrapper[4860]: I1211 08:11:09.591405 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:09 crc kubenswrapper[4860]: I1211 08:11:09.591536 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:09 crc kubenswrapper[4860]: I1211 08:11:09.591801 4860 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="3498078e25d1156a89ddc17f4823182b1aeb8b283f8b46bf66df67d67c5c5e44" exitCode=0 Dec 11 08:11:09 crc kubenswrapper[4860]: I1211 08:11:09.591882 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"3498078e25d1156a89ddc17f4823182b1aeb8b283f8b46bf66df67d67c5c5e44"} Dec 11 08:11:09 crc kubenswrapper[4860]: I1211 08:11:09.591985 4860 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 08:11:09 crc kubenswrapper[4860]: I1211 08:11:09.592749 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:09 crc kubenswrapper[4860]: I1211 08:11:09.592777 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:09 crc kubenswrapper[4860]: I1211 08:11:09.592789 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:09 crc kubenswrapper[4860]: I1211 08:11:09.596333 4860 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="4602fc6635e6188bf4cb75c5c7cf1e71a64fc0c5420626bfafd87beae150722d" exitCode=0 Dec 11 08:11:09 crc kubenswrapper[4860]: I1211 08:11:09.596396 4860 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 08:11:09 crc kubenswrapper[4860]: I1211 08:11:09.596388 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"4602fc6635e6188bf4cb75c5c7cf1e71a64fc0c5420626bfafd87beae150722d"} Dec 11 08:11:09 crc kubenswrapper[4860]: I1211 08:11:09.600136 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:09 crc kubenswrapper[4860]: I1211 08:11:09.600165 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:09 crc kubenswrapper[4860]: I1211 08:11:09.600177 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:09 crc kubenswrapper[4860]: I1211 08:11:09.602149 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"a5483384f1d2ebc1fefa9706c9e67a3dc5a00ff643f803e3f2707c252e27ad94"} Dec 11 08:11:09 crc kubenswrapper[4860]: I1211 08:11:09.602184 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"3d7c919899b74a2b0fcd9221f3f769a9d013267354b200a282669d342d6413d5"} Dec 11 08:11:09 crc kubenswrapper[4860]: I1211 08:11:09.602197 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"5ad4ba1b769c6cb2b34073798f8e3c25eb0177e6d3059484bf365e6981cab710"} Dec 11 08:11:09 crc kubenswrapper[4860]: I1211 08:11:09.602209 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"077aa1dd2fe5189ff60a0f2365e47a6a545cca0de7aa6379370908db4ea5838a"} Dec 11 08:11:09 crc kubenswrapper[4860]: I1211 08:11:09.602283 4860 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 08:11:09 crc kubenswrapper[4860]: I1211 08:11:09.604359 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:09 crc kubenswrapper[4860]: I1211 08:11:09.604390 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:09 crc kubenswrapper[4860]: I1211 08:11:09.604402 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:09 crc kubenswrapper[4860]: I1211 08:11:09.608446 4860 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="cf0aff8a6ab816b6617b1feeca9146d41bb5e8808bf1db96f8d493500c135d85" exitCode=0 Dec 11 08:11:09 crc kubenswrapper[4860]: I1211 08:11:09.608510 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"cf0aff8a6ab816b6617b1feeca9146d41bb5e8808bf1db96f8d493500c135d85"} Dec 11 08:11:09 crc kubenswrapper[4860]: I1211 08:11:09.608690 4860 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 08:11:09 crc kubenswrapper[4860]: I1211 08:11:09.609816 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:09 crc kubenswrapper[4860]: I1211 08:11:09.609989 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:09 crc kubenswrapper[4860]: I1211 08:11:09.610106 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:09 crc kubenswrapper[4860]: I1211 08:11:09.614879 4860 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 08:11:09 crc kubenswrapper[4860]: I1211 08:11:09.615870 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:09 crc kubenswrapper[4860]: I1211 08:11:09.615906 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:09 crc kubenswrapper[4860]: I1211 08:11:09.615918 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:09 crc kubenswrapper[4860]: I1211 08:11:09.623155 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 11 08:11:10 crc kubenswrapper[4860]: I1211 08:11:10.614906 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"867cdac386b0576ee2958a520afd522a2558a8878ea6b5af977b136dc7bba4a5"} Dec 11 08:11:10 crc kubenswrapper[4860]: I1211 08:11:10.615008 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"1348b83ed7b1e8c25f9dfb1b2d2d035b5adc96dc23854ee90236f06fb155f155"} Dec 11 08:11:10 crc kubenswrapper[4860]: I1211 08:11:10.615026 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"122eb7e14a042027de3aafaa8db89562d80565e7df20f302f872e48d412a26b3"} Dec 11 08:11:10 crc kubenswrapper[4860]: I1211 08:11:10.614952 4860 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 08:11:10 crc kubenswrapper[4860]: I1211 08:11:10.615974 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:10 crc kubenswrapper[4860]: I1211 08:11:10.616030 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:10 crc kubenswrapper[4860]: I1211 08:11:10.616046 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:10 crc kubenswrapper[4860]: I1211 08:11:10.618820 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"2090cab02ad74ddfd6bb57ff50eba35d18f56b0cde1c230fe011c01994b693bc"} Dec 11 08:11:10 crc kubenswrapper[4860]: I1211 08:11:10.618876 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"f2274fb298ca7d2b37b060b4dbf9b166d39ba77a108b08edeeb4d67196e72cb3"} Dec 11 08:11:10 crc kubenswrapper[4860]: I1211 08:11:10.618896 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"5a498aa90f914e1f0370df01dc5ef174df113a2a16d86b2e3f24b19a1b99b4c0"} Dec 11 08:11:10 crc kubenswrapper[4860]: I1211 08:11:10.618913 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"f8caebd2f4dfcbb89c597029a07c335171889761d2d200206987537fb0a228db"} Dec 11 08:11:10 crc kubenswrapper[4860]: I1211 08:11:10.618928 4860 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 08:11:10 crc kubenswrapper[4860]: I1211 08:11:10.618930 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"e831db1a15976d3967d56479689572c26f2599a766baa0402897a3008428c56f"} Dec 11 08:11:10 crc kubenswrapper[4860]: I1211 08:11:10.619705 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:10 crc kubenswrapper[4860]: I1211 08:11:10.619728 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:10 crc kubenswrapper[4860]: I1211 08:11:10.619737 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:10 crc kubenswrapper[4860]: I1211 08:11:10.620567 4860 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="08ce42b974fba00c409c0af971f0de351e50f18b2f06f8e722150da3239bb908" exitCode=0 Dec 11 08:11:10 crc kubenswrapper[4860]: I1211 08:11:10.620661 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"08ce42b974fba00c409c0af971f0de351e50f18b2f06f8e722150da3239bb908"} Dec 11 08:11:10 crc kubenswrapper[4860]: I1211 08:11:10.620673 4860 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 08:11:10 crc kubenswrapper[4860]: I1211 08:11:10.621362 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:10 crc kubenswrapper[4860]: I1211 08:11:10.621389 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:10 crc kubenswrapper[4860]: I1211 08:11:10.621398 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:10 crc kubenswrapper[4860]: I1211 08:11:10.622485 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"77c0c831f046149002cd12564aa9b1544a813da1fcd81e61451eeb504f0450b8"} Dec 11 08:11:10 crc kubenswrapper[4860]: I1211 08:11:10.622524 4860 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 08:11:10 crc kubenswrapper[4860]: I1211 08:11:10.622528 4860 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 08:11:10 crc kubenswrapper[4860]: I1211 08:11:10.623370 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:10 crc kubenswrapper[4860]: I1211 08:11:10.623416 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:10 crc kubenswrapper[4860]: I1211 08:11:10.623429 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:10 crc kubenswrapper[4860]: I1211 08:11:10.623940 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:10 crc kubenswrapper[4860]: I1211 08:11:10.623962 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:10 crc kubenswrapper[4860]: I1211 08:11:10.623970 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:10 crc kubenswrapper[4860]: I1211 08:11:10.817261 4860 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 08:11:10 crc kubenswrapper[4860]: I1211 08:11:10.820454 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:10 crc kubenswrapper[4860]: I1211 08:11:10.820502 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:10 crc kubenswrapper[4860]: I1211 08:11:10.820520 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:10 crc kubenswrapper[4860]: I1211 08:11:10.820551 4860 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 11 08:11:10 crc kubenswrapper[4860]: I1211 08:11:10.991315 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 11 08:11:10 crc kubenswrapper[4860]: I1211 08:11:10.996931 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 11 08:11:11 crc kubenswrapper[4860]: I1211 08:11:11.627482 4860 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="cfac6be03e3cfa0b0f3450b48b67957f2d5172b9c65e27f14c2a260bda5d423f" exitCode=0 Dec 11 08:11:11 crc kubenswrapper[4860]: I1211 08:11:11.627583 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"cfac6be03e3cfa0b0f3450b48b67957f2d5172b9c65e27f14c2a260bda5d423f"} Dec 11 08:11:11 crc kubenswrapper[4860]: I1211 08:11:11.627630 4860 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 08:11:11 crc kubenswrapper[4860]: I1211 08:11:11.627667 4860 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 08:11:11 crc kubenswrapper[4860]: I1211 08:11:11.627717 4860 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 08:11:11 crc kubenswrapper[4860]: I1211 08:11:11.627730 4860 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 08:11:11 crc kubenswrapper[4860]: I1211 08:11:11.627800 4860 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 11 08:11:11 crc kubenswrapper[4860]: I1211 08:11:11.627884 4860 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 08:11:11 crc kubenswrapper[4860]: I1211 08:11:11.628929 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 08:11:11 crc kubenswrapper[4860]: I1211 08:11:11.629131 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 11 08:11:11 crc kubenswrapper[4860]: I1211 08:11:11.631956 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:11 crc kubenswrapper[4860]: I1211 08:11:11.631964 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:11 crc kubenswrapper[4860]: I1211 08:11:11.631983 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:11 crc kubenswrapper[4860]: I1211 08:11:11.631994 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:11 crc kubenswrapper[4860]: I1211 08:11:11.632006 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:11 crc kubenswrapper[4860]: I1211 08:11:11.632034 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:11 crc kubenswrapper[4860]: I1211 08:11:11.632064 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:11 crc kubenswrapper[4860]: I1211 08:11:11.632080 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:11 crc kubenswrapper[4860]: I1211 08:11:11.632088 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:11 crc kubenswrapper[4860]: I1211 08:11:11.632114 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:11 crc kubenswrapper[4860]: I1211 08:11:11.632132 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:11 crc kubenswrapper[4860]: I1211 08:11:11.632142 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:11 crc kubenswrapper[4860]: I1211 08:11:11.632068 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:11 crc kubenswrapper[4860]: I1211 08:11:11.631996 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:11 crc kubenswrapper[4860]: I1211 08:11:11.632582 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:12 crc kubenswrapper[4860]: I1211 08:11:12.624037 4860 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 11 08:11:12 crc kubenswrapper[4860]: I1211 08:11:12.624109 4860 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 11 08:11:12 crc kubenswrapper[4860]: I1211 08:11:12.633397 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"eedc9988e1a4a3c8cae730dc4a352103a6d665535b825e157d0f87c331b137c5"} Dec 11 08:11:12 crc kubenswrapper[4860]: I1211 08:11:12.633448 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"bbe00b46063d3968d59c393b87178e135f982abb7d0f32fa798b212d4b83df6f"} Dec 11 08:11:12 crc kubenswrapper[4860]: I1211 08:11:12.633466 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"6f74a32a8fefea704cc73ffd1114083dab68aa44dc33f8d9d432d2d24216546c"} Dec 11 08:11:12 crc kubenswrapper[4860]: I1211 08:11:12.633478 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"f683e48f3577a68862eb6302383538376e41f49b20b1e0ea3f1b3b143f24872a"} Dec 11 08:11:12 crc kubenswrapper[4860]: I1211 08:11:12.633500 4860 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 08:11:12 crc kubenswrapper[4860]: I1211 08:11:12.633501 4860 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 08:11:12 crc kubenswrapper[4860]: I1211 08:11:12.634751 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:12 crc kubenswrapper[4860]: I1211 08:11:12.634781 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:12 crc kubenswrapper[4860]: I1211 08:11:12.634791 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:12 crc kubenswrapper[4860]: I1211 08:11:12.634821 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:12 crc kubenswrapper[4860]: I1211 08:11:12.634848 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:12 crc kubenswrapper[4860]: I1211 08:11:12.634887 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:13 crc kubenswrapper[4860]: I1211 08:11:13.593399 4860 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Dec 11 08:11:13 crc kubenswrapper[4860]: I1211 08:11:13.638770 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"2137fee4b7abc9765f9517d7b3b6ea796601a31d03151bf167d17136ced8b25a"} Dec 11 08:11:13 crc kubenswrapper[4860]: I1211 08:11:13.638919 4860 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 08:11:13 crc kubenswrapper[4860]: I1211 08:11:13.639745 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:13 crc kubenswrapper[4860]: I1211 08:11:13.639786 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:13 crc kubenswrapper[4860]: I1211 08:11:13.639799 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:14 crc kubenswrapper[4860]: I1211 08:11:14.058139 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 11 08:11:14 crc kubenswrapper[4860]: I1211 08:11:14.058744 4860 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 08:11:14 crc kubenswrapper[4860]: I1211 08:11:14.060276 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:14 crc kubenswrapper[4860]: I1211 08:11:14.060340 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:14 crc kubenswrapper[4860]: I1211 08:11:14.060358 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:14 crc kubenswrapper[4860]: I1211 08:11:14.641427 4860 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 08:11:14 crc kubenswrapper[4860]: I1211 08:11:14.642313 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:14 crc kubenswrapper[4860]: I1211 08:11:14.642367 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:14 crc kubenswrapper[4860]: I1211 08:11:14.642386 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:14 crc kubenswrapper[4860]: I1211 08:11:14.920843 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 08:11:14 crc kubenswrapper[4860]: I1211 08:11:14.921067 4860 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 08:11:14 crc kubenswrapper[4860]: I1211 08:11:14.922580 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:14 crc kubenswrapper[4860]: I1211 08:11:14.922683 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:14 crc kubenswrapper[4860]: I1211 08:11:14.922708 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:15 crc kubenswrapper[4860]: I1211 08:11:15.222724 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 08:11:15 crc kubenswrapper[4860]: I1211 08:11:15.644333 4860 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 08:11:15 crc kubenswrapper[4860]: I1211 08:11:15.646005 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:15 crc kubenswrapper[4860]: I1211 08:11:15.646061 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:15 crc kubenswrapper[4860]: I1211 08:11:15.646126 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:15 crc kubenswrapper[4860]: I1211 08:11:15.713505 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Dec 11 08:11:15 crc kubenswrapper[4860]: I1211 08:11:15.713787 4860 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 08:11:15 crc kubenswrapper[4860]: I1211 08:11:15.715240 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:15 crc kubenswrapper[4860]: I1211 08:11:15.715291 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:15 crc kubenswrapper[4860]: I1211 08:11:15.715308 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:17 crc kubenswrapper[4860]: I1211 08:11:17.516436 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Dec 11 08:11:17 crc kubenswrapper[4860]: I1211 08:11:17.516727 4860 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 08:11:17 crc kubenswrapper[4860]: I1211 08:11:17.518078 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:17 crc kubenswrapper[4860]: I1211 08:11:17.518117 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:17 crc kubenswrapper[4860]: I1211 08:11:17.518128 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:17 crc kubenswrapper[4860]: E1211 08:11:17.718696 4860 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 11 08:11:18 crc kubenswrapper[4860]: I1211 08:11:18.224181 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 11 08:11:18 crc kubenswrapper[4860]: I1211 08:11:18.224432 4860 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 08:11:18 crc kubenswrapper[4860]: I1211 08:11:18.226221 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:18 crc kubenswrapper[4860]: I1211 08:11:18.226280 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:18 crc kubenswrapper[4860]: I1211 08:11:18.226296 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:18 crc kubenswrapper[4860]: I1211 08:11:18.230084 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 11 08:11:18 crc kubenswrapper[4860]: I1211 08:11:18.650602 4860 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 08:11:18 crc kubenswrapper[4860]: I1211 08:11:18.651638 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:18 crc kubenswrapper[4860]: I1211 08:11:18.651845 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:18 crc kubenswrapper[4860]: I1211 08:11:18.651969 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:20 crc kubenswrapper[4860]: I1211 08:11:20.412951 4860 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Dec 11 08:11:20 crc kubenswrapper[4860]: I1211 08:11:20.413022 4860 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Dec 11 08:11:20 crc kubenswrapper[4860]: I1211 08:11:20.506240 4860 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Dec 11 08:11:20 crc kubenswrapper[4860]: E1211 08:11:20.516559 4860 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" interval="3.2s" Dec 11 08:11:20 crc kubenswrapper[4860]: E1211 08:11:20.822008 4860 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": net/http: TLS handshake timeout" node="crc" Dec 11 08:11:20 crc kubenswrapper[4860]: W1211 08:11:20.888240 4860 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout Dec 11 08:11:20 crc kubenswrapper[4860]: I1211 08:11:20.888498 4860 trace.go:236] Trace[1325041008]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (11-Dec-2025 08:11:10.886) (total time: 10001ms): Dec 11 08:11:20 crc kubenswrapper[4860]: Trace[1325041008]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (08:11:20.888) Dec 11 08:11:20 crc kubenswrapper[4860]: Trace[1325041008]: [10.001903965s] [10.001903965s] END Dec 11 08:11:20 crc kubenswrapper[4860]: E1211 08:11:20.888660 4860 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Dec 11 08:11:20 crc kubenswrapper[4860]: W1211 08:11:20.965167 4860 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout Dec 11 08:11:20 crc kubenswrapper[4860]: I1211 08:11:20.965913 4860 trace.go:236] Trace[28038762]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (11-Dec-2025 08:11:10.963) (total time: 10002ms): Dec 11 08:11:20 crc kubenswrapper[4860]: Trace[28038762]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10002ms (08:11:20.965) Dec 11 08:11:20 crc kubenswrapper[4860]: Trace[28038762]: [10.002483252s] [10.002483252s] END Dec 11 08:11:20 crc kubenswrapper[4860]: E1211 08:11:20.966268 4860 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Dec 11 08:11:21 crc kubenswrapper[4860]: W1211 08:11:21.224720 4860 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout Dec 11 08:11:21 crc kubenswrapper[4860]: I1211 08:11:21.224820 4860 trace.go:236] Trace[566572862]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (11-Dec-2025 08:11:11.222) (total time: 10001ms): Dec 11 08:11:21 crc kubenswrapper[4860]: Trace[566572862]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (08:11:21.224) Dec 11 08:11:21 crc kubenswrapper[4860]: Trace[566572862]: [10.001816631s] [10.001816631s] END Dec 11 08:11:21 crc kubenswrapper[4860]: E1211 08:11:21.224844 4860 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Dec 11 08:11:21 crc kubenswrapper[4860]: I1211 08:11:21.250220 4860 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 11 08:11:21 crc kubenswrapper[4860]: I1211 08:11:21.250479 4860 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 11 08:11:21 crc kubenswrapper[4860]: I1211 08:11:21.262425 4860 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 11 08:11:21 crc kubenswrapper[4860]: I1211 08:11:21.262499 4860 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 11 08:11:22 crc kubenswrapper[4860]: I1211 08:11:22.623879 4860 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 11 08:11:22 crc kubenswrapper[4860]: I1211 08:11:22.623972 4860 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 11 08:11:24 crc kubenswrapper[4860]: I1211 08:11:24.022731 4860 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 08:11:24 crc kubenswrapper[4860]: I1211 08:11:24.024060 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:24 crc kubenswrapper[4860]: I1211 08:11:24.024231 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:24 crc kubenswrapper[4860]: I1211 08:11:24.024326 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:24 crc kubenswrapper[4860]: I1211 08:11:24.024417 4860 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 11 08:11:24 crc kubenswrapper[4860]: E1211 08:11:24.028436 4860 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Dec 11 08:11:25 crc kubenswrapper[4860]: I1211 08:11:25.227861 4860 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 11 08:11:25 crc kubenswrapper[4860]: I1211 08:11:25.233557 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 08:11:25 crc kubenswrapper[4860]: I1211 08:11:25.234206 4860 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 08:11:25 crc kubenswrapper[4860]: I1211 08:11:25.236505 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:25 crc kubenswrapper[4860]: I1211 08:11:25.236561 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:25 crc kubenswrapper[4860]: I1211 08:11:25.236583 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:25 crc kubenswrapper[4860]: I1211 08:11:25.241905 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 08:11:25 crc kubenswrapper[4860]: I1211 08:11:25.543078 4860 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 11 08:11:25 crc kubenswrapper[4860]: I1211 08:11:25.671157 4860 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 08:11:25 crc kubenswrapper[4860]: I1211 08:11:25.671919 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:25 crc kubenswrapper[4860]: I1211 08:11:25.671944 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:25 crc kubenswrapper[4860]: I1211 08:11:25.671951 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.253174 4860 trace.go:236] Trace[239727596]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (11-Dec-2025 08:11:11.609) (total time: 14643ms): Dec 11 08:11:26 crc kubenswrapper[4860]: Trace[239727596]: ---"Objects listed" error: 14643ms (08:11:26.253) Dec 11 08:11:26 crc kubenswrapper[4860]: Trace[239727596]: [14.643693153s] [14.643693153s] END Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.253200 4860 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.253845 4860 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.258420 4860 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.275791 4860 csr.go:261] certificate signing request csr-ntf64 is approved, waiting to be issued Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.283575 4860 csr.go:257] certificate signing request csr-ntf64 is issued Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.531109 4860 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:58852->192.168.126.11:17697: read: connection reset by peer" start-of-body= Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.531265 4860 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:58852->192.168.126.11:17697: read: connection reset by peer" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.531933 4860 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.532016 4860 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.600741 4860 apiserver.go:52] "Watching apiserver" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.604936 4860 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.605335 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb"] Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.605832 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.605922 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.606107 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:11:26 crc kubenswrapper[4860]: E1211 08:11:26.606178 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.606281 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 11 08:11:26 crc kubenswrapper[4860]: E1211 08:11:26.605947 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.606380 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.606430 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 11 08:11:26 crc kubenswrapper[4860]: E1211 08:11:26.606449 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.608442 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.609268 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.609372 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.609495 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.609733 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.610491 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.610545 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.612048 4860 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.612849 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.612855 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.645078 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.656582 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.656634 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.656676 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.657107 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.657182 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.657209 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.657224 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.657212 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.657289 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.657240 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.657323 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.657368 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.657401 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.657432 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.657458 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.657470 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.657484 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.657530 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.657556 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.657580 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.657592 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.657604 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.657626 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.657655 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.657660 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.657706 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.657729 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.657754 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.657772 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.657792 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.657814 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.657838 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.657855 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.657859 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.657871 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.657943 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.657971 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.658005 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.658033 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.658054 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.658076 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.658098 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.658118 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.658137 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.658156 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.658180 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.658201 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.658218 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.658244 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.658269 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.658299 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.658324 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.658348 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.658374 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.658399 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.658425 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.658451 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.658490 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.658517 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.658544 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.658565 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.658587 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.658609 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.658630 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.658676 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.658700 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.658726 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.658765 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.658823 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.658853 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.658884 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.658908 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.658930 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.658945 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.658965 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.658992 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.659009 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.659069 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.659095 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.659124 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.659146 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.659163 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.659183 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.659199 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.659224 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.659249 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.659273 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.659298 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.659320 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.659337 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.659353 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.659372 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.659391 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.659413 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.659434 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.659451 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.659471 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.659489 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.659521 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.659549 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.659581 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.659607 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.659686 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.659722 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.659764 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.659785 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.659807 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.659837 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.659860 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.659883 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.659903 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.659926 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.659951 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.659976 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.659995 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.660011 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.660030 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.660046 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.660066 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.660082 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.660101 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.660117 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.660136 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.660151 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.660171 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.660187 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.660206 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.660225 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.660246 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.660262 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.660279 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.660298 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.660334 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.660364 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.660485 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.660506 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.660528 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.660545 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.660565 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.660582 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.660599 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.660616 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.660633 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.660995 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.661017 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.661037 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.661054 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.661073 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.661111 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.661140 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.661548 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.661573 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.661599 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.661617 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.661635 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.661670 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.661748 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.661771 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.661791 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.661812 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.661832 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.661854 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.661871 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.661889 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.661906 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.661924 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.661959 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.661978 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.661998 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.662016 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.662032 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.662050 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.662069 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.662088 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.662113 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.662138 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.662164 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.662189 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.662211 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.662232 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.662260 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.662287 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.662315 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.662346 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.662368 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.662390 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.662414 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.662434 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.662452 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.662485 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.662516 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.662535 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.662559 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.662578 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.662598 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.662618 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.664928 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.664970 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.664990 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.665009 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.665028 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.665049 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.665067 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.665124 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.665162 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.665189 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.665224 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.665247 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.658851 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.672606 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.672725 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.672756 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.672879 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.672909 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.672932 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.672956 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.672976 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.673003 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.673072 4860 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.673086 4860 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.673096 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.673129 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.673143 4860 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.673154 4860 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.673164 4860 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.673176 4860 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.673189 4860 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.658947 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.659063 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.659238 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.675868 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.672554 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.676203 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.676319 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.676370 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.676406 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.676597 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.676691 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.676828 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.676950 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: E1211 08:11:26.677086 4860 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 11 08:11:26 crc kubenswrapper[4860]: E1211 08:11:26.677154 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-11 08:11:27.177131748 +0000 UTC m=+19.905650943 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.677209 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.677674 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.677887 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.677945 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.677982 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.678046 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: E1211 08:11:26.678108 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:11:27.178083579 +0000 UTC m=+19.906602834 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.678285 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.678383 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.678452 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.678456 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.659311 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.659468 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.659485 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.659620 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.659744 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.659839 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.659984 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.660028 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.660200 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.660367 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.660500 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.660626 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.660759 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.660868 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.660805 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.661076 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.661251 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.661357 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.661466 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.662529 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.662709 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.662802 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.662807 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.663077 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.663203 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.663251 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.663540 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.663572 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.663592 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.663698 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.664050 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.664131 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.664256 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.664428 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.678847 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.664516 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.664568 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.664839 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.665115 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.665566 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.665084 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.666010 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.666067 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.666150 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.666562 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.666722 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.666827 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.666867 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.666872 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.666974 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.667635 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.668072 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.668266 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.668420 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.669874 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.671104 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.671418 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.671454 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.671505 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.671709 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.671857 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.672032 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.672044 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.672176 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.672329 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.672412 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.673178 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.673241 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.673282 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.673471 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.673806 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.673892 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.674096 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.674201 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.674147 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.674524 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.674667 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.674710 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.675073 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.675280 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.679144 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.679468 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.679685 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.679724 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.679788 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.680086 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.680140 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.680596 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.680792 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.680864 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.680880 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.681099 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.681187 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.681219 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.681269 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.681281 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.681506 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.681524 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.681593 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.681449 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.681727 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.682613 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.682851 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.659271 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.684166 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.684221 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.684252 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.684260 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.684350 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.684417 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.684441 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.684499 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.684572 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.684599 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.684694 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.684745 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.685170 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.685263 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.685340 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.685386 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.685493 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.685684 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.685865 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.686145 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.686372 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.686663 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.688182 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.688409 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.688845 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: E1211 08:11:26.689188 4860 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 11 08:11:26 crc kubenswrapper[4860]: E1211 08:11:26.689211 4860 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 11 08:11:26 crc kubenswrapper[4860]: E1211 08:11:26.689228 4860 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 08:11:26 crc kubenswrapper[4860]: E1211 08:11:26.689288 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-11 08:11:27.189266249 +0000 UTC m=+19.917785524 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.690014 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.690034 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.690426 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.690458 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.690748 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.690783 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.690799 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.691132 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.691219 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.691354 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.691677 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.691690 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.691858 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.691893 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: E1211 08:11:26.691959 4860 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 11 08:11:26 crc kubenswrapper[4860]: E1211 08:11:26.691994 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-11 08:11:27.191983947 +0000 UTC m=+19.920503002 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.692011 4860 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.693223 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.693506 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.693600 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 11 08:11:26 crc kubenswrapper[4860]: E1211 08:11:26.693704 4860 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 11 08:11:26 crc kubenswrapper[4860]: E1211 08:11:26.693725 4860 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 11 08:11:26 crc kubenswrapper[4860]: E1211 08:11:26.693748 4860 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.693792 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: E1211 08:11:26.693800 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-11 08:11:27.193784445 +0000 UTC m=+19.922303720 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.693829 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.693928 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.693996 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.694103 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.694298 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.694353 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.694436 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.694877 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.695190 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.695453 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.695518 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.695777 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.696137 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.696261 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.696509 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.697599 4860 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="2090cab02ad74ddfd6bb57ff50eba35d18f56b0cde1c230fe011c01994b693bc" exitCode=255 Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.697740 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"2090cab02ad74ddfd6bb57ff50eba35d18f56b0cde1c230fe011c01994b693bc"} Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.699499 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.699809 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.701737 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.704891 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.709487 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.709615 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.717103 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.717716 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.717839 4860 scope.go:117] "RemoveContainer" containerID="2090cab02ad74ddfd6bb57ff50eba35d18f56b0cde1c230fe011c01994b693bc" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.720027 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.728249 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.731053 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.736304 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.739313 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.740391 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.747941 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.749963 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.759882 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.773550 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.773713 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.773943 4860 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.774071 4860 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.774146 4860 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.774218 4860 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.774282 4860 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.774342 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.774401 4860 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.774482 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.774552 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.774617 4860 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.773968 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.774708 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.774746 4860 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.774758 4860 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.774768 4860 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.774778 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.774787 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.774795 4860 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.774804 4860 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.774813 4860 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.774821 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.774831 4860 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.774840 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.774849 4860 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.774858 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.774866 4860 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.774875 4860 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.774911 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.774925 4860 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.774936 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.774948 4860 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.774958 4860 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.774966 4860 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.774975 4860 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.774984 4860 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.774993 4860 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.775002 4860 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.775012 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.775021 4860 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.775030 4860 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.775039 4860 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.775049 4860 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.775058 4860 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.775070 4860 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.775080 4860 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.775098 4860 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.775107 4860 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.775116 4860 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.773625 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.775125 4860 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.775263 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.775276 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.775286 4860 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.775299 4860 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.775313 4860 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.775325 4860 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.775333 4860 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.775343 4860 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.775352 4860 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.775360 4860 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.775369 4860 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.775378 4860 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.775387 4860 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.775395 4860 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.775403 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.775412 4860 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.775420 4860 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.775429 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.775437 4860 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.775447 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.775456 4860 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.775464 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.775474 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.775485 4860 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.775494 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.775502 4860 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.775511 4860 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.775519 4860 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.775528 4860 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.775537 4860 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.775548 4860 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.775561 4860 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.775572 4860 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.775580 4860 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.775590 4860 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.775601 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.775609 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.775619 4860 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.775630 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.775653 4860 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.775665 4860 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.775675 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.775685 4860 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.775694 4860 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.775702 4860 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.775711 4860 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.775720 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.775728 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.775737 4860 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.775745 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.775755 4860 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.775764 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.775773 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.775782 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.775791 4860 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.775800 4860 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.775808 4860 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.775817 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.775825 4860 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.775834 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.775842 4860 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.775851 4860 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.775859 4860 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.775869 4860 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.775877 4860 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.775885 4860 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.775894 4860 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.775904 4860 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.775912 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.775923 4860 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.775935 4860 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.775944 4860 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.775954 4860 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.775963 4860 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.775971 4860 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.775979 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.775988 4860 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.775996 4860 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.776005 4860 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.776015 4860 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.776024 4860 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.776034 4860 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.776044 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.776055 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.776065 4860 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.776074 4860 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.776083 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.776092 4860 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.776100 4860 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.776108 4860 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.776116 4860 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.776124 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.776133 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.776141 4860 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.776150 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.776159 4860 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.776167 4860 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.776175 4860 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.776183 4860 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.776191 4860 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.776199 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.776209 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.776218 4860 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.776226 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.776235 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.776244 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.776253 4860 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.776264 4860 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.776274 4860 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.776282 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.776292 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.776301 4860 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.776310 4860 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.776319 4860 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.776328 4860 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.776337 4860 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.776348 4860 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.776357 4860 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.776368 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.776378 4860 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.776388 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.776405 4860 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.776414 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.776422 4860 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.776431 4860 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.776440 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.776448 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.776457 4860 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.776466 4860 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.776474 4860 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.776482 4860 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.776493 4860 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.776502 4860 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.776510 4860 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.776518 4860 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.776526 4860 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.776535 4860 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.784152 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.793710 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.806081 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.821318 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.836546 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ec3758-85f1-4a22-a4ec-c805472607a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e831db1a15976d3967d56479689572c26f2599a766baa0402897a3008428c56f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a498aa90f914e1f0370df01dc5ef174df113a2a16d86b2e3f24b19a1b99b4c0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8caebd2f4dfcbb89c597029a07c335171889761d2d200206987537fb0a228db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2090cab02ad74ddfd6bb57ff50eba35d18f56b0cde1c230fe011c01994b693bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2090cab02ad74ddfd6bb57ff50eba35d18f56b0cde1c230fe011c01994b693bc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 08:11:21.171011 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 08:11:21.177674 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-644352688/tls.crt::/tmp/serving-cert-644352688/tls.key\\\\\\\"\\\\nI1211 08:11:26.508606 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 08:11:26.512741 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 08:11:26.512778 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 08:11:26.512839 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 08:11:26.512858 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 08:11:26.521684 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1211 08:11:26.521700 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1211 08:11:26.521722 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:11:26.521732 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:11:26.521738 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 08:11:26.521746 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 08:11:26.521749 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 08:11:26.521753 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1211 08:11:26.524999 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2274fb298ca7d2b37b060b4dbf9b166d39ba77a108b08edeeb4d67196e72cb3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf0aff8a6ab816b6617b1feeca9146d41bb5e8808bf1db96f8d493500c135d85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf0aff8a6ab816b6617b1feeca9146d41bb5e8808bf1db96f8d493500c135d85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.848096 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.929068 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.939828 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 11 08:11:26 crc kubenswrapper[4860]: W1211 08:11:26.940720 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-bfd5cca58e5048557c1c019618caa033951895fb4da0befe935f8e0f27777dd4 WatchSource:0}: Error finding container bfd5cca58e5048557c1c019618caa033951895fb4da0befe935f8e0f27777dd4: Status 404 returned error can't find the container with id bfd5cca58e5048557c1c019618caa033951895fb4da0befe935f8e0f27777dd4 Dec 11 08:11:26 crc kubenswrapper[4860]: I1211 08:11:26.942529 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 11 08:11:26 crc kubenswrapper[4860]: W1211 08:11:26.953231 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-5e1342102c245383831a701f17298edb1fbbc294f3551ab4c7a875476f6de13e WatchSource:0}: Error finding container 5e1342102c245383831a701f17298edb1fbbc294f3551ab4c7a875476f6de13e: Status 404 returned error can't find the container with id 5e1342102c245383831a701f17298edb1fbbc294f3551ab4c7a875476f6de13e Dec 11 08:11:26 crc kubenswrapper[4860]: W1211 08:11:26.966590 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-4f7f60dcd6e08a16382278485d84903d84731e697c69761f036fb45709af6934 WatchSource:0}: Error finding container 4f7f60dcd6e08a16382278485d84903d84731e697c69761f036fb45709af6934: Status 404 returned error can't find the container with id 4f7f60dcd6e08a16382278485d84903d84731e697c69761f036fb45709af6934 Dec 11 08:11:27 crc kubenswrapper[4860]: I1211 08:11:27.180121 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:11:27 crc kubenswrapper[4860]: I1211 08:11:27.180202 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:11:27 crc kubenswrapper[4860]: E1211 08:11:27.180301 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:11:28.180278907 +0000 UTC m=+20.908797962 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:11:27 crc kubenswrapper[4860]: E1211 08:11:27.180380 4860 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 11 08:11:27 crc kubenswrapper[4860]: E1211 08:11:27.180467 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-11 08:11:28.180445243 +0000 UTC m=+20.908964358 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 11 08:11:27 crc kubenswrapper[4860]: I1211 08:11:27.280764 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:11:27 crc kubenswrapper[4860]: I1211 08:11:27.280805 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:11:27 crc kubenswrapper[4860]: I1211 08:11:27.280831 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:11:27 crc kubenswrapper[4860]: E1211 08:11:27.280933 4860 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 11 08:11:27 crc kubenswrapper[4860]: E1211 08:11:27.280949 4860 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 11 08:11:27 crc kubenswrapper[4860]: E1211 08:11:27.280959 4860 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 08:11:27 crc kubenswrapper[4860]: E1211 08:11:27.280969 4860 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 11 08:11:27 crc kubenswrapper[4860]: E1211 08:11:27.281012 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-11 08:11:28.280997724 +0000 UTC m=+21.009516779 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 08:11:27 crc kubenswrapper[4860]: E1211 08:11:27.281072 4860 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 11 08:11:27 crc kubenswrapper[4860]: E1211 08:11:27.281101 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-11 08:11:28.281071547 +0000 UTC m=+21.009590642 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 11 08:11:27 crc kubenswrapper[4860]: E1211 08:11:27.281105 4860 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 11 08:11:27 crc kubenswrapper[4860]: E1211 08:11:27.281123 4860 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 08:11:27 crc kubenswrapper[4860]: E1211 08:11:27.281176 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-11 08:11:28.28115877 +0000 UTC m=+21.009677825 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 08:11:27 crc kubenswrapper[4860]: I1211 08:11:27.284788 4860 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-12-11 08:06:26 +0000 UTC, rotation deadline is 2026-09-26 21:38:06.224762684 +0000 UTC Dec 11 08:11:27 crc kubenswrapper[4860]: I1211 08:11:27.284857 4860 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 6949h26m38.939907995s for next certificate rotation Dec 11 08:11:27 crc kubenswrapper[4860]: I1211 08:11:27.295454 4860 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 11 08:11:27 crc kubenswrapper[4860]: I1211 08:11:27.426336 4860 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Dec 11 08:11:27 crc kubenswrapper[4860]: W1211 08:11:27.426632 4860 reflector.go:484] object-"openshift-network-operator"/"openshift-service-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-network-operator"/"openshift-service-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Dec 11 08:11:27 crc kubenswrapper[4860]: W1211 08:11:27.426703 4860 reflector.go:484] object-"openshift-network-node-identity"/"openshift-service-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-network-node-identity"/"openshift-service-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Dec 11 08:11:27 crc kubenswrapper[4860]: W1211 08:11:27.426744 4860 reflector.go:484] object-"openshift-network-node-identity"/"ovnkube-identity-cm": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-network-node-identity"/"ovnkube-identity-cm": Unexpected watch close - watch lasted less than a second and no items received Dec 11 08:11:27 crc kubenswrapper[4860]: W1211 08:11:27.426737 4860 reflector.go:484] object-"openshift-network-node-identity"/"kube-root-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-network-node-identity"/"kube-root-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Dec 11 08:11:27 crc kubenswrapper[4860]: E1211 08:11:27.426760 4860 controller.go:145] "Failed to ensure lease exists, will retry" err="Post \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases?timeout=10s\": read tcp 38.102.83.169:48920->38.102.83.169:6443: use of closed network connection" interval="6.4s" Dec 11 08:11:27 crc kubenswrapper[4860]: W1211 08:11:27.426807 4860 reflector.go:484] object-"openshift-network-operator"/"kube-root-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-network-operator"/"kube-root-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Dec 11 08:11:27 crc kubenswrapper[4860]: W1211 08:11:27.426833 4860 reflector.go:484] k8s.io/client-go/informers/factory.go:160: watch of *v1.RuntimeClass ended with: very short watch: k8s.io/client-go/informers/factory.go:160: Unexpected watch close - watch lasted less than a second and no items received Dec 11 08:11:27 crc kubenswrapper[4860]: W1211 08:11:27.426840 4860 reflector.go:484] object-"openshift-network-node-identity"/"env-overrides": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-network-node-identity"/"env-overrides": Unexpected watch close - watch lasted less than a second and no items received Dec 11 08:11:27 crc kubenswrapper[4860]: W1211 08:11:27.426633 4860 reflector.go:484] object-"openshift-network-operator"/"metrics-tls": watch of *v1.Secret ended with: very short watch: object-"openshift-network-operator"/"metrics-tls": Unexpected watch close - watch lasted less than a second and no items received Dec 11 08:11:27 crc kubenswrapper[4860]: W1211 08:11:27.426866 4860 reflector.go:484] object-"openshift-network-operator"/"iptables-alerter-script": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-network-operator"/"iptables-alerter-script": Unexpected watch close - watch lasted less than a second and no items received Dec 11 08:11:27 crc kubenswrapper[4860]: W1211 08:11:27.426869 4860 reflector.go:484] object-"openshift-network-node-identity"/"network-node-identity-cert": watch of *v1.Secret ended with: very short watch: object-"openshift-network-node-identity"/"network-node-identity-cert": Unexpected watch close - watch lasted less than a second and no items received Dec 11 08:11:27 crc kubenswrapper[4860]: E1211 08:11:27.426852 4860 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/events\": read tcp 38.102.83.169:48920->38.102.83.169:6443: use of closed network connection" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.18801aef20b68255 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 08:11:08.277572181 +0000 UTC m=+1.006091236,LastTimestamp:2025-12-11 08:11:08.277572181 +0000 UTC m=+1.006091236,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 08:11:27 crc kubenswrapper[4860]: I1211 08:11:27.541460 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Dec 11 08:11:27 crc kubenswrapper[4860]: I1211 08:11:27.555789 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:27Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:27 crc kubenswrapper[4860]: I1211 08:11:27.558148 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Dec 11 08:11:27 crc kubenswrapper[4860]: I1211 08:11:27.565837 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Dec 11 08:11:27 crc kubenswrapper[4860]: I1211 08:11:27.574871 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:27Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:27 crc kubenswrapper[4860]: I1211 08:11:27.581483 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Dec 11 08:11:27 crc kubenswrapper[4860]: I1211 08:11:27.581975 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Dec 11 08:11:27 crc kubenswrapper[4860]: I1211 08:11:27.583379 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Dec 11 08:11:27 crc kubenswrapper[4860]: I1211 08:11:27.584085 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Dec 11 08:11:27 crc kubenswrapper[4860]: I1211 08:11:27.585121 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Dec 11 08:11:27 crc kubenswrapper[4860]: I1211 08:11:27.585670 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Dec 11 08:11:27 crc kubenswrapper[4860]: I1211 08:11:27.586225 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Dec 11 08:11:27 crc kubenswrapper[4860]: I1211 08:11:27.587191 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Dec 11 08:11:27 crc kubenswrapper[4860]: I1211 08:11:27.587793 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Dec 11 08:11:27 crc kubenswrapper[4860]: I1211 08:11:27.588667 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Dec 11 08:11:27 crc kubenswrapper[4860]: I1211 08:11:27.589191 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Dec 11 08:11:27 crc kubenswrapper[4860]: I1211 08:11:27.590291 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Dec 11 08:11:27 crc kubenswrapper[4860]: I1211 08:11:27.590921 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Dec 11 08:11:27 crc kubenswrapper[4860]: I1211 08:11:27.591432 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Dec 11 08:11:27 crc kubenswrapper[4860]: I1211 08:11:27.592414 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:27Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:27 crc kubenswrapper[4860]: I1211 08:11:27.592802 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Dec 11 08:11:27 crc kubenswrapper[4860]: I1211 08:11:27.593285 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Dec 11 08:11:27 crc kubenswrapper[4860]: I1211 08:11:27.594196 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Dec 11 08:11:27 crc kubenswrapper[4860]: I1211 08:11:27.594560 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Dec 11 08:11:27 crc kubenswrapper[4860]: I1211 08:11:27.595210 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Dec 11 08:11:27 crc kubenswrapper[4860]: I1211 08:11:27.596883 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Dec 11 08:11:27 crc kubenswrapper[4860]: I1211 08:11:27.597316 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Dec 11 08:11:27 crc kubenswrapper[4860]: I1211 08:11:27.598241 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Dec 11 08:11:27 crc kubenswrapper[4860]: I1211 08:11:27.598719 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Dec 11 08:11:27 crc kubenswrapper[4860]: I1211 08:11:27.599666 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Dec 11 08:11:27 crc kubenswrapper[4860]: I1211 08:11:27.600080 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Dec 11 08:11:27 crc kubenswrapper[4860]: I1211 08:11:27.600634 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Dec 11 08:11:27 crc kubenswrapper[4860]: I1211 08:11:27.601620 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Dec 11 08:11:27 crc kubenswrapper[4860]: I1211 08:11:27.602085 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Dec 11 08:11:27 crc kubenswrapper[4860]: I1211 08:11:27.603023 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Dec 11 08:11:27 crc kubenswrapper[4860]: I1211 08:11:27.603459 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Dec 11 08:11:27 crc kubenswrapper[4860]: I1211 08:11:27.604355 4860 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Dec 11 08:11:27 crc kubenswrapper[4860]: I1211 08:11:27.604456 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Dec 11 08:11:27 crc kubenswrapper[4860]: I1211 08:11:27.606270 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Dec 11 08:11:27 crc kubenswrapper[4860]: I1211 08:11:27.607191 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Dec 11 08:11:27 crc kubenswrapper[4860]: I1211 08:11:27.607741 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Dec 11 08:11:27 crc kubenswrapper[4860]: I1211 08:11:27.609321 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:27Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:27 crc kubenswrapper[4860]: I1211 08:11:27.609675 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Dec 11 08:11:27 crc kubenswrapper[4860]: I1211 08:11:27.610327 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Dec 11 08:11:27 crc kubenswrapper[4860]: I1211 08:11:27.613201 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Dec 11 08:11:27 crc kubenswrapper[4860]: I1211 08:11:27.613811 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Dec 11 08:11:27 crc kubenswrapper[4860]: I1211 08:11:27.614844 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Dec 11 08:11:27 crc kubenswrapper[4860]: I1211 08:11:27.615427 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Dec 11 08:11:27 crc kubenswrapper[4860]: I1211 08:11:27.616381 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Dec 11 08:11:27 crc kubenswrapper[4860]: I1211 08:11:27.617118 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Dec 11 08:11:27 crc kubenswrapper[4860]: I1211 08:11:27.618882 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Dec 11 08:11:27 crc kubenswrapper[4860]: I1211 08:11:27.619529 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Dec 11 08:11:27 crc kubenswrapper[4860]: I1211 08:11:27.620497 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Dec 11 08:11:27 crc kubenswrapper[4860]: I1211 08:11:27.621112 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Dec 11 08:11:27 crc kubenswrapper[4860]: I1211 08:11:27.622267 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Dec 11 08:11:27 crc kubenswrapper[4860]: I1211 08:11:27.622734 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Dec 11 08:11:27 crc kubenswrapper[4860]: I1211 08:11:27.623832 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Dec 11 08:11:27 crc kubenswrapper[4860]: I1211 08:11:27.624451 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Dec 11 08:11:27 crc kubenswrapper[4860]: I1211 08:11:27.625871 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Dec 11 08:11:27 crc kubenswrapper[4860]: I1211 08:11:27.626562 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ec3758-85f1-4a22-a4ec-c805472607a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e831db1a15976d3967d56479689572c26f2599a766baa0402897a3008428c56f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a498aa90f914e1f0370df01dc5ef174df113a2a16d86b2e3f24b19a1b99b4c0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8caebd2f4dfcbb89c597029a07c335171889761d2d200206987537fb0a228db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2090cab02ad74ddfd6bb57ff50eba35d18f56b0cde1c230fe011c01994b693bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2090cab02ad74ddfd6bb57ff50eba35d18f56b0cde1c230fe011c01994b693bc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 08:11:21.171011 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 08:11:21.177674 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-644352688/tls.crt::/tmp/serving-cert-644352688/tls.key\\\\\\\"\\\\nI1211 08:11:26.508606 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 08:11:26.512741 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 08:11:26.512778 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 08:11:26.512839 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 08:11:26.512858 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 08:11:26.521684 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1211 08:11:26.521700 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1211 08:11:26.521722 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:11:26.521732 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:11:26.521738 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 08:11:26.521746 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 08:11:26.521749 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 08:11:26.521753 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1211 08:11:26.524999 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2274fb298ca7d2b37b060b4dbf9b166d39ba77a108b08edeeb4d67196e72cb3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf0aff8a6ab816b6617b1feeca9146d41bb5e8808bf1db96f8d493500c135d85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf0aff8a6ab816b6617b1feeca9146d41bb5e8808bf1db96f8d493500c135d85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:27Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:27 crc kubenswrapper[4860]: I1211 08:11:27.628743 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Dec 11 08:11:27 crc kubenswrapper[4860]: I1211 08:11:27.630060 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Dec 11 08:11:27 crc kubenswrapper[4860]: I1211 08:11:27.640609 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:27Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:27 crc kubenswrapper[4860]: I1211 08:11:27.659024 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:27Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:27 crc kubenswrapper[4860]: I1211 08:11:27.674180 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:27Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:27 crc kubenswrapper[4860]: I1211 08:11:27.690845 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:27Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:27 crc kubenswrapper[4860]: I1211 08:11:27.700268 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"5e1342102c245383831a701f17298edb1fbbc294f3551ab4c7a875476f6de13e"} Dec 11 08:11:27 crc kubenswrapper[4860]: I1211 08:11:27.701265 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"2d8c2f28628dc2f1ffbdd98a55670e22cf6cdc8cf2c6442e9d1f8e5466b2f12e"} Dec 11 08:11:27 crc kubenswrapper[4860]: I1211 08:11:27.701283 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"074ea63beac52bd196507ee7cfb1c3d7fc2d37d5ec6faa4d47dde62b922e1c43"} Dec 11 08:11:27 crc kubenswrapper[4860]: I1211 08:11:27.701291 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"bfd5cca58e5048557c1c019618caa033951895fb4da0befe935f8e0f27777dd4"} Dec 11 08:11:27 crc kubenswrapper[4860]: I1211 08:11:27.706321 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 11 08:11:27 crc kubenswrapper[4860]: I1211 08:11:27.707882 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"34e91f615b76445ff3241a304bb2964fde59de7e8a8dfff6841b2e628a31e86d"} Dec 11 08:11:27 crc kubenswrapper[4860]: I1211 08:11:27.708055 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 08:11:27 crc kubenswrapper[4860]: I1211 08:11:27.708746 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"a41f35db087418d1be16953f10377042abd3f4488e5ac1bc29ca5dcfc00a37e6"} Dec 11 08:11:27 crc kubenswrapper[4860]: I1211 08:11:27.708772 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"4f7f60dcd6e08a16382278485d84903d84731e697c69761f036fb45709af6934"} Dec 11 08:11:27 crc kubenswrapper[4860]: I1211 08:11:27.725240 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9872be1e-b995-4a00-904b-875dff96ad56\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f74a32a8fefea704cc73ffd1114083dab68aa44dc33f8d9d432d2d24216546c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbe00b46063d3968d59c393b87178e135f982abb7d0f32fa798b212d4b83df6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eedc9988e1a4a3c8cae730dc4a352103a6d665535b825e157d0f87c331b137c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2137fee4b7abc9765f9517d7b3b6ea796601a31d03151bf167d17136ced8b25a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f683e48f3577a68862eb6302383538376e41f49b20b1e0ea3f1b3b143f24872a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5de794187456b3576b969069a3b839d75c7918418a97fa35dff8bb40bb76e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5de794187456b3576b969069a3b839d75c7918418a97fa35dff8bb40bb76e39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08ce42b974fba00c409c0af971f0de351e50f18b2f06f8e722150da3239bb908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08ce42b974fba00c409c0af971f0de351e50f18b2f06f8e722150da3239bb908\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cfac6be03e3cfa0b0f3450b48b67957f2d5172b9c65e27f14c2a260bda5d423f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cfac6be03e3cfa0b0f3450b48b67957f2d5172b9c65e27f14c2a260bda5d423f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:27Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:27 crc kubenswrapper[4860]: I1211 08:11:27.741241 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ec3758-85f1-4a22-a4ec-c805472607a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e831db1a15976d3967d56479689572c26f2599a766baa0402897a3008428c56f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a498aa90f914e1f0370df01dc5ef174df113a2a16d86b2e3f24b19a1b99b4c0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8caebd2f4dfcbb89c597029a07c335171889761d2d200206987537fb0a228db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2090cab02ad74ddfd6bb57ff50eba35d18f56b0cde1c230fe011c01994b693bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2090cab02ad74ddfd6bb57ff50eba35d18f56b0cde1c230fe011c01994b693bc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 08:11:21.171011 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 08:11:21.177674 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-644352688/tls.crt::/tmp/serving-cert-644352688/tls.key\\\\\\\"\\\\nI1211 08:11:26.508606 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 08:11:26.512741 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 08:11:26.512778 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 08:11:26.512839 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 08:11:26.512858 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 08:11:26.521684 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1211 08:11:26.521700 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1211 08:11:26.521722 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:11:26.521732 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:11:26.521738 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 08:11:26.521746 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 08:11:26.521749 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 08:11:26.521753 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1211 08:11:26.524999 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2274fb298ca7d2b37b060b4dbf9b166d39ba77a108b08edeeb4d67196e72cb3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf0aff8a6ab816b6617b1feeca9146d41bb5e8808bf1db96f8d493500c135d85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf0aff8a6ab816b6617b1feeca9146d41bb5e8808bf1db96f8d493500c135d85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:27Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:27 crc kubenswrapper[4860]: I1211 08:11:27.754356 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:27Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:27 crc kubenswrapper[4860]: I1211 08:11:27.767545 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:27Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:27 crc kubenswrapper[4860]: I1211 08:11:27.778304 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:27Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:27 crc kubenswrapper[4860]: I1211 08:11:27.792194 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:27Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:27 crc kubenswrapper[4860]: I1211 08:11:27.807880 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:27Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:27 crc kubenswrapper[4860]: I1211 08:11:27.819502 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:27Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:27 crc kubenswrapper[4860]: I1211 08:11:27.831915 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:27Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:27 crc kubenswrapper[4860]: I1211 08:11:27.854588 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d8c2f28628dc2f1ffbdd98a55670e22cf6cdc8cf2c6442e9d1f8e5466b2f12e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://074ea63beac52bd196507ee7cfb1c3d7fc2d37d5ec6faa4d47dde62b922e1c43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:27Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:27 crc kubenswrapper[4860]: I1211 08:11:27.886965 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9872be1e-b995-4a00-904b-875dff96ad56\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f74a32a8fefea704cc73ffd1114083dab68aa44dc33f8d9d432d2d24216546c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbe00b46063d3968d59c393b87178e135f982abb7d0f32fa798b212d4b83df6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eedc9988e1a4a3c8cae730dc4a352103a6d665535b825e157d0f87c331b137c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2137fee4b7abc9765f9517d7b3b6ea796601a31d03151bf167d17136ced8b25a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f683e48f3577a68862eb6302383538376e41f49b20b1e0ea3f1b3b143f24872a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5de794187456b3576b969069a3b839d75c7918418a97fa35dff8bb40bb76e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5de794187456b3576b969069a3b839d75c7918418a97fa35dff8bb40bb76e39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08ce42b974fba00c409c0af971f0de351e50f18b2f06f8e722150da3239bb908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08ce42b974fba00c409c0af971f0de351e50f18b2f06f8e722150da3239bb908\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cfac6be03e3cfa0b0f3450b48b67957f2d5172b9c65e27f14c2a260bda5d423f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cfac6be03e3cfa0b0f3450b48b67957f2d5172b9c65e27f14c2a260bda5d423f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:27Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:27 crc kubenswrapper[4860]: I1211 08:11:27.912068 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ec3758-85f1-4a22-a4ec-c805472607a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e831db1a15976d3967d56479689572c26f2599a766baa0402897a3008428c56f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a498aa90f914e1f0370df01dc5ef174df113a2a16d86b2e3f24b19a1b99b4c0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8caebd2f4dfcbb89c597029a07c335171889761d2d200206987537fb0a228db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34e91f615b76445ff3241a304bb2964fde59de7e8a8dfff6841b2e628a31e86d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2090cab02ad74ddfd6bb57ff50eba35d18f56b0cde1c230fe011c01994b693bc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 08:11:21.171011 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 08:11:21.177674 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-644352688/tls.crt::/tmp/serving-cert-644352688/tls.key\\\\\\\"\\\\nI1211 08:11:26.508606 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 08:11:26.512741 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 08:11:26.512778 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 08:11:26.512839 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 08:11:26.512858 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 08:11:26.521684 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1211 08:11:26.521700 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1211 08:11:26.521722 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:11:26.521732 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:11:26.521738 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 08:11:26.521746 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 08:11:26.521749 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 08:11:26.521753 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1211 08:11:26.524999 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2274fb298ca7d2b37b060b4dbf9b166d39ba77a108b08edeeb4d67196e72cb3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf0aff8a6ab816b6617b1feeca9146d41bb5e8808bf1db96f8d493500c135d85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf0aff8a6ab816b6617b1feeca9146d41bb5e8808bf1db96f8d493500c135d85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:27Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:27 crc kubenswrapper[4860]: I1211 08:11:27.943541 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:27Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:27 crc kubenswrapper[4860]: I1211 08:11:27.961254 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a41f35db087418d1be16953f10377042abd3f4488e5ac1bc29ca5dcfc00a37e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:27Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:27 crc kubenswrapper[4860]: I1211 08:11:27.983926 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9872be1e-b995-4a00-904b-875dff96ad56\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f74a32a8fefea704cc73ffd1114083dab68aa44dc33f8d9d432d2d24216546c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbe00b46063d3968d59c393b87178e135f982abb7d0f32fa798b212d4b83df6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eedc9988e1a4a3c8cae730dc4a352103a6d665535b825e157d0f87c331b137c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2137fee4b7abc9765f9517d7b3b6ea796601a31d03151bf167d17136ced8b25a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f683e48f3577a68862eb6302383538376e41f49b20b1e0ea3f1b3b143f24872a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5de794187456b3576b969069a3b839d75c7918418a97fa35dff8bb40bb76e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5de794187456b3576b969069a3b839d75c7918418a97fa35dff8bb40bb76e39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08ce42b974fba00c409c0af971f0de351e50f18b2f06f8e722150da3239bb908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08ce42b974fba00c409c0af971f0de351e50f18b2f06f8e722150da3239bb908\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cfac6be03e3cfa0b0f3450b48b67957f2d5172b9c65e27f14c2a260bda5d423f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cfac6be03e3cfa0b0f3450b48b67957f2d5172b9c65e27f14c2a260bda5d423f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:27Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.039569 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ec3758-85f1-4a22-a4ec-c805472607a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e831db1a15976d3967d56479689572c26f2599a766baa0402897a3008428c56f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a498aa90f914e1f0370df01dc5ef174df113a2a16d86b2e3f24b19a1b99b4c0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8caebd2f4dfcbb89c597029a07c335171889761d2d200206987537fb0a228db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34e91f615b76445ff3241a304bb2964fde59de7e8a8dfff6841b2e628a31e86d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2090cab02ad74ddfd6bb57ff50eba35d18f56b0cde1c230fe011c01994b693bc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 08:11:21.171011 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 08:11:21.177674 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-644352688/tls.crt::/tmp/serving-cert-644352688/tls.key\\\\\\\"\\\\nI1211 08:11:26.508606 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 08:11:26.512741 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 08:11:26.512778 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 08:11:26.512839 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 08:11:26.512858 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 08:11:26.521684 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1211 08:11:26.521700 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1211 08:11:26.521722 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:11:26.521732 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:11:26.521738 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 08:11:26.521746 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 08:11:26.521749 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 08:11:26.521753 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1211 08:11:26.524999 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2274fb298ca7d2b37b060b4dbf9b166d39ba77a108b08edeeb4d67196e72cb3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf0aff8a6ab816b6617b1feeca9146d41bb5e8808bf1db96f8d493500c135d85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf0aff8a6ab816b6617b1feeca9146d41bb5e8808bf1db96f8d493500c135d85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:28Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.061540 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:28Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.080710 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a41f35db087418d1be16953f10377042abd3f4488e5ac1bc29ca5dcfc00a37e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:28Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.097060 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:28Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.116309 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:28Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.131543 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d8c2f28628dc2f1ffbdd98a55670e22cf6cdc8cf2c6442e9d1f8e5466b2f12e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://074ea63beac52bd196507ee7cfb1c3d7fc2d37d5ec6faa4d47dde62b922e1c43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:28Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.141962 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:28Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.186198 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.186317 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:11:28 crc kubenswrapper[4860]: E1211 08:11:28.186451 4860 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 11 08:11:28 crc kubenswrapper[4860]: E1211 08:11:28.186448 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:11:30.186410192 +0000 UTC m=+22.914929267 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:11:28 crc kubenswrapper[4860]: E1211 08:11:28.186525 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-11 08:11:30.186505976 +0000 UTC m=+22.915025091 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.282887 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-6xq7j"] Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.283263 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-6xq7j" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.285783 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.286034 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.286200 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.286683 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.286736 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.286785 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:11:28 crc kubenswrapper[4860]: E1211 08:11:28.286859 4860 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 11 08:11:28 crc kubenswrapper[4860]: E1211 08:11:28.286876 4860 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 11 08:11:28 crc kubenswrapper[4860]: E1211 08:11:28.286889 4860 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 08:11:28 crc kubenswrapper[4860]: E1211 08:11:28.286934 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-11 08:11:30.286918412 +0000 UTC m=+23.015437467 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 08:11:28 crc kubenswrapper[4860]: E1211 08:11:28.286948 4860 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 11 08:11:28 crc kubenswrapper[4860]: E1211 08:11:28.286972 4860 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 11 08:11:28 crc kubenswrapper[4860]: E1211 08:11:28.286990 4860 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 08:11:28 crc kubenswrapper[4860]: E1211 08:11:28.287033 4860 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 11 08:11:28 crc kubenswrapper[4860]: E1211 08:11:28.287050 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-11 08:11:30.287029416 +0000 UTC m=+23.015548501 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 08:11:28 crc kubenswrapper[4860]: E1211 08:11:28.287078 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-11 08:11:30.287066637 +0000 UTC m=+23.015585792 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.305655 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:28Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.321854 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a41f35db087418d1be16953f10377042abd3f4488e5ac1bc29ca5dcfc00a37e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:28Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.327832 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.353789 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9872be1e-b995-4a00-904b-875dff96ad56\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f74a32a8fefea704cc73ffd1114083dab68aa44dc33f8d9d432d2d24216546c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbe00b46063d3968d59c393b87178e135f982abb7d0f32fa798b212d4b83df6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eedc9988e1a4a3c8cae730dc4a352103a6d665535b825e157d0f87c331b137c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2137fee4b7abc9765f9517d7b3b6ea796601a31d03151bf167d17136ced8b25a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f683e48f3577a68862eb6302383538376e41f49b20b1e0ea3f1b3b143f24872a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5de794187456b3576b969069a3b839d75c7918418a97fa35dff8bb40bb76e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5de794187456b3576b969069a3b839d75c7918418a97fa35dff8bb40bb76e39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08ce42b974fba00c409c0af971f0de351e50f18b2f06f8e722150da3239bb908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08ce42b974fba00c409c0af971f0de351e50f18b2f06f8e722150da3239bb908\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cfac6be03e3cfa0b0f3450b48b67957f2d5172b9c65e27f14c2a260bda5d423f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cfac6be03e3cfa0b0f3450b48b67957f2d5172b9c65e27f14c2a260bda5d423f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:28Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.368979 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ec3758-85f1-4a22-a4ec-c805472607a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e831db1a15976d3967d56479689572c26f2599a766baa0402897a3008428c56f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a498aa90f914e1f0370df01dc5ef174df113a2a16d86b2e3f24b19a1b99b4c0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8caebd2f4dfcbb89c597029a07c335171889761d2d200206987537fb0a228db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34e91f615b76445ff3241a304bb2964fde59de7e8a8dfff6841b2e628a31e86d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2090cab02ad74ddfd6bb57ff50eba35d18f56b0cde1c230fe011c01994b693bc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 08:11:21.171011 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 08:11:21.177674 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-644352688/tls.crt::/tmp/serving-cert-644352688/tls.key\\\\\\\"\\\\nI1211 08:11:26.508606 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 08:11:26.512741 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 08:11:26.512778 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 08:11:26.512839 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 08:11:26.512858 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 08:11:26.521684 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1211 08:11:26.521700 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1211 08:11:26.521722 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:11:26.521732 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:11:26.521738 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 08:11:26.521746 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 08:11:26.521749 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 08:11:26.521753 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1211 08:11:26.524999 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2274fb298ca7d2b37b060b4dbf9b166d39ba77a108b08edeeb4d67196e72cb3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf0aff8a6ab816b6617b1feeca9146d41bb5e8808bf1db96f8d493500c135d85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf0aff8a6ab816b6617b1feeca9146d41bb5e8808bf1db96f8d493500c135d85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:28Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.381956 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:28Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.388158 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/6a75ce3c-5e1e-4b9f-bbde-7a3b0cce5eae-hosts-file\") pod \"node-resolver-6xq7j\" (UID: \"6a75ce3c-5e1e-4b9f-bbde-7a3b0cce5eae\") " pod="openshift-dns/node-resolver-6xq7j" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.388213 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5c6nb\" (UniqueName: \"kubernetes.io/projected/6a75ce3c-5e1e-4b9f-bbde-7a3b0cce5eae-kube-api-access-5c6nb\") pod \"node-resolver-6xq7j\" (UID: \"6a75ce3c-5e1e-4b9f-bbde-7a3b0cce5eae\") " pod="openshift-dns/node-resolver-6xq7j" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.391085 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.398882 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d8c2f28628dc2f1ffbdd98a55670e22cf6cdc8cf2c6442e9d1f8e5466b2f12e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://074ea63beac52bd196507ee7cfb1c3d7fc2d37d5ec6faa4d47dde62b922e1c43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:28Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.410950 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.414280 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:28Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.425691 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:28Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.436011 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6xq7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a75ce3c-5e1e-4b9f-bbde-7a3b0cce5eae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5c6nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6xq7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:28Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.489403 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5c6nb\" (UniqueName: \"kubernetes.io/projected/6a75ce3c-5e1e-4b9f-bbde-7a3b0cce5eae-kube-api-access-5c6nb\") pod \"node-resolver-6xq7j\" (UID: \"6a75ce3c-5e1e-4b9f-bbde-7a3b0cce5eae\") " pod="openshift-dns/node-resolver-6xq7j" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.489478 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/6a75ce3c-5e1e-4b9f-bbde-7a3b0cce5eae-hosts-file\") pod \"node-resolver-6xq7j\" (UID: \"6a75ce3c-5e1e-4b9f-bbde-7a3b0cce5eae\") " pod="openshift-dns/node-resolver-6xq7j" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.489551 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/6a75ce3c-5e1e-4b9f-bbde-7a3b0cce5eae-hosts-file\") pod \"node-resolver-6xq7j\" (UID: \"6a75ce3c-5e1e-4b9f-bbde-7a3b0cce5eae\") " pod="openshift-dns/node-resolver-6xq7j" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.513778 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5c6nb\" (UniqueName: \"kubernetes.io/projected/6a75ce3c-5e1e-4b9f-bbde-7a3b0cce5eae-kube-api-access-5c6nb\") pod \"node-resolver-6xq7j\" (UID: \"6a75ce3c-5e1e-4b9f-bbde-7a3b0cce5eae\") " pod="openshift-dns/node-resolver-6xq7j" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.537577 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.568138 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.578490 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.578537 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.578496 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:11:28 crc kubenswrapper[4860]: E1211 08:11:28.578631 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:11:28 crc kubenswrapper[4860]: E1211 08:11:28.578713 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:11:28 crc kubenswrapper[4860]: E1211 08:11:28.578769 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.596320 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-6xq7j" Dec 11 08:11:28 crc kubenswrapper[4860]: W1211 08:11:28.610891 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6a75ce3c_5e1e_4b9f_bbde_7a3b0cce5eae.slice/crio-dd4da3c244caeb9d4c825f0ce4b7a18027cecfcd6262e7962a7aa507afb491bf WatchSource:0}: Error finding container dd4da3c244caeb9d4c825f0ce4b7a18027cecfcd6262e7962a7aa507afb491bf: Status 404 returned error can't find the container with id dd4da3c244caeb9d4c825f0ce4b7a18027cecfcd6262e7962a7aa507afb491bf Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.715832 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-99qgp"] Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.716204 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-6xq7j" event={"ID":"6a75ce3c-5e1e-4b9f-bbde-7a3b0cce5eae","Type":"ContainerStarted","Data":"dd4da3c244caeb9d4c825f0ce4b7a18027cecfcd6262e7962a7aa507afb491bf"} Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.716352 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.716354 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-26nc6"] Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.717086 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-26nc6" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.719505 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-pznwl"] Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.719773 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-pznwl" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.722154 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.722347 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.723432 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.723810 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.738134 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.738428 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.738451 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.738229 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.738183 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.738456 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.738830 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.738924 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.753667 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:28Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.769690 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a41f35db087418d1be16953f10377042abd3f4488e5ac1bc29ca5dcfc00a37e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:28Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.777203 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.787391 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9872be1e-b995-4a00-904b-875dff96ad56\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f74a32a8fefea704cc73ffd1114083dab68aa44dc33f8d9d432d2d24216546c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbe00b46063d3968d59c393b87178e135f982abb7d0f32fa798b212d4b83df6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eedc9988e1a4a3c8cae730dc4a352103a6d665535b825e157d0f87c331b137c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2137fee4b7abc9765f9517d7b3b6ea796601a31d03151bf167d17136ced8b25a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f683e48f3577a68862eb6302383538376e41f49b20b1e0ea3f1b3b143f24872a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5de794187456b3576b969069a3b839d75c7918418a97fa35dff8bb40bb76e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5de794187456b3576b969069a3b839d75c7918418a97fa35dff8bb40bb76e39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08ce42b974fba00c409c0af971f0de351e50f18b2f06f8e722150da3239bb908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08ce42b974fba00c409c0af971f0de351e50f18b2f06f8e722150da3239bb908\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cfac6be03e3cfa0b0f3450b48b67957f2d5172b9c65e27f14c2a260bda5d423f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cfac6be03e3cfa0b0f3450b48b67957f2d5172b9c65e27f14c2a260bda5d423f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:28Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.791592 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/5310c1fc-66c6-40aa-b1bf-5a59a2410f9c-multus-daemon-config\") pod \"multus-pznwl\" (UID: \"5310c1fc-66c6-40aa-b1bf-5a59a2410f9c\") " pod="openshift-multus/multus-pznwl" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.791624 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/a1faa733-f408-41fd-b7aa-0cd86283976a-os-release\") pod \"multus-additional-cni-plugins-26nc6\" (UID: \"a1faa733-f408-41fd-b7aa-0cd86283976a\") " pod="openshift-multus/multus-additional-cni-plugins-26nc6" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.791661 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/a1faa733-f408-41fd-b7aa-0cd86283976a-system-cni-dir\") pod \"multus-additional-cni-plugins-26nc6\" (UID: \"a1faa733-f408-41fd-b7aa-0cd86283976a\") " pod="openshift-multus/multus-additional-cni-plugins-26nc6" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.791678 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/a1faa733-f408-41fd-b7aa-0cd86283976a-cni-binary-copy\") pod \"multus-additional-cni-plugins-26nc6\" (UID: \"a1faa733-f408-41fd-b7aa-0cd86283976a\") " pod="openshift-multus/multus-additional-cni-plugins-26nc6" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.791694 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31c30642-6e60-41b4-a477-0d802424e0aa-proxy-tls\") pod \"machine-config-daemon-99qgp\" (UID: \"31c30642-6e60-41b4-a477-0d802424e0aa\") " pod="openshift-machine-config-operator/machine-config-daemon-99qgp" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.791745 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/5310c1fc-66c6-40aa-b1bf-5a59a2410f9c-multus-cni-dir\") pod \"multus-pznwl\" (UID: \"5310c1fc-66c6-40aa-b1bf-5a59a2410f9c\") " pod="openshift-multus/multus-pznwl" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.791773 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5310c1fc-66c6-40aa-b1bf-5a59a2410f9c-etc-kubernetes\") pod \"multus-pznwl\" (UID: \"5310c1fc-66c6-40aa-b1bf-5a59a2410f9c\") " pod="openshift-multus/multus-pznwl" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.791810 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/a1faa733-f408-41fd-b7aa-0cd86283976a-tuning-conf-dir\") pod \"multus-additional-cni-plugins-26nc6\" (UID: \"a1faa733-f408-41fd-b7aa-0cd86283976a\") " pod="openshift-multus/multus-additional-cni-plugins-26nc6" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.791831 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/a1faa733-f408-41fd-b7aa-0cd86283976a-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-26nc6\" (UID: \"a1faa733-f408-41fd-b7aa-0cd86283976a\") " pod="openshift-multus/multus-additional-cni-plugins-26nc6" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.791853 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/5310c1fc-66c6-40aa-b1bf-5a59a2410f9c-host-var-lib-cni-bin\") pod \"multus-pznwl\" (UID: \"5310c1fc-66c6-40aa-b1bf-5a59a2410f9c\") " pod="openshift-multus/multus-pznwl" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.791871 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/5310c1fc-66c6-40aa-b1bf-5a59a2410f9c-cni-binary-copy\") pod \"multus-pznwl\" (UID: \"5310c1fc-66c6-40aa-b1bf-5a59a2410f9c\") " pod="openshift-multus/multus-pznwl" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.791934 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/5310c1fc-66c6-40aa-b1bf-5a59a2410f9c-host-var-lib-kubelet\") pod \"multus-pznwl\" (UID: \"5310c1fc-66c6-40aa-b1bf-5a59a2410f9c\") " pod="openshift-multus/multus-pznwl" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.791980 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/5310c1fc-66c6-40aa-b1bf-5a59a2410f9c-os-release\") pod \"multus-pznwl\" (UID: \"5310c1fc-66c6-40aa-b1bf-5a59a2410f9c\") " pod="openshift-multus/multus-pznwl" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.791996 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/5310c1fc-66c6-40aa-b1bf-5a59a2410f9c-host-run-k8s-cni-cncf-io\") pod \"multus-pznwl\" (UID: \"5310c1fc-66c6-40aa-b1bf-5a59a2410f9c\") " pod="openshift-multus/multus-pznwl" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.792013 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hkt48\" (UniqueName: \"kubernetes.io/projected/5310c1fc-66c6-40aa-b1bf-5a59a2410f9c-kube-api-access-hkt48\") pod \"multus-pznwl\" (UID: \"5310c1fc-66c6-40aa-b1bf-5a59a2410f9c\") " pod="openshift-multus/multus-pznwl" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.792029 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/5310c1fc-66c6-40aa-b1bf-5a59a2410f9c-system-cni-dir\") pod \"multus-pznwl\" (UID: \"5310c1fc-66c6-40aa-b1bf-5a59a2410f9c\") " pod="openshift-multus/multus-pznwl" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.792042 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/5310c1fc-66c6-40aa-b1bf-5a59a2410f9c-hostroot\") pod \"multus-pznwl\" (UID: \"5310c1fc-66c6-40aa-b1bf-5a59a2410f9c\") " pod="openshift-multus/multus-pznwl" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.792056 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/a1faa733-f408-41fd-b7aa-0cd86283976a-cnibin\") pod \"multus-additional-cni-plugins-26nc6\" (UID: \"a1faa733-f408-41fd-b7aa-0cd86283976a\") " pod="openshift-multus/multus-additional-cni-plugins-26nc6" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.792071 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9nf9r\" (UniqueName: \"kubernetes.io/projected/a1faa733-f408-41fd-b7aa-0cd86283976a-kube-api-access-9nf9r\") pod \"multus-additional-cni-plugins-26nc6\" (UID: \"a1faa733-f408-41fd-b7aa-0cd86283976a\") " pod="openshift-multus/multus-additional-cni-plugins-26nc6" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.792087 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/5310c1fc-66c6-40aa-b1bf-5a59a2410f9c-host-var-lib-cni-multus\") pod \"multus-pznwl\" (UID: \"5310c1fc-66c6-40aa-b1bf-5a59a2410f9c\") " pod="openshift-multus/multus-pznwl" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.792104 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/5310c1fc-66c6-40aa-b1bf-5a59a2410f9c-host-run-multus-certs\") pod \"multus-pznwl\" (UID: \"5310c1fc-66c6-40aa-b1bf-5a59a2410f9c\") " pod="openshift-multus/multus-pznwl" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.792127 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/31c30642-6e60-41b4-a477-0d802424e0aa-rootfs\") pod \"machine-config-daemon-99qgp\" (UID: \"31c30642-6e60-41b4-a477-0d802424e0aa\") " pod="openshift-machine-config-operator/machine-config-daemon-99qgp" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.792143 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31c30642-6e60-41b4-a477-0d802424e0aa-mcd-auth-proxy-config\") pod \"machine-config-daemon-99qgp\" (UID: \"31c30642-6e60-41b4-a477-0d802424e0aa\") " pod="openshift-machine-config-operator/machine-config-daemon-99qgp" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.792159 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/5310c1fc-66c6-40aa-b1bf-5a59a2410f9c-multus-conf-dir\") pod \"multus-pznwl\" (UID: \"5310c1fc-66c6-40aa-b1bf-5a59a2410f9c\") " pod="openshift-multus/multus-pznwl" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.792176 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7qc8q\" (UniqueName: \"kubernetes.io/projected/31c30642-6e60-41b4-a477-0d802424e0aa-kube-api-access-7qc8q\") pod \"machine-config-daemon-99qgp\" (UID: \"31c30642-6e60-41b4-a477-0d802424e0aa\") " pod="openshift-machine-config-operator/machine-config-daemon-99qgp" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.792206 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/5310c1fc-66c6-40aa-b1bf-5a59a2410f9c-multus-socket-dir-parent\") pod \"multus-pznwl\" (UID: \"5310c1fc-66c6-40aa-b1bf-5a59a2410f9c\") " pod="openshift-multus/multus-pznwl" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.792226 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/5310c1fc-66c6-40aa-b1bf-5a59a2410f9c-cnibin\") pod \"multus-pznwl\" (UID: \"5310c1fc-66c6-40aa-b1bf-5a59a2410f9c\") " pod="openshift-multus/multus-pznwl" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.792239 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/5310c1fc-66c6-40aa-b1bf-5a59a2410f9c-host-run-netns\") pod \"multus-pznwl\" (UID: \"5310c1fc-66c6-40aa-b1bf-5a59a2410f9c\") " pod="openshift-multus/multus-pznwl" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.800706 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ec3758-85f1-4a22-a4ec-c805472607a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e831db1a15976d3967d56479689572c26f2599a766baa0402897a3008428c56f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a498aa90f914e1f0370df01dc5ef174df113a2a16d86b2e3f24b19a1b99b4c0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8caebd2f4dfcbb89c597029a07c335171889761d2d200206987537fb0a228db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34e91f615b76445ff3241a304bb2964fde59de7e8a8dfff6841b2e628a31e86d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2090cab02ad74ddfd6bb57ff50eba35d18f56b0cde1c230fe011c01994b693bc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 08:11:21.171011 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 08:11:21.177674 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-644352688/tls.crt::/tmp/serving-cert-644352688/tls.key\\\\\\\"\\\\nI1211 08:11:26.508606 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 08:11:26.512741 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 08:11:26.512778 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 08:11:26.512839 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 08:11:26.512858 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 08:11:26.521684 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1211 08:11:26.521700 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1211 08:11:26.521722 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:11:26.521732 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:11:26.521738 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 08:11:26.521746 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 08:11:26.521749 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 08:11:26.521753 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1211 08:11:26.524999 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2274fb298ca7d2b37b060b4dbf9b166d39ba77a108b08edeeb4d67196e72cb3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf0aff8a6ab816b6617b1feeca9146d41bb5e8808bf1db96f8d493500c135d85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf0aff8a6ab816b6617b1feeca9146d41bb5e8808bf1db96f8d493500c135d85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:28Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.816123 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:28Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.826758 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d8c2f28628dc2f1ffbdd98a55670e22cf6cdc8cf2c6442e9d1f8e5466b2f12e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://074ea63beac52bd196507ee7cfb1c3d7fc2d37d5ec6faa4d47dde62b922e1c43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:28Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.837486 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:28Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.847803 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:28Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.856854 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6xq7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a75ce3c-5e1e-4b9f-bbde-7a3b0cce5eae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5c6nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6xq7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:28Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.869662 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31c30642-6e60-41b4-a477-0d802424e0aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-99qgp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:28Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.879831 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.885691 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-26nc6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1faa733-f408-41fd-b7aa-0cd86283976a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-26nc6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:28Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.893274 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7qc8q\" (UniqueName: \"kubernetes.io/projected/31c30642-6e60-41b4-a477-0d802424e0aa-kube-api-access-7qc8q\") pod \"machine-config-daemon-99qgp\" (UID: \"31c30642-6e60-41b4-a477-0d802424e0aa\") " pod="openshift-machine-config-operator/machine-config-daemon-99qgp" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.893560 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/5310c1fc-66c6-40aa-b1bf-5a59a2410f9c-multus-socket-dir-parent\") pod \"multus-pznwl\" (UID: \"5310c1fc-66c6-40aa-b1bf-5a59a2410f9c\") " pod="openshift-multus/multus-pznwl" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.893658 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/5310c1fc-66c6-40aa-b1bf-5a59a2410f9c-cnibin\") pod \"multus-pznwl\" (UID: \"5310c1fc-66c6-40aa-b1bf-5a59a2410f9c\") " pod="openshift-multus/multus-pznwl" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.893742 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/5310c1fc-66c6-40aa-b1bf-5a59a2410f9c-host-run-netns\") pod \"multus-pznwl\" (UID: \"5310c1fc-66c6-40aa-b1bf-5a59a2410f9c\") " pod="openshift-multus/multus-pznwl" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.893708 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/5310c1fc-66c6-40aa-b1bf-5a59a2410f9c-multus-socket-dir-parent\") pod \"multus-pznwl\" (UID: \"5310c1fc-66c6-40aa-b1bf-5a59a2410f9c\") " pod="openshift-multus/multus-pznwl" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.893838 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/5310c1fc-66c6-40aa-b1bf-5a59a2410f9c-cnibin\") pod \"multus-pznwl\" (UID: \"5310c1fc-66c6-40aa-b1bf-5a59a2410f9c\") " pod="openshift-multus/multus-pznwl" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.893893 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/5310c1fc-66c6-40aa-b1bf-5a59a2410f9c-host-run-netns\") pod \"multus-pznwl\" (UID: \"5310c1fc-66c6-40aa-b1bf-5a59a2410f9c\") " pod="openshift-multus/multus-pznwl" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.893945 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/5310c1fc-66c6-40aa-b1bf-5a59a2410f9c-multus-daemon-config\") pod \"multus-pznwl\" (UID: \"5310c1fc-66c6-40aa-b1bf-5a59a2410f9c\") " pod="openshift-multus/multus-pznwl" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.894024 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/a1faa733-f408-41fd-b7aa-0cd86283976a-os-release\") pod \"multus-additional-cni-plugins-26nc6\" (UID: \"a1faa733-f408-41fd-b7aa-0cd86283976a\") " pod="openshift-multus/multus-additional-cni-plugins-26nc6" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.894106 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/a1faa733-f408-41fd-b7aa-0cd86283976a-system-cni-dir\") pod \"multus-additional-cni-plugins-26nc6\" (UID: \"a1faa733-f408-41fd-b7aa-0cd86283976a\") " pod="openshift-multus/multus-additional-cni-plugins-26nc6" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.894183 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/a1faa733-f408-41fd-b7aa-0cd86283976a-os-release\") pod \"multus-additional-cni-plugins-26nc6\" (UID: \"a1faa733-f408-41fd-b7aa-0cd86283976a\") " pod="openshift-multus/multus-additional-cni-plugins-26nc6" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.894206 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/a1faa733-f408-41fd-b7aa-0cd86283976a-system-cni-dir\") pod \"multus-additional-cni-plugins-26nc6\" (UID: \"a1faa733-f408-41fd-b7aa-0cd86283976a\") " pod="openshift-multus/multus-additional-cni-plugins-26nc6" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.894174 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/a1faa733-f408-41fd-b7aa-0cd86283976a-cni-binary-copy\") pod \"multus-additional-cni-plugins-26nc6\" (UID: \"a1faa733-f408-41fd-b7aa-0cd86283976a\") " pod="openshift-multus/multus-additional-cni-plugins-26nc6" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.894262 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31c30642-6e60-41b4-a477-0d802424e0aa-proxy-tls\") pod \"machine-config-daemon-99qgp\" (UID: \"31c30642-6e60-41b4-a477-0d802424e0aa\") " pod="openshift-machine-config-operator/machine-config-daemon-99qgp" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.894314 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/5310c1fc-66c6-40aa-b1bf-5a59a2410f9c-multus-cni-dir\") pod \"multus-pznwl\" (UID: \"5310c1fc-66c6-40aa-b1bf-5a59a2410f9c\") " pod="openshift-multus/multus-pznwl" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.894339 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5310c1fc-66c6-40aa-b1bf-5a59a2410f9c-etc-kubernetes\") pod \"multus-pznwl\" (UID: \"5310c1fc-66c6-40aa-b1bf-5a59a2410f9c\") " pod="openshift-multus/multus-pznwl" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.894360 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/a1faa733-f408-41fd-b7aa-0cd86283976a-tuning-conf-dir\") pod \"multus-additional-cni-plugins-26nc6\" (UID: \"a1faa733-f408-41fd-b7aa-0cd86283976a\") " pod="openshift-multus/multus-additional-cni-plugins-26nc6" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.894383 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/a1faa733-f408-41fd-b7aa-0cd86283976a-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-26nc6\" (UID: \"a1faa733-f408-41fd-b7aa-0cd86283976a\") " pod="openshift-multus/multus-additional-cni-plugins-26nc6" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.894405 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/5310c1fc-66c6-40aa-b1bf-5a59a2410f9c-host-var-lib-cni-bin\") pod \"multus-pznwl\" (UID: \"5310c1fc-66c6-40aa-b1bf-5a59a2410f9c\") " pod="openshift-multus/multus-pznwl" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.894426 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/5310c1fc-66c6-40aa-b1bf-5a59a2410f9c-cni-binary-copy\") pod \"multus-pznwl\" (UID: \"5310c1fc-66c6-40aa-b1bf-5a59a2410f9c\") " pod="openshift-multus/multus-pznwl" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.894446 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/5310c1fc-66c6-40aa-b1bf-5a59a2410f9c-host-var-lib-kubelet\") pod \"multus-pznwl\" (UID: \"5310c1fc-66c6-40aa-b1bf-5a59a2410f9c\") " pod="openshift-multus/multus-pznwl" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.894473 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/5310c1fc-66c6-40aa-b1bf-5a59a2410f9c-os-release\") pod \"multus-pznwl\" (UID: \"5310c1fc-66c6-40aa-b1bf-5a59a2410f9c\") " pod="openshift-multus/multus-pznwl" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.894483 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/5310c1fc-66c6-40aa-b1bf-5a59a2410f9c-multus-daemon-config\") pod \"multus-pznwl\" (UID: \"5310c1fc-66c6-40aa-b1bf-5a59a2410f9c\") " pod="openshift-multus/multus-pznwl" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.894493 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/5310c1fc-66c6-40aa-b1bf-5a59a2410f9c-host-run-k8s-cni-cncf-io\") pod \"multus-pznwl\" (UID: \"5310c1fc-66c6-40aa-b1bf-5a59a2410f9c\") " pod="openshift-multus/multus-pznwl" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.894515 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hkt48\" (UniqueName: \"kubernetes.io/projected/5310c1fc-66c6-40aa-b1bf-5a59a2410f9c-kube-api-access-hkt48\") pod \"multus-pznwl\" (UID: \"5310c1fc-66c6-40aa-b1bf-5a59a2410f9c\") " pod="openshift-multus/multus-pznwl" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.894543 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/5310c1fc-66c6-40aa-b1bf-5a59a2410f9c-system-cni-dir\") pod \"multus-pznwl\" (UID: \"5310c1fc-66c6-40aa-b1bf-5a59a2410f9c\") " pod="openshift-multus/multus-pznwl" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.894565 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/5310c1fc-66c6-40aa-b1bf-5a59a2410f9c-hostroot\") pod \"multus-pznwl\" (UID: \"5310c1fc-66c6-40aa-b1bf-5a59a2410f9c\") " pod="openshift-multus/multus-pznwl" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.894587 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/a1faa733-f408-41fd-b7aa-0cd86283976a-cnibin\") pod \"multus-additional-cni-plugins-26nc6\" (UID: \"a1faa733-f408-41fd-b7aa-0cd86283976a\") " pod="openshift-multus/multus-additional-cni-plugins-26nc6" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.894611 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9nf9r\" (UniqueName: \"kubernetes.io/projected/a1faa733-f408-41fd-b7aa-0cd86283976a-kube-api-access-9nf9r\") pod \"multus-additional-cni-plugins-26nc6\" (UID: \"a1faa733-f408-41fd-b7aa-0cd86283976a\") " pod="openshift-multus/multus-additional-cni-plugins-26nc6" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.894634 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/5310c1fc-66c6-40aa-b1bf-5a59a2410f9c-host-var-lib-cni-multus\") pod \"multus-pznwl\" (UID: \"5310c1fc-66c6-40aa-b1bf-5a59a2410f9c\") " pod="openshift-multus/multus-pznwl" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.894675 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/5310c1fc-66c6-40aa-b1bf-5a59a2410f9c-host-run-multus-certs\") pod \"multus-pznwl\" (UID: \"5310c1fc-66c6-40aa-b1bf-5a59a2410f9c\") " pod="openshift-multus/multus-pznwl" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.894696 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/31c30642-6e60-41b4-a477-0d802424e0aa-rootfs\") pod \"machine-config-daemon-99qgp\" (UID: \"31c30642-6e60-41b4-a477-0d802424e0aa\") " pod="openshift-machine-config-operator/machine-config-daemon-99qgp" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.894718 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31c30642-6e60-41b4-a477-0d802424e0aa-mcd-auth-proxy-config\") pod \"machine-config-daemon-99qgp\" (UID: \"31c30642-6e60-41b4-a477-0d802424e0aa\") " pod="openshift-machine-config-operator/machine-config-daemon-99qgp" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.894742 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/5310c1fc-66c6-40aa-b1bf-5a59a2410f9c-multus-conf-dir\") pod \"multus-pznwl\" (UID: \"5310c1fc-66c6-40aa-b1bf-5a59a2410f9c\") " pod="openshift-multus/multus-pznwl" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.894809 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/5310c1fc-66c6-40aa-b1bf-5a59a2410f9c-multus-conf-dir\") pod \"multus-pznwl\" (UID: \"5310c1fc-66c6-40aa-b1bf-5a59a2410f9c\") " pod="openshift-multus/multus-pznwl" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.894845 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/5310c1fc-66c6-40aa-b1bf-5a59a2410f9c-host-var-lib-cni-bin\") pod \"multus-pznwl\" (UID: \"5310c1fc-66c6-40aa-b1bf-5a59a2410f9c\") " pod="openshift-multus/multus-pznwl" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.894952 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/a1faa733-f408-41fd-b7aa-0cd86283976a-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-26nc6\" (UID: \"a1faa733-f408-41fd-b7aa-0cd86283976a\") " pod="openshift-multus/multus-additional-cni-plugins-26nc6" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.895000 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5310c1fc-66c6-40aa-b1bf-5a59a2410f9c-etc-kubernetes\") pod \"multus-pznwl\" (UID: \"5310c1fc-66c6-40aa-b1bf-5a59a2410f9c\") " pod="openshift-multus/multus-pznwl" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.895036 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/5310c1fc-66c6-40aa-b1bf-5a59a2410f9c-host-run-k8s-cni-cncf-io\") pod \"multus-pznwl\" (UID: \"5310c1fc-66c6-40aa-b1bf-5a59a2410f9c\") " pod="openshift-multus/multus-pznwl" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.895065 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/5310c1fc-66c6-40aa-b1bf-5a59a2410f9c-host-var-lib-kubelet\") pod \"multus-pznwl\" (UID: \"5310c1fc-66c6-40aa-b1bf-5a59a2410f9c\") " pod="openshift-multus/multus-pznwl" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.895092 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/5310c1fc-66c6-40aa-b1bf-5a59a2410f9c-host-run-multus-certs\") pod \"multus-pznwl\" (UID: \"5310c1fc-66c6-40aa-b1bf-5a59a2410f9c\") " pod="openshift-multus/multus-pznwl" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.895093 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/5310c1fc-66c6-40aa-b1bf-5a59a2410f9c-hostroot\") pod \"multus-pznwl\" (UID: \"5310c1fc-66c6-40aa-b1bf-5a59a2410f9c\") " pod="openshift-multus/multus-pznwl" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.895140 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/5310c1fc-66c6-40aa-b1bf-5a59a2410f9c-os-release\") pod \"multus-pznwl\" (UID: \"5310c1fc-66c6-40aa-b1bf-5a59a2410f9c\") " pod="openshift-multus/multus-pznwl" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.895118 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/a1faa733-f408-41fd-b7aa-0cd86283976a-cnibin\") pod \"multus-additional-cni-plugins-26nc6\" (UID: \"a1faa733-f408-41fd-b7aa-0cd86283976a\") " pod="openshift-multus/multus-additional-cni-plugins-26nc6" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.895138 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/5310c1fc-66c6-40aa-b1bf-5a59a2410f9c-host-var-lib-cni-multus\") pod \"multus-pznwl\" (UID: \"5310c1fc-66c6-40aa-b1bf-5a59a2410f9c\") " pod="openshift-multus/multus-pznwl" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.895096 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/31c30642-6e60-41b4-a477-0d802424e0aa-rootfs\") pod \"machine-config-daemon-99qgp\" (UID: \"31c30642-6e60-41b4-a477-0d802424e0aa\") " pod="openshift-machine-config-operator/machine-config-daemon-99qgp" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.901687 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/a1faa733-f408-41fd-b7aa-0cd86283976a-tuning-conf-dir\") pod \"multus-additional-cni-plugins-26nc6\" (UID: \"a1faa733-f408-41fd-b7aa-0cd86283976a\") " pod="openshift-multus/multus-additional-cni-plugins-26nc6" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.901866 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/5310c1fc-66c6-40aa-b1bf-5a59a2410f9c-multus-cni-dir\") pod \"multus-pznwl\" (UID: \"5310c1fc-66c6-40aa-b1bf-5a59a2410f9c\") " pod="openshift-multus/multus-pznwl" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.901964 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/5310c1fc-66c6-40aa-b1bf-5a59a2410f9c-system-cni-dir\") pod \"multus-pznwl\" (UID: \"5310c1fc-66c6-40aa-b1bf-5a59a2410f9c\") " pod="openshift-multus/multus-pznwl" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.902751 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31c30642-6e60-41b4-a477-0d802424e0aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-99qgp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:28Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.902864 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31c30642-6e60-41b4-a477-0d802424e0aa-mcd-auth-proxy-config\") pod \"machine-config-daemon-99qgp\" (UID: \"31c30642-6e60-41b4-a477-0d802424e0aa\") " pod="openshift-machine-config-operator/machine-config-daemon-99qgp" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.902892 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/a1faa733-f408-41fd-b7aa-0cd86283976a-cni-binary-copy\") pod \"multus-additional-cni-plugins-26nc6\" (UID: \"a1faa733-f408-41fd-b7aa-0cd86283976a\") " pod="openshift-multus/multus-additional-cni-plugins-26nc6" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.903077 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/5310c1fc-66c6-40aa-b1bf-5a59a2410f9c-cni-binary-copy\") pod \"multus-pznwl\" (UID: \"5310c1fc-66c6-40aa-b1bf-5a59a2410f9c\") " pod="openshift-multus/multus-pznwl" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.908065 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31c30642-6e60-41b4-a477-0d802424e0aa-proxy-tls\") pod \"machine-config-daemon-99qgp\" (UID: \"31c30642-6e60-41b4-a477-0d802424e0aa\") " pod="openshift-machine-config-operator/machine-config-daemon-99qgp" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.917316 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7qc8q\" (UniqueName: \"kubernetes.io/projected/31c30642-6e60-41b4-a477-0d802424e0aa-kube-api-access-7qc8q\") pod \"machine-config-daemon-99qgp\" (UID: \"31c30642-6e60-41b4-a477-0d802424e0aa\") " pod="openshift-machine-config-operator/machine-config-daemon-99qgp" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.920036 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hkt48\" (UniqueName: \"kubernetes.io/projected/5310c1fc-66c6-40aa-b1bf-5a59a2410f9c-kube-api-access-hkt48\") pod \"multus-pznwl\" (UID: \"5310c1fc-66c6-40aa-b1bf-5a59a2410f9c\") " pod="openshift-multus/multus-pznwl" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.920753 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9nf9r\" (UniqueName: \"kubernetes.io/projected/a1faa733-f408-41fd-b7aa-0cd86283976a-kube-api-access-9nf9r\") pod \"multus-additional-cni-plugins-26nc6\" (UID: \"a1faa733-f408-41fd-b7aa-0cd86283976a\") " pod="openshift-multus/multus-additional-cni-plugins-26nc6" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.928305 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9872be1e-b995-4a00-904b-875dff96ad56\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f74a32a8fefea704cc73ffd1114083dab68aa44dc33f8d9d432d2d24216546c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbe00b46063d3968d59c393b87178e135f982abb7d0f32fa798b212d4b83df6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eedc9988e1a4a3c8cae730dc4a352103a6d665535b825e157d0f87c331b137c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2137fee4b7abc9765f9517d7b3b6ea796601a31d03151bf167d17136ced8b25a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f683e48f3577a68862eb6302383538376e41f49b20b1e0ea3f1b3b143f24872a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5de794187456b3576b969069a3b839d75c7918418a97fa35dff8bb40bb76e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5de794187456b3576b969069a3b839d75c7918418a97fa35dff8bb40bb76e39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08ce42b974fba00c409c0af971f0de351e50f18b2f06f8e722150da3239bb908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08ce42b974fba00c409c0af971f0de351e50f18b2f06f8e722150da3239bb908\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cfac6be03e3cfa0b0f3450b48b67957f2d5172b9c65e27f14c2a260bda5d423f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cfac6be03e3cfa0b0f3450b48b67957f2d5172b9c65e27f14c2a260bda5d423f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:28Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.941301 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-pznwl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5310c1fc-66c6-40aa-b1bf-5a59a2410f9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hkt48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-pznwl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:28Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.952695 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:28Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.964062 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:28Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.970554 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.973952 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6xq7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a75ce3c-5e1e-4b9f-bbde-7a3b0cce5eae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5c6nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6xq7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:28Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.983806 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:28Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:28 crc kubenswrapper[4860]: I1211 08:11:28.994480 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d8c2f28628dc2f1ffbdd98a55670e22cf6cdc8cf2c6442e9d1f8e5466b2f12e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://074ea63beac52bd196507ee7cfb1c3d7fc2d37d5ec6faa4d47dde62b922e1c43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:28Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.005877 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ec3758-85f1-4a22-a4ec-c805472607a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e831db1a15976d3967d56479689572c26f2599a766baa0402897a3008428c56f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a498aa90f914e1f0370df01dc5ef174df113a2a16d86b2e3f24b19a1b99b4c0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8caebd2f4dfcbb89c597029a07c335171889761d2d200206987537fb0a228db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34e91f615b76445ff3241a304bb2964fde59de7e8a8dfff6841b2e628a31e86d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2090cab02ad74ddfd6bb57ff50eba35d18f56b0cde1c230fe011c01994b693bc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 08:11:21.171011 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 08:11:21.177674 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-644352688/tls.crt::/tmp/serving-cert-644352688/tls.key\\\\\\\"\\\\nI1211 08:11:26.508606 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 08:11:26.512741 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 08:11:26.512778 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 08:11:26.512839 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 08:11:26.512858 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 08:11:26.521684 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1211 08:11:26.521700 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1211 08:11:26.521722 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:11:26.521732 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:11:26.521738 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 08:11:26.521746 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 08:11:26.521749 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 08:11:26.521753 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1211 08:11:26.524999 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2274fb298ca7d2b37b060b4dbf9b166d39ba77a108b08edeeb4d67196e72cb3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf0aff8a6ab816b6617b1feeca9146d41bb5e8808bf1db96f8d493500c135d85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf0aff8a6ab816b6617b1feeca9146d41bb5e8808bf1db96f8d493500c135d85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:29Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.015961 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:29Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.021708 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.026430 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a41f35db087418d1be16953f10377042abd3f4488e5ac1bc29ca5dcfc00a37e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:29Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.028610 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.039484 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-26nc6" Dec 11 08:11:29 crc kubenswrapper[4860]: W1211 08:11:29.039807 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod31c30642_6e60_41b4_a477_0d802424e0aa.slice/crio-9051fe5ee8ee538eda2f22376a445d34e6f32bf01353b0244c8605e22730aef1 WatchSource:0}: Error finding container 9051fe5ee8ee538eda2f22376a445d34e6f32bf01353b0244c8605e22730aef1: Status 404 returned error can't find the container with id 9051fe5ee8ee538eda2f22376a445d34e6f32bf01353b0244c8605e22730aef1 Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.047085 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-pznwl" Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.095929 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-xvnqp"] Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.096886 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.099318 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.099690 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.099818 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.100062 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.100180 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.100743 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.101262 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.110915 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:29Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.121554 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a41f35db087418d1be16953f10377042abd3f4488e5ac1bc29ca5dcfc00a37e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:29Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.134253 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ec3758-85f1-4a22-a4ec-c805472607a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e831db1a15976d3967d56479689572c26f2599a766baa0402897a3008428c56f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a498aa90f914e1f0370df01dc5ef174df113a2a16d86b2e3f24b19a1b99b4c0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8caebd2f4dfcbb89c597029a07c335171889761d2d200206987537fb0a228db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34e91f615b76445ff3241a304bb2964fde59de7e8a8dfff6841b2e628a31e86d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2090cab02ad74ddfd6bb57ff50eba35d18f56b0cde1c230fe011c01994b693bc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 08:11:21.171011 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 08:11:21.177674 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-644352688/tls.crt::/tmp/serving-cert-644352688/tls.key\\\\\\\"\\\\nI1211 08:11:26.508606 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 08:11:26.512741 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 08:11:26.512778 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 08:11:26.512839 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 08:11:26.512858 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 08:11:26.521684 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1211 08:11:26.521700 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1211 08:11:26.521722 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:11:26.521732 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:11:26.521738 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 08:11:26.521746 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 08:11:26.521749 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 08:11:26.521753 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1211 08:11:26.524999 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2274fb298ca7d2b37b060b4dbf9b166d39ba77a108b08edeeb4d67196e72cb3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf0aff8a6ab816b6617b1feeca9146d41bb5e8808bf1db96f8d493500c135d85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf0aff8a6ab816b6617b1feeca9146d41bb5e8808bf1db96f8d493500c135d85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:29Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.151001 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31c30642-6e60-41b4-a477-0d802424e0aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-99qgp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:29Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.168205 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-26nc6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1faa733-f408-41fd-b7aa-0cd86283976a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-26nc6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:29Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.185485 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0589a204-c98c-417d-8256-bfe3bf747660\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xvnqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:29Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.200126 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-pznwl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5310c1fc-66c6-40aa-b1bf-5a59a2410f9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hkt48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-pznwl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:29Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.201092 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/0589a204-c98c-417d-8256-bfe3bf747660-node-log\") pod \"ovnkube-node-xvnqp\" (UID: \"0589a204-c98c-417d-8256-bfe3bf747660\") " pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.201170 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/0589a204-c98c-417d-8256-bfe3bf747660-log-socket\") pod \"ovnkube-node-xvnqp\" (UID: \"0589a204-c98c-417d-8256-bfe3bf747660\") " pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.201238 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/0589a204-c98c-417d-8256-bfe3bf747660-run-ovn\") pod \"ovnkube-node-xvnqp\" (UID: \"0589a204-c98c-417d-8256-bfe3bf747660\") " pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.201509 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0589a204-c98c-417d-8256-bfe3bf747660-etc-openvswitch\") pod \"ovnkube-node-xvnqp\" (UID: \"0589a204-c98c-417d-8256-bfe3bf747660\") " pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.201580 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/0589a204-c98c-417d-8256-bfe3bf747660-ovnkube-script-lib\") pod \"ovnkube-node-xvnqp\" (UID: \"0589a204-c98c-417d-8256-bfe3bf747660\") " pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.201610 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0589a204-c98c-417d-8256-bfe3bf747660-var-lib-openvswitch\") pod \"ovnkube-node-xvnqp\" (UID: \"0589a204-c98c-417d-8256-bfe3bf747660\") " pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.201648 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/0589a204-c98c-417d-8256-bfe3bf747660-host-cni-bin\") pod \"ovnkube-node-xvnqp\" (UID: \"0589a204-c98c-417d-8256-bfe3bf747660\") " pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.201676 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0589a204-c98c-417d-8256-bfe3bf747660-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-xvnqp\" (UID: \"0589a204-c98c-417d-8256-bfe3bf747660\") " pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.201706 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/0589a204-c98c-417d-8256-bfe3bf747660-systemd-units\") pod \"ovnkube-node-xvnqp\" (UID: \"0589a204-c98c-417d-8256-bfe3bf747660\") " pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.201726 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/0589a204-c98c-417d-8256-bfe3bf747660-host-run-netns\") pod \"ovnkube-node-xvnqp\" (UID: \"0589a204-c98c-417d-8256-bfe3bf747660\") " pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.201751 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/0589a204-c98c-417d-8256-bfe3bf747660-host-kubelet\") pod \"ovnkube-node-xvnqp\" (UID: \"0589a204-c98c-417d-8256-bfe3bf747660\") " pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.201769 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0589a204-c98c-417d-8256-bfe3bf747660-host-run-ovn-kubernetes\") pod \"ovnkube-node-xvnqp\" (UID: \"0589a204-c98c-417d-8256-bfe3bf747660\") " pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.201790 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/0589a204-c98c-417d-8256-bfe3bf747660-host-slash\") pod \"ovnkube-node-xvnqp\" (UID: \"0589a204-c98c-417d-8256-bfe3bf747660\") " pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.201809 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/0589a204-c98c-417d-8256-bfe3bf747660-env-overrides\") pod \"ovnkube-node-xvnqp\" (UID: \"0589a204-c98c-417d-8256-bfe3bf747660\") " pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.201826 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0589a204-c98c-417d-8256-bfe3bf747660-run-openvswitch\") pod \"ovnkube-node-xvnqp\" (UID: \"0589a204-c98c-417d-8256-bfe3bf747660\") " pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.201847 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/0589a204-c98c-417d-8256-bfe3bf747660-ovnkube-config\") pod \"ovnkube-node-xvnqp\" (UID: \"0589a204-c98c-417d-8256-bfe3bf747660\") " pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.201893 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/0589a204-c98c-417d-8256-bfe3bf747660-host-cni-netd\") pod \"ovnkube-node-xvnqp\" (UID: \"0589a204-c98c-417d-8256-bfe3bf747660\") " pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.201915 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/0589a204-c98c-417d-8256-bfe3bf747660-ovn-node-metrics-cert\") pod \"ovnkube-node-xvnqp\" (UID: \"0589a204-c98c-417d-8256-bfe3bf747660\") " pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.202138 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/0589a204-c98c-417d-8256-bfe3bf747660-run-systemd\") pod \"ovnkube-node-xvnqp\" (UID: \"0589a204-c98c-417d-8256-bfe3bf747660\") " pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.202203 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tftsj\" (UniqueName: \"kubernetes.io/projected/0589a204-c98c-417d-8256-bfe3bf747660-kube-api-access-tftsj\") pod \"ovnkube-node-xvnqp\" (UID: \"0589a204-c98c-417d-8256-bfe3bf747660\") " pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.222702 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9872be1e-b995-4a00-904b-875dff96ad56\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f74a32a8fefea704cc73ffd1114083dab68aa44dc33f8d9d432d2d24216546c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbe00b46063d3968d59c393b87178e135f982abb7d0f32fa798b212d4b83df6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eedc9988e1a4a3c8cae730dc4a352103a6d665535b825e157d0f87c331b137c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2137fee4b7abc9765f9517d7b3b6ea796601a31d03151bf167d17136ced8b25a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f683e48f3577a68862eb6302383538376e41f49b20b1e0ea3f1b3b143f24872a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5de794187456b3576b969069a3b839d75c7918418a97fa35dff8bb40bb76e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5de794187456b3576b969069a3b839d75c7918418a97fa35dff8bb40bb76e39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08ce42b974fba00c409c0af971f0de351e50f18b2f06f8e722150da3239bb908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08ce42b974fba00c409c0af971f0de351e50f18b2f06f8e722150da3239bb908\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cfac6be03e3cfa0b0f3450b48b67957f2d5172b9c65e27f14c2a260bda5d423f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cfac6be03e3cfa0b0f3450b48b67957f2d5172b9c65e27f14c2a260bda5d423f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:29Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.235011 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:29Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.245834 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d8c2f28628dc2f1ffbdd98a55670e22cf6cdc8cf2c6442e9d1f8e5466b2f12e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://074ea63beac52bd196507ee7cfb1c3d7fc2d37d5ec6faa4d47dde62b922e1c43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:29Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.257288 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:29Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.271275 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:29Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.282948 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6xq7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a75ce3c-5e1e-4b9f-bbde-7a3b0cce5eae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5c6nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6xq7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:29Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.305137 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/0589a204-c98c-417d-8256-bfe3bf747660-systemd-units\") pod \"ovnkube-node-xvnqp\" (UID: \"0589a204-c98c-417d-8256-bfe3bf747660\") " pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.305174 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/0589a204-c98c-417d-8256-bfe3bf747660-host-run-netns\") pod \"ovnkube-node-xvnqp\" (UID: \"0589a204-c98c-417d-8256-bfe3bf747660\") " pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.305191 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/0589a204-c98c-417d-8256-bfe3bf747660-host-kubelet\") pod \"ovnkube-node-xvnqp\" (UID: \"0589a204-c98c-417d-8256-bfe3bf747660\") " pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.305209 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0589a204-c98c-417d-8256-bfe3bf747660-host-run-ovn-kubernetes\") pod \"ovnkube-node-xvnqp\" (UID: \"0589a204-c98c-417d-8256-bfe3bf747660\") " pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.305225 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/0589a204-c98c-417d-8256-bfe3bf747660-host-slash\") pod \"ovnkube-node-xvnqp\" (UID: \"0589a204-c98c-417d-8256-bfe3bf747660\") " pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.305241 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/0589a204-c98c-417d-8256-bfe3bf747660-env-overrides\") pod \"ovnkube-node-xvnqp\" (UID: \"0589a204-c98c-417d-8256-bfe3bf747660\") " pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.305263 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0589a204-c98c-417d-8256-bfe3bf747660-run-openvswitch\") pod \"ovnkube-node-xvnqp\" (UID: \"0589a204-c98c-417d-8256-bfe3bf747660\") " pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.305278 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/0589a204-c98c-417d-8256-bfe3bf747660-ovnkube-config\") pod \"ovnkube-node-xvnqp\" (UID: \"0589a204-c98c-417d-8256-bfe3bf747660\") " pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.305297 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/0589a204-c98c-417d-8256-bfe3bf747660-host-cni-netd\") pod \"ovnkube-node-xvnqp\" (UID: \"0589a204-c98c-417d-8256-bfe3bf747660\") " pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.305319 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/0589a204-c98c-417d-8256-bfe3bf747660-ovn-node-metrics-cert\") pod \"ovnkube-node-xvnqp\" (UID: \"0589a204-c98c-417d-8256-bfe3bf747660\") " pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.305316 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/0589a204-c98c-417d-8256-bfe3bf747660-host-run-netns\") pod \"ovnkube-node-xvnqp\" (UID: \"0589a204-c98c-417d-8256-bfe3bf747660\") " pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.305367 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/0589a204-c98c-417d-8256-bfe3bf747660-run-systemd\") pod \"ovnkube-node-xvnqp\" (UID: \"0589a204-c98c-417d-8256-bfe3bf747660\") " pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.305371 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/0589a204-c98c-417d-8256-bfe3bf747660-host-kubelet\") pod \"ovnkube-node-xvnqp\" (UID: \"0589a204-c98c-417d-8256-bfe3bf747660\") " pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.305372 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/0589a204-c98c-417d-8256-bfe3bf747660-systemd-units\") pod \"ovnkube-node-xvnqp\" (UID: \"0589a204-c98c-417d-8256-bfe3bf747660\") " pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.305416 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0589a204-c98c-417d-8256-bfe3bf747660-host-run-ovn-kubernetes\") pod \"ovnkube-node-xvnqp\" (UID: \"0589a204-c98c-417d-8256-bfe3bf747660\") " pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.305340 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/0589a204-c98c-417d-8256-bfe3bf747660-host-slash\") pod \"ovnkube-node-xvnqp\" (UID: \"0589a204-c98c-417d-8256-bfe3bf747660\") " pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.305336 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/0589a204-c98c-417d-8256-bfe3bf747660-run-systemd\") pod \"ovnkube-node-xvnqp\" (UID: \"0589a204-c98c-417d-8256-bfe3bf747660\") " pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.305401 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/0589a204-c98c-417d-8256-bfe3bf747660-host-cni-netd\") pod \"ovnkube-node-xvnqp\" (UID: \"0589a204-c98c-417d-8256-bfe3bf747660\") " pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.305375 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0589a204-c98c-417d-8256-bfe3bf747660-run-openvswitch\") pod \"ovnkube-node-xvnqp\" (UID: \"0589a204-c98c-417d-8256-bfe3bf747660\") " pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.305533 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tftsj\" (UniqueName: \"kubernetes.io/projected/0589a204-c98c-417d-8256-bfe3bf747660-kube-api-access-tftsj\") pod \"ovnkube-node-xvnqp\" (UID: \"0589a204-c98c-417d-8256-bfe3bf747660\") " pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.305828 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/0589a204-c98c-417d-8256-bfe3bf747660-node-log\") pod \"ovnkube-node-xvnqp\" (UID: \"0589a204-c98c-417d-8256-bfe3bf747660\") " pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.305880 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/0589a204-c98c-417d-8256-bfe3bf747660-log-socket\") pod \"ovnkube-node-xvnqp\" (UID: \"0589a204-c98c-417d-8256-bfe3bf747660\") " pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.305905 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/0589a204-c98c-417d-8256-bfe3bf747660-run-ovn\") pod \"ovnkube-node-xvnqp\" (UID: \"0589a204-c98c-417d-8256-bfe3bf747660\") " pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.305940 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/0589a204-c98c-417d-8256-bfe3bf747660-env-overrides\") pod \"ovnkube-node-xvnqp\" (UID: \"0589a204-c98c-417d-8256-bfe3bf747660\") " pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.305951 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0589a204-c98c-417d-8256-bfe3bf747660-etc-openvswitch\") pod \"ovnkube-node-xvnqp\" (UID: \"0589a204-c98c-417d-8256-bfe3bf747660\") " pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.305970 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/0589a204-c98c-417d-8256-bfe3bf747660-node-log\") pod \"ovnkube-node-xvnqp\" (UID: \"0589a204-c98c-417d-8256-bfe3bf747660\") " pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.305978 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/0589a204-c98c-417d-8256-bfe3bf747660-ovnkube-script-lib\") pod \"ovnkube-node-xvnqp\" (UID: \"0589a204-c98c-417d-8256-bfe3bf747660\") " pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.306001 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/0589a204-c98c-417d-8256-bfe3bf747660-run-ovn\") pod \"ovnkube-node-xvnqp\" (UID: \"0589a204-c98c-417d-8256-bfe3bf747660\") " pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.306014 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0589a204-c98c-417d-8256-bfe3bf747660-etc-openvswitch\") pod \"ovnkube-node-xvnqp\" (UID: \"0589a204-c98c-417d-8256-bfe3bf747660\") " pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.306021 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0589a204-c98c-417d-8256-bfe3bf747660-var-lib-openvswitch\") pod \"ovnkube-node-xvnqp\" (UID: \"0589a204-c98c-417d-8256-bfe3bf747660\") " pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.305983 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/0589a204-c98c-417d-8256-bfe3bf747660-log-socket\") pod \"ovnkube-node-xvnqp\" (UID: \"0589a204-c98c-417d-8256-bfe3bf747660\") " pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.306050 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/0589a204-c98c-417d-8256-bfe3bf747660-host-cni-bin\") pod \"ovnkube-node-xvnqp\" (UID: \"0589a204-c98c-417d-8256-bfe3bf747660\") " pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.306064 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0589a204-c98c-417d-8256-bfe3bf747660-var-lib-openvswitch\") pod \"ovnkube-node-xvnqp\" (UID: \"0589a204-c98c-417d-8256-bfe3bf747660\") " pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.306074 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0589a204-c98c-417d-8256-bfe3bf747660-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-xvnqp\" (UID: \"0589a204-c98c-417d-8256-bfe3bf747660\") " pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.306099 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/0589a204-c98c-417d-8256-bfe3bf747660-host-cni-bin\") pod \"ovnkube-node-xvnqp\" (UID: \"0589a204-c98c-417d-8256-bfe3bf747660\") " pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.306166 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0589a204-c98c-417d-8256-bfe3bf747660-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-xvnqp\" (UID: \"0589a204-c98c-417d-8256-bfe3bf747660\") " pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.306200 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/0589a204-c98c-417d-8256-bfe3bf747660-ovnkube-config\") pod \"ovnkube-node-xvnqp\" (UID: \"0589a204-c98c-417d-8256-bfe3bf747660\") " pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.306470 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/0589a204-c98c-417d-8256-bfe3bf747660-ovnkube-script-lib\") pod \"ovnkube-node-xvnqp\" (UID: \"0589a204-c98c-417d-8256-bfe3bf747660\") " pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.308399 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/0589a204-c98c-417d-8256-bfe3bf747660-ovn-node-metrics-cert\") pod \"ovnkube-node-xvnqp\" (UID: \"0589a204-c98c-417d-8256-bfe3bf747660\") " pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.321461 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tftsj\" (UniqueName: \"kubernetes.io/projected/0589a204-c98c-417d-8256-bfe3bf747660-kube-api-access-tftsj\") pod \"ovnkube-node-xvnqp\" (UID: \"0589a204-c98c-417d-8256-bfe3bf747660\") " pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.410722 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.629001 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.632496 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.636908 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.638502 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6xq7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a75ce3c-5e1e-4b9f-bbde-7a3b0cce5eae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5c6nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6xq7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:29Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.651408 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:29Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.664204 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d8c2f28628dc2f1ffbdd98a55670e22cf6cdc8cf2c6442e9d1f8e5466b2f12e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://074ea63beac52bd196507ee7cfb1c3d7fc2d37d5ec6faa4d47dde62b922e1c43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:29Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.674973 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:29Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.688138 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:29Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.699508 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ec3758-85f1-4a22-a4ec-c805472607a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e831db1a15976d3967d56479689572c26f2599a766baa0402897a3008428c56f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a498aa90f914e1f0370df01dc5ef174df113a2a16d86b2e3f24b19a1b99b4c0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8caebd2f4dfcbb89c597029a07c335171889761d2d200206987537fb0a228db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34e91f615b76445ff3241a304bb2964fde59de7e8a8dfff6841b2e628a31e86d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2090cab02ad74ddfd6bb57ff50eba35d18f56b0cde1c230fe011c01994b693bc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 08:11:21.171011 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 08:11:21.177674 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-644352688/tls.crt::/tmp/serving-cert-644352688/tls.key\\\\\\\"\\\\nI1211 08:11:26.508606 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 08:11:26.512741 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 08:11:26.512778 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 08:11:26.512839 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 08:11:26.512858 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 08:11:26.521684 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1211 08:11:26.521700 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1211 08:11:26.521722 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:11:26.521732 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:11:26.521738 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 08:11:26.521746 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 08:11:26.521749 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 08:11:26.521753 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1211 08:11:26.524999 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2274fb298ca7d2b37b060b4dbf9b166d39ba77a108b08edeeb4d67196e72cb3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf0aff8a6ab816b6617b1feeca9146d41bb5e8808bf1db96f8d493500c135d85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf0aff8a6ab816b6617b1feeca9146d41bb5e8808bf1db96f8d493500c135d85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:29Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.709434 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:29Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.721320 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-pznwl" event={"ID":"5310c1fc-66c6-40aa-b1bf-5a59a2410f9c","Type":"ContainerStarted","Data":"d2d6ff00ce322a28607bc3f6db05c5e4fb14acb6777ea315786ef53963ca504a"} Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.721369 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-pznwl" event={"ID":"5310c1fc-66c6-40aa-b1bf-5a59a2410f9c","Type":"ContainerStarted","Data":"180d323f5cc00bd6852fa596c4fb247371988e2cff0db183354711594dc8912a"} Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.723705 4860 generic.go:334] "Generic (PLEG): container finished" podID="0589a204-c98c-417d-8256-bfe3bf747660" containerID="17cbcb49c06f3a9059cd5a0f05eedb8c785da30a3e502e31bae2f57b3486cb14" exitCode=0 Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.723759 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" event={"ID":"0589a204-c98c-417d-8256-bfe3bf747660","Type":"ContainerDied","Data":"17cbcb49c06f3a9059cd5a0f05eedb8c785da30a3e502e31bae2f57b3486cb14"} Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.723782 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" event={"ID":"0589a204-c98c-417d-8256-bfe3bf747660","Type":"ContainerStarted","Data":"7604d81409ead4ab386e79a3646e8237268fa238b5c89b25f14fdcfbc5e2b819"} Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.725034 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"25a6a154da2b74952951ae94efc914ba078ea4bbe5893360b3c84d46be6bf973"} Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.725957 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-26nc6" event={"ID":"a1faa733-f408-41fd-b7aa-0cd86283976a","Type":"ContainerStarted","Data":"8ac45db4f858a1c18a6b87472fc1cea2300f5845f1ad62c22260374995548a0b"} Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.725991 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-26nc6" event={"ID":"a1faa733-f408-41fd-b7aa-0cd86283976a","Type":"ContainerStarted","Data":"9a0126e05af1b8847868d43486c1a0bde77939b3c7a1980cc6fb8579dd48bfb2"} Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.727279 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" event={"ID":"31c30642-6e60-41b4-a477-0d802424e0aa","Type":"ContainerStarted","Data":"79969402f5c33852a26fc839f37a45febf4f2377d77823284fbb01769e5585fe"} Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.727301 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" event={"ID":"31c30642-6e60-41b4-a477-0d802424e0aa","Type":"ContainerStarted","Data":"b2d4d5de7106175087d91acb1f0c7552014f0c08bfcd65cbbf184d3f8d7826ff"} Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.727310 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" event={"ID":"31c30642-6e60-41b4-a477-0d802424e0aa","Type":"ContainerStarted","Data":"9051fe5ee8ee538eda2f22376a445d34e6f32bf01353b0244c8605e22730aef1"} Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.731390 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a41f35db087418d1be16953f10377042abd3f4488e5ac1bc29ca5dcfc00a37e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:29Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.737717 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-6xq7j" event={"ID":"6a75ce3c-5e1e-4b9f-bbde-7a3b0cce5eae","Type":"ContainerStarted","Data":"fe8d4d675567b41728e476351811a93c47b19868ccbadab8c613a4eb8eca8562"} Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.742455 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31c30642-6e60-41b4-a477-0d802424e0aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-99qgp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:29Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.756384 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-26nc6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1faa733-f408-41fd-b7aa-0cd86283976a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-26nc6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:29Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.771753 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0589a204-c98c-417d-8256-bfe3bf747660\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xvnqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:29Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.797810 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9872be1e-b995-4a00-904b-875dff96ad56\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f74a32a8fefea704cc73ffd1114083dab68aa44dc33f8d9d432d2d24216546c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbe00b46063d3968d59c393b87178e135f982abb7d0f32fa798b212d4b83df6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eedc9988e1a4a3c8cae730dc4a352103a6d665535b825e157d0f87c331b137c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2137fee4b7abc9765f9517d7b3b6ea796601a31d03151bf167d17136ced8b25a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f683e48f3577a68862eb6302383538376e41f49b20b1e0ea3f1b3b143f24872a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5de794187456b3576b969069a3b839d75c7918418a97fa35dff8bb40bb76e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5de794187456b3576b969069a3b839d75c7918418a97fa35dff8bb40bb76e39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08ce42b974fba00c409c0af971f0de351e50f18b2f06f8e722150da3239bb908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08ce42b974fba00c409c0af971f0de351e50f18b2f06f8e722150da3239bb908\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cfac6be03e3cfa0b0f3450b48b67957f2d5172b9c65e27f14c2a260bda5d423f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cfac6be03e3cfa0b0f3450b48b67957f2d5172b9c65e27f14c2a260bda5d423f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:29Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.831671 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-pznwl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5310c1fc-66c6-40aa-b1bf-5a59a2410f9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hkt48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-pznwl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:29Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.873353 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ec3758-85f1-4a22-a4ec-c805472607a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e831db1a15976d3967d56479689572c26f2599a766baa0402897a3008428c56f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a498aa90f914e1f0370df01dc5ef174df113a2a16d86b2e3f24b19a1b99b4c0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8caebd2f4dfcbb89c597029a07c335171889761d2d200206987537fb0a228db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34e91f615b76445ff3241a304bb2964fde59de7e8a8dfff6841b2e628a31e86d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2090cab02ad74ddfd6bb57ff50eba35d18f56b0cde1c230fe011c01994b693bc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 08:11:21.171011 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 08:11:21.177674 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-644352688/tls.crt::/tmp/serving-cert-644352688/tls.key\\\\\\\"\\\\nI1211 08:11:26.508606 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 08:11:26.512741 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 08:11:26.512778 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 08:11:26.512839 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 08:11:26.512858 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 08:11:26.521684 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1211 08:11:26.521700 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1211 08:11:26.521722 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:11:26.521732 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:11:26.521738 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 08:11:26.521746 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 08:11:26.521749 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 08:11:26.521753 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1211 08:11:26.524999 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2274fb298ca7d2b37b060b4dbf9b166d39ba77a108b08edeeb4d67196e72cb3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf0aff8a6ab816b6617b1feeca9146d41bb5e8808bf1db96f8d493500c135d85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf0aff8a6ab816b6617b1feeca9146d41bb5e8808bf1db96f8d493500c135d85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:29Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.911076 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:29Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.952008 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a41f35db087418d1be16953f10377042abd3f4488e5ac1bc29ca5dcfc00a37e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:29Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:29 crc kubenswrapper[4860]: I1211 08:11:29.993521 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-26nc6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1faa733-f408-41fd-b7aa-0cd86283976a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ac45db4f858a1c18a6b87472fc1cea2300f5845f1ad62c22260374995548a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-26nc6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:29Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:30 crc kubenswrapper[4860]: I1211 08:11:30.037022 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0589a204-c98c-417d-8256-bfe3bf747660\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xvnqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:30Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:30 crc kubenswrapper[4860]: I1211 08:11:30.071805 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9aa0aed-62b3-4ea2-82ae-9a653c1101a5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ad4ba1b769c6cb2b34073798f8e3c25eb0177e6d3059484bf365e6981cab710\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://077aa1dd2fe5189ff60a0f2365e47a6a545cca0de7aa6379370908db4ea5838a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d7c919899b74a2b0fcd9221f3f769a9d013267354b200a282669d342d6413d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5483384f1d2ebc1fefa9706c9e67a3dc5a00ff643f803e3f2707c252e27ad94\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:30Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:30 crc kubenswrapper[4860]: I1211 08:11:30.117062 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31c30642-6e60-41b4-a477-0d802424e0aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79969402f5c33852a26fc839f37a45febf4f2377d77823284fbb01769e5585fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2d4d5de7106175087d91acb1f0c7552014f0c08bfcd65cbbf184d3f8d7826ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-99qgp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:30Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:30 crc kubenswrapper[4860]: I1211 08:11:30.182045 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9872be1e-b995-4a00-904b-875dff96ad56\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f74a32a8fefea704cc73ffd1114083dab68aa44dc33f8d9d432d2d24216546c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbe00b46063d3968d59c393b87178e135f982abb7d0f32fa798b212d4b83df6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eedc9988e1a4a3c8cae730dc4a352103a6d665535b825e157d0f87c331b137c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2137fee4b7abc9765f9517d7b3b6ea796601a31d03151bf167d17136ced8b25a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f683e48f3577a68862eb6302383538376e41f49b20b1e0ea3f1b3b143f24872a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5de794187456b3576b969069a3b839d75c7918418a97fa35dff8bb40bb76e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5de794187456b3576b969069a3b839d75c7918418a97fa35dff8bb40bb76e39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08ce42b974fba00c409c0af971f0de351e50f18b2f06f8e722150da3239bb908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08ce42b974fba00c409c0af971f0de351e50f18b2f06f8e722150da3239bb908\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cfac6be03e3cfa0b0f3450b48b67957f2d5172b9c65e27f14c2a260bda5d423f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cfac6be03e3cfa0b0f3450b48b67957f2d5172b9c65e27f14c2a260bda5d423f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:30Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:30 crc kubenswrapper[4860]: I1211 08:11:30.202950 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-pznwl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5310c1fc-66c6-40aa-b1bf-5a59a2410f9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2d6ff00ce322a28607bc3f6db05c5e4fb14acb6777ea315786ef53963ca504a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hkt48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-pznwl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:30Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:30 crc kubenswrapper[4860]: I1211 08:11:30.219012 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:11:30 crc kubenswrapper[4860]: I1211 08:11:30.219130 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:11:30 crc kubenswrapper[4860]: E1211 08:11:30.219203 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:11:34.219175693 +0000 UTC m=+26.947694748 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:11:30 crc kubenswrapper[4860]: E1211 08:11:30.219221 4860 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 11 08:11:30 crc kubenswrapper[4860]: E1211 08:11:30.219270 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-11 08:11:34.219256346 +0000 UTC m=+26.947775401 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 11 08:11:30 crc kubenswrapper[4860]: I1211 08:11:30.237359 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:30Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:30 crc kubenswrapper[4860]: I1211 08:11:30.271827 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25a6a154da2b74952951ae94efc914ba078ea4bbe5893360b3c84d46be6bf973\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:30Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:30 crc kubenswrapper[4860]: I1211 08:11:30.310203 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6xq7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a75ce3c-5e1e-4b9f-bbde-7a3b0cce5eae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe8d4d675567b41728e476351811a93c47b19868ccbadab8c613a4eb8eca8562\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5c6nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6xq7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:30Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:30 crc kubenswrapper[4860]: I1211 08:11:30.320624 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:11:30 crc kubenswrapper[4860]: I1211 08:11:30.320681 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:11:30 crc kubenswrapper[4860]: I1211 08:11:30.320706 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:11:30 crc kubenswrapper[4860]: E1211 08:11:30.320800 4860 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 11 08:11:30 crc kubenswrapper[4860]: E1211 08:11:30.320800 4860 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 11 08:11:30 crc kubenswrapper[4860]: E1211 08:11:30.320828 4860 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 11 08:11:30 crc kubenswrapper[4860]: E1211 08:11:30.320842 4860 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 08:11:30 crc kubenswrapper[4860]: E1211 08:11:30.320841 4860 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 11 08:11:30 crc kubenswrapper[4860]: E1211 08:11:30.320815 4860 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 11 08:11:30 crc kubenswrapper[4860]: E1211 08:11:30.320874 4860 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 08:11:30 crc kubenswrapper[4860]: E1211 08:11:30.320887 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-11 08:11:34.320874172 +0000 UTC m=+27.049393227 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 11 08:11:30 crc kubenswrapper[4860]: E1211 08:11:30.320904 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-11 08:11:34.320893022 +0000 UTC m=+27.049412077 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 08:11:30 crc kubenswrapper[4860]: E1211 08:11:30.320919 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-11 08:11:34.320911573 +0000 UTC m=+27.049430628 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 08:11:30 crc kubenswrapper[4860]: I1211 08:11:30.353563 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:30Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:30 crc kubenswrapper[4860]: I1211 08:11:30.395877 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d8c2f28628dc2f1ffbdd98a55670e22cf6cdc8cf2c6442e9d1f8e5466b2f12e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://074ea63beac52bd196507ee7cfb1c3d7fc2d37d5ec6faa4d47dde62b922e1c43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:30Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:30 crc kubenswrapper[4860]: I1211 08:11:30.428889 4860 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 11 08:11:30 crc kubenswrapper[4860]: I1211 08:11:30.430907 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:30 crc kubenswrapper[4860]: I1211 08:11:30.430938 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:30 crc kubenswrapper[4860]: I1211 08:11:30.430948 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:30 crc kubenswrapper[4860]: I1211 08:11:30.431035 4860 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 11 08:11:30 crc kubenswrapper[4860]: I1211 08:11:30.436304 4860 kubelet_node_status.go:115] "Node was previously registered" node="crc" Dec 11 08:11:30 crc kubenswrapper[4860]: I1211 08:11:30.436566 4860 kubelet_node_status.go:79] "Successfully registered node" node="crc" Dec 11 08:11:30 crc kubenswrapper[4860]: I1211 08:11:30.437553 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:30 crc kubenswrapper[4860]: I1211 08:11:30.437588 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:30 crc kubenswrapper[4860]: I1211 08:11:30.437599 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:30 crc kubenswrapper[4860]: I1211 08:11:30.437616 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:30 crc kubenswrapper[4860]: I1211 08:11:30.437628 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:30Z","lastTransitionTime":"2025-12-11T08:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:30 crc kubenswrapper[4860]: E1211 08:11:30.456437 4860 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"52631a1a-bc27-4dd5-8309-5dc39b5dcc41\\\",\\\"systemUUID\\\":\\\"299d3cd2-e1ad-4f5c-8af4-f12c0a5d76e7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:30Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:30 crc kubenswrapper[4860]: I1211 08:11:30.459632 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:30 crc kubenswrapper[4860]: I1211 08:11:30.459677 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:30 crc kubenswrapper[4860]: I1211 08:11:30.459686 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:30 crc kubenswrapper[4860]: I1211 08:11:30.459699 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:30 crc kubenswrapper[4860]: I1211 08:11:30.459708 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:30Z","lastTransitionTime":"2025-12-11T08:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:30 crc kubenswrapper[4860]: E1211 08:11:30.471031 4860 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"52631a1a-bc27-4dd5-8309-5dc39b5dcc41\\\",\\\"systemUUID\\\":\\\"299d3cd2-e1ad-4f5c-8af4-f12c0a5d76e7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:30Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:30 crc kubenswrapper[4860]: I1211 08:11:30.473965 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:30 crc kubenswrapper[4860]: I1211 08:11:30.473995 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:30 crc kubenswrapper[4860]: I1211 08:11:30.474002 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:30 crc kubenswrapper[4860]: I1211 08:11:30.474014 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:30 crc kubenswrapper[4860]: I1211 08:11:30.474024 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:30Z","lastTransitionTime":"2025-12-11T08:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:30 crc kubenswrapper[4860]: E1211 08:11:30.484916 4860 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"52631a1a-bc27-4dd5-8309-5dc39b5dcc41\\\",\\\"systemUUID\\\":\\\"299d3cd2-e1ad-4f5c-8af4-f12c0a5d76e7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:30Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:30 crc kubenswrapper[4860]: I1211 08:11:30.489208 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:30 crc kubenswrapper[4860]: I1211 08:11:30.489242 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:30 crc kubenswrapper[4860]: I1211 08:11:30.489252 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:30 crc kubenswrapper[4860]: I1211 08:11:30.489267 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:30 crc kubenswrapper[4860]: I1211 08:11:30.489278 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:30Z","lastTransitionTime":"2025-12-11T08:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:30 crc kubenswrapper[4860]: E1211 08:11:30.501957 4860 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"52631a1a-bc27-4dd5-8309-5dc39b5dcc41\\\",\\\"systemUUID\\\":\\\"299d3cd2-e1ad-4f5c-8af4-f12c0a5d76e7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:30Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:30 crc kubenswrapper[4860]: I1211 08:11:30.507955 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:30 crc kubenswrapper[4860]: I1211 08:11:30.507989 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:30 crc kubenswrapper[4860]: I1211 08:11:30.508000 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:30 crc kubenswrapper[4860]: I1211 08:11:30.508016 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:30 crc kubenswrapper[4860]: I1211 08:11:30.508028 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:30Z","lastTransitionTime":"2025-12-11T08:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:30 crc kubenswrapper[4860]: E1211 08:11:30.520613 4860 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"52631a1a-bc27-4dd5-8309-5dc39b5dcc41\\\",\\\"systemUUID\\\":\\\"299d3cd2-e1ad-4f5c-8af4-f12c0a5d76e7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:30Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:30 crc kubenswrapper[4860]: E1211 08:11:30.520746 4860 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 11 08:11:30 crc kubenswrapper[4860]: I1211 08:11:30.522724 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:30 crc kubenswrapper[4860]: I1211 08:11:30.522758 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:30 crc kubenswrapper[4860]: I1211 08:11:30.522779 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:30 crc kubenswrapper[4860]: I1211 08:11:30.522797 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:30 crc kubenswrapper[4860]: I1211 08:11:30.522808 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:30Z","lastTransitionTime":"2025-12-11T08:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:30 crc kubenswrapper[4860]: I1211 08:11:30.566281 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-gsx9l"] Dec 11 08:11:30 crc kubenswrapper[4860]: I1211 08:11:30.566580 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-gsx9l" Dec 11 08:11:30 crc kubenswrapper[4860]: I1211 08:11:30.570434 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 11 08:11:30 crc kubenswrapper[4860]: I1211 08:11:30.571020 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 11 08:11:30 crc kubenswrapper[4860]: I1211 08:11:30.571290 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 11 08:11:30 crc kubenswrapper[4860]: I1211 08:11:30.571422 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 11 08:11:30 crc kubenswrapper[4860]: I1211 08:11:30.577897 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:11:30 crc kubenswrapper[4860]: I1211 08:11:30.577983 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:11:30 crc kubenswrapper[4860]: E1211 08:11:30.578094 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:11:30 crc kubenswrapper[4860]: I1211 08:11:30.578115 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:11:30 crc kubenswrapper[4860]: E1211 08:11:30.578168 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:11:30 crc kubenswrapper[4860]: E1211 08:11:30.578248 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:11:30 crc kubenswrapper[4860]: I1211 08:11:30.587214 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9aa0aed-62b3-4ea2-82ae-9a653c1101a5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ad4ba1b769c6cb2b34073798f8e3c25eb0177e6d3059484bf365e6981cab710\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://077aa1dd2fe5189ff60a0f2365e47a6a545cca0de7aa6379370908db4ea5838a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d7c919899b74a2b0fcd9221f3f769a9d013267354b200a282669d342d6413d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5483384f1d2ebc1fefa9706c9e67a3dc5a00ff643f803e3f2707c252e27ad94\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:30Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:30 crc kubenswrapper[4860]: I1211 08:11:30.602343 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31c30642-6e60-41b4-a477-0d802424e0aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79969402f5c33852a26fc839f37a45febf4f2377d77823284fbb01769e5585fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2d4d5de7106175087d91acb1f0c7552014f0c08bfcd65cbbf184d3f8d7826ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-99qgp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:30Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:30 crc kubenswrapper[4860]: I1211 08:11:30.624616 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:30 crc kubenswrapper[4860]: I1211 08:11:30.624661 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:30 crc kubenswrapper[4860]: I1211 08:11:30.624671 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:30 crc kubenswrapper[4860]: I1211 08:11:30.624684 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:30 crc kubenswrapper[4860]: I1211 08:11:30.624693 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:30Z","lastTransitionTime":"2025-12-11T08:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:30 crc kubenswrapper[4860]: I1211 08:11:30.635003 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-26nc6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1faa733-f408-41fd-b7aa-0cd86283976a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ac45db4f858a1c18a6b87472fc1cea2300f5845f1ad62c22260374995548a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-26nc6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:30Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:30 crc kubenswrapper[4860]: I1211 08:11:30.678208 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0589a204-c98c-417d-8256-bfe3bf747660\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xvnqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:30Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:30 crc kubenswrapper[4860]: I1211 08:11:30.711765 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gsx9l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aba5b6b9-f88d-4763-9562-6b4dec606df9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6zw9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gsx9l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:30Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:30 crc kubenswrapper[4860]: I1211 08:11:30.724010 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6zw9s\" (UniqueName: \"kubernetes.io/projected/aba5b6b9-f88d-4763-9562-6b4dec606df9-kube-api-access-6zw9s\") pod \"node-ca-gsx9l\" (UID: \"aba5b6b9-f88d-4763-9562-6b4dec606df9\") " pod="openshift-image-registry/node-ca-gsx9l" Dec 11 08:11:30 crc kubenswrapper[4860]: I1211 08:11:30.724087 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/aba5b6b9-f88d-4763-9562-6b4dec606df9-serviceca\") pod \"node-ca-gsx9l\" (UID: \"aba5b6b9-f88d-4763-9562-6b4dec606df9\") " pod="openshift-image-registry/node-ca-gsx9l" Dec 11 08:11:30 crc kubenswrapper[4860]: I1211 08:11:30.724135 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/aba5b6b9-f88d-4763-9562-6b4dec606df9-host\") pod \"node-ca-gsx9l\" (UID: \"aba5b6b9-f88d-4763-9562-6b4dec606df9\") " pod="openshift-image-registry/node-ca-gsx9l" Dec 11 08:11:30 crc kubenswrapper[4860]: I1211 08:11:30.726360 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:30 crc kubenswrapper[4860]: I1211 08:11:30.726388 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:30 crc kubenswrapper[4860]: I1211 08:11:30.726398 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:30 crc kubenswrapper[4860]: I1211 08:11:30.726414 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:30 crc kubenswrapper[4860]: I1211 08:11:30.726425 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:30Z","lastTransitionTime":"2025-12-11T08:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:30 crc kubenswrapper[4860]: I1211 08:11:30.741415 4860 generic.go:334] "Generic (PLEG): container finished" podID="a1faa733-f408-41fd-b7aa-0cd86283976a" containerID="8ac45db4f858a1c18a6b87472fc1cea2300f5845f1ad62c22260374995548a0b" exitCode=0 Dec 11 08:11:30 crc kubenswrapper[4860]: I1211 08:11:30.742377 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-26nc6" event={"ID":"a1faa733-f408-41fd-b7aa-0cd86283976a","Type":"ContainerDied","Data":"8ac45db4f858a1c18a6b87472fc1cea2300f5845f1ad62c22260374995548a0b"} Dec 11 08:11:30 crc kubenswrapper[4860]: I1211 08:11:30.772358 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9872be1e-b995-4a00-904b-875dff96ad56\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f74a32a8fefea704cc73ffd1114083dab68aa44dc33f8d9d432d2d24216546c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbe00b46063d3968d59c393b87178e135f982abb7d0f32fa798b212d4b83df6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eedc9988e1a4a3c8cae730dc4a352103a6d665535b825e157d0f87c331b137c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2137fee4b7abc9765f9517d7b3b6ea796601a31d03151bf167d17136ced8b25a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f683e48f3577a68862eb6302383538376e41f49b20b1e0ea3f1b3b143f24872a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5de794187456b3576b969069a3b839d75c7918418a97fa35dff8bb40bb76e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5de794187456b3576b969069a3b839d75c7918418a97fa35dff8bb40bb76e39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08ce42b974fba00c409c0af971f0de351e50f18b2f06f8e722150da3239bb908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08ce42b974fba00c409c0af971f0de351e50f18b2f06f8e722150da3239bb908\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cfac6be03e3cfa0b0f3450b48b67957f2d5172b9c65e27f14c2a260bda5d423f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cfac6be03e3cfa0b0f3450b48b67957f2d5172b9c65e27f14c2a260bda5d423f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:30Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:30 crc kubenswrapper[4860]: I1211 08:11:30.792120 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-pznwl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5310c1fc-66c6-40aa-b1bf-5a59a2410f9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2d6ff00ce322a28607bc3f6db05c5e4fb14acb6777ea315786ef53963ca504a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hkt48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-pznwl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:30Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:30 crc kubenswrapper[4860]: I1211 08:11:30.825737 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6zw9s\" (UniqueName: \"kubernetes.io/projected/aba5b6b9-f88d-4763-9562-6b4dec606df9-kube-api-access-6zw9s\") pod \"node-ca-gsx9l\" (UID: \"aba5b6b9-f88d-4763-9562-6b4dec606df9\") " pod="openshift-image-registry/node-ca-gsx9l" Dec 11 08:11:30 crc kubenswrapper[4860]: I1211 08:11:30.825907 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/aba5b6b9-f88d-4763-9562-6b4dec606df9-serviceca\") pod \"node-ca-gsx9l\" (UID: \"aba5b6b9-f88d-4763-9562-6b4dec606df9\") " pod="openshift-image-registry/node-ca-gsx9l" Dec 11 08:11:30 crc kubenswrapper[4860]: I1211 08:11:30.826001 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/aba5b6b9-f88d-4763-9562-6b4dec606df9-host\") pod \"node-ca-gsx9l\" (UID: \"aba5b6b9-f88d-4763-9562-6b4dec606df9\") " pod="openshift-image-registry/node-ca-gsx9l" Dec 11 08:11:30 crc kubenswrapper[4860]: I1211 08:11:30.826072 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/aba5b6b9-f88d-4763-9562-6b4dec606df9-host\") pod \"node-ca-gsx9l\" (UID: \"aba5b6b9-f88d-4763-9562-6b4dec606df9\") " pod="openshift-image-registry/node-ca-gsx9l" Dec 11 08:11:30 crc kubenswrapper[4860]: I1211 08:11:30.829264 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/aba5b6b9-f88d-4763-9562-6b4dec606df9-serviceca\") pod \"node-ca-gsx9l\" (UID: \"aba5b6b9-f88d-4763-9562-6b4dec606df9\") " pod="openshift-image-registry/node-ca-gsx9l" Dec 11 08:11:30 crc kubenswrapper[4860]: I1211 08:11:30.835685 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:30Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:30 crc kubenswrapper[4860]: I1211 08:11:30.836721 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:30 crc kubenswrapper[4860]: I1211 08:11:30.836743 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:30 crc kubenswrapper[4860]: I1211 08:11:30.836750 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:30 crc kubenswrapper[4860]: I1211 08:11:30.836762 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:30 crc kubenswrapper[4860]: I1211 08:11:30.836769 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:30Z","lastTransitionTime":"2025-12-11T08:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:30 crc kubenswrapper[4860]: I1211 08:11:30.871453 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6zw9s\" (UniqueName: \"kubernetes.io/projected/aba5b6b9-f88d-4763-9562-6b4dec606df9-kube-api-access-6zw9s\") pod \"node-ca-gsx9l\" (UID: \"aba5b6b9-f88d-4763-9562-6b4dec606df9\") " pod="openshift-image-registry/node-ca-gsx9l" Dec 11 08:11:30 crc kubenswrapper[4860]: I1211 08:11:30.885192 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-gsx9l" Dec 11 08:11:30 crc kubenswrapper[4860]: I1211 08:11:30.893409 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d8c2f28628dc2f1ffbdd98a55670e22cf6cdc8cf2c6442e9d1f8e5466b2f12e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://074ea63beac52bd196507ee7cfb1c3d7fc2d37d5ec6faa4d47dde62b922e1c43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:30Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:30 crc kubenswrapper[4860]: I1211 08:11:30.934714 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:30Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:30 crc kubenswrapper[4860]: I1211 08:11:30.939046 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:30 crc kubenswrapper[4860]: I1211 08:11:30.939072 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:30 crc kubenswrapper[4860]: I1211 08:11:30.939082 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:30 crc kubenswrapper[4860]: I1211 08:11:30.939096 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:30 crc kubenswrapper[4860]: I1211 08:11:30.939113 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:30Z","lastTransitionTime":"2025-12-11T08:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:30 crc kubenswrapper[4860]: I1211 08:11:30.970671 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25a6a154da2b74952951ae94efc914ba078ea4bbe5893360b3c84d46be6bf973\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:30Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:31 crc kubenswrapper[4860]: I1211 08:11:31.009683 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6xq7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a75ce3c-5e1e-4b9f-bbde-7a3b0cce5eae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe8d4d675567b41728e476351811a93c47b19868ccbadab8c613a4eb8eca8562\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5c6nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6xq7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:31Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:31 crc kubenswrapper[4860]: I1211 08:11:31.041356 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:31 crc kubenswrapper[4860]: I1211 08:11:31.041568 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:31 crc kubenswrapper[4860]: I1211 08:11:31.041627 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:31 crc kubenswrapper[4860]: I1211 08:11:31.041715 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:31 crc kubenswrapper[4860]: I1211 08:11:31.041776 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:31Z","lastTransitionTime":"2025-12-11T08:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:31 crc kubenswrapper[4860]: I1211 08:11:31.061446 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ec3758-85f1-4a22-a4ec-c805472607a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e831db1a15976d3967d56479689572c26f2599a766baa0402897a3008428c56f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a498aa90f914e1f0370df01dc5ef174df113a2a16d86b2e3f24b19a1b99b4c0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8caebd2f4dfcbb89c597029a07c335171889761d2d200206987537fb0a228db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34e91f615b76445ff3241a304bb2964fde59de7e8a8dfff6841b2e628a31e86d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2090cab02ad74ddfd6bb57ff50eba35d18f56b0cde1c230fe011c01994b693bc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 08:11:21.171011 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 08:11:21.177674 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-644352688/tls.crt::/tmp/serving-cert-644352688/tls.key\\\\\\\"\\\\nI1211 08:11:26.508606 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 08:11:26.512741 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 08:11:26.512778 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 08:11:26.512839 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 08:11:26.512858 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 08:11:26.521684 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1211 08:11:26.521700 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1211 08:11:26.521722 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:11:26.521732 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:11:26.521738 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 08:11:26.521746 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 08:11:26.521749 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 08:11:26.521753 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1211 08:11:26.524999 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2274fb298ca7d2b37b060b4dbf9b166d39ba77a108b08edeeb4d67196e72cb3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf0aff8a6ab816b6617b1feeca9146d41bb5e8808bf1db96f8d493500c135d85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf0aff8a6ab816b6617b1feeca9146d41bb5e8808bf1db96f8d493500c135d85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:31Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:31 crc kubenswrapper[4860]: I1211 08:11:31.090769 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:31Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:31 crc kubenswrapper[4860]: I1211 08:11:31.134887 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a41f35db087418d1be16953f10377042abd3f4488e5ac1bc29ca5dcfc00a37e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:31Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:31 crc kubenswrapper[4860]: I1211 08:11:31.144514 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:31 crc kubenswrapper[4860]: I1211 08:11:31.144559 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:31 crc kubenswrapper[4860]: I1211 08:11:31.144575 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:31 crc kubenswrapper[4860]: I1211 08:11:31.144596 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:31 crc kubenswrapper[4860]: I1211 08:11:31.144612 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:31Z","lastTransitionTime":"2025-12-11T08:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:31 crc kubenswrapper[4860]: I1211 08:11:31.171071 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6xq7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a75ce3c-5e1e-4b9f-bbde-7a3b0cce5eae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe8d4d675567b41728e476351811a93c47b19868ccbadab8c613a4eb8eca8562\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5c6nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6xq7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:31Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:31 crc kubenswrapper[4860]: I1211 08:11:31.215034 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:31Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:31 crc kubenswrapper[4860]: I1211 08:11:31.246878 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:31 crc kubenswrapper[4860]: I1211 08:11:31.246921 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:31 crc kubenswrapper[4860]: I1211 08:11:31.246931 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:31 crc kubenswrapper[4860]: I1211 08:11:31.246946 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:31 crc kubenswrapper[4860]: I1211 08:11:31.246957 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:31Z","lastTransitionTime":"2025-12-11T08:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:31 crc kubenswrapper[4860]: I1211 08:11:31.254078 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d8c2f28628dc2f1ffbdd98a55670e22cf6cdc8cf2c6442e9d1f8e5466b2f12e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://074ea63beac52bd196507ee7cfb1c3d7fc2d37d5ec6faa4d47dde62b922e1c43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:31Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:31 crc kubenswrapper[4860]: I1211 08:11:31.301838 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:31Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:31 crc kubenswrapper[4860]: I1211 08:11:31.331445 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25a6a154da2b74952951ae94efc914ba078ea4bbe5893360b3c84d46be6bf973\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:31Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:31 crc kubenswrapper[4860]: I1211 08:11:31.351664 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:31 crc kubenswrapper[4860]: I1211 08:11:31.351720 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:31 crc kubenswrapper[4860]: I1211 08:11:31.351732 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:31 crc kubenswrapper[4860]: I1211 08:11:31.351755 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:31 crc kubenswrapper[4860]: I1211 08:11:31.351772 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:31Z","lastTransitionTime":"2025-12-11T08:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:31 crc kubenswrapper[4860]: I1211 08:11:31.387945 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ec3758-85f1-4a22-a4ec-c805472607a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e831db1a15976d3967d56479689572c26f2599a766baa0402897a3008428c56f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a498aa90f914e1f0370df01dc5ef174df113a2a16d86b2e3f24b19a1b99b4c0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8caebd2f4dfcbb89c597029a07c335171889761d2d200206987537fb0a228db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34e91f615b76445ff3241a304bb2964fde59de7e8a8dfff6841b2e628a31e86d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2090cab02ad74ddfd6bb57ff50eba35d18f56b0cde1c230fe011c01994b693bc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 08:11:21.171011 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 08:11:21.177674 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-644352688/tls.crt::/tmp/serving-cert-644352688/tls.key\\\\\\\"\\\\nI1211 08:11:26.508606 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 08:11:26.512741 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 08:11:26.512778 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 08:11:26.512839 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 08:11:26.512858 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 08:11:26.521684 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1211 08:11:26.521700 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1211 08:11:26.521722 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:11:26.521732 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:11:26.521738 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 08:11:26.521746 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 08:11:26.521749 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 08:11:26.521753 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1211 08:11:26.524999 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2274fb298ca7d2b37b060b4dbf9b166d39ba77a108b08edeeb4d67196e72cb3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf0aff8a6ab816b6617b1feeca9146d41bb5e8808bf1db96f8d493500c135d85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf0aff8a6ab816b6617b1feeca9146d41bb5e8808bf1db96f8d493500c135d85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:31Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:31 crc kubenswrapper[4860]: I1211 08:11:31.419048 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:31Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:31 crc kubenswrapper[4860]: I1211 08:11:31.453603 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:31 crc kubenswrapper[4860]: I1211 08:11:31.453635 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:31 crc kubenswrapper[4860]: I1211 08:11:31.453654 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:31 crc kubenswrapper[4860]: I1211 08:11:31.453667 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:31 crc kubenswrapper[4860]: I1211 08:11:31.453676 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:31Z","lastTransitionTime":"2025-12-11T08:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:31 crc kubenswrapper[4860]: I1211 08:11:31.458185 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a41f35db087418d1be16953f10377042abd3f4488e5ac1bc29ca5dcfc00a37e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:31Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:31 crc kubenswrapper[4860]: I1211 08:11:31.497996 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gsx9l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aba5b6b9-f88d-4763-9562-6b4dec606df9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6zw9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gsx9l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:31Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:31 crc kubenswrapper[4860]: I1211 08:11:31.537939 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9aa0aed-62b3-4ea2-82ae-9a653c1101a5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ad4ba1b769c6cb2b34073798f8e3c25eb0177e6d3059484bf365e6981cab710\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://077aa1dd2fe5189ff60a0f2365e47a6a545cca0de7aa6379370908db4ea5838a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d7c919899b74a2b0fcd9221f3f769a9d013267354b200a282669d342d6413d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5483384f1d2ebc1fefa9706c9e67a3dc5a00ff643f803e3f2707c252e27ad94\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:31Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:31 crc kubenswrapper[4860]: I1211 08:11:31.556408 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:31 crc kubenswrapper[4860]: I1211 08:11:31.556452 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:31 crc kubenswrapper[4860]: I1211 08:11:31.556464 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:31 crc kubenswrapper[4860]: I1211 08:11:31.556482 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:31 crc kubenswrapper[4860]: I1211 08:11:31.556493 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:31Z","lastTransitionTime":"2025-12-11T08:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:31 crc kubenswrapper[4860]: I1211 08:11:31.572679 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31c30642-6e60-41b4-a477-0d802424e0aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79969402f5c33852a26fc839f37a45febf4f2377d77823284fbb01769e5585fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2d4d5de7106175087d91acb1f0c7552014f0c08bfcd65cbbf184d3f8d7826ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-99qgp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:31Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:31 crc kubenswrapper[4860]: I1211 08:11:31.613891 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-26nc6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1faa733-f408-41fd-b7aa-0cd86283976a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ac45db4f858a1c18a6b87472fc1cea2300f5845f1ad62c22260374995548a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ac45db4f858a1c18a6b87472fc1cea2300f5845f1ad62c22260374995548a0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-26nc6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:31Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:31 crc kubenswrapper[4860]: I1211 08:11:31.657264 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0589a204-c98c-417d-8256-bfe3bf747660\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17cbcb49c06f3a9059cd5a0f05eedb8c785da30a3e502e31bae2f57b3486cb14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17cbcb49c06f3a9059cd5a0f05eedb8c785da30a3e502e31bae2f57b3486cb14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xvnqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:31Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:31 crc kubenswrapper[4860]: I1211 08:11:31.658993 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:31 crc kubenswrapper[4860]: I1211 08:11:31.659035 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:31 crc kubenswrapper[4860]: I1211 08:11:31.659044 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:31 crc kubenswrapper[4860]: I1211 08:11:31.659058 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:31 crc kubenswrapper[4860]: I1211 08:11:31.659072 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:31Z","lastTransitionTime":"2025-12-11T08:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:31 crc kubenswrapper[4860]: I1211 08:11:31.697720 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9872be1e-b995-4a00-904b-875dff96ad56\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f74a32a8fefea704cc73ffd1114083dab68aa44dc33f8d9d432d2d24216546c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbe00b46063d3968d59c393b87178e135f982abb7d0f32fa798b212d4b83df6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eedc9988e1a4a3c8cae730dc4a352103a6d665535b825e157d0f87c331b137c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2137fee4b7abc9765f9517d7b3b6ea796601a31d03151bf167d17136ced8b25a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f683e48f3577a68862eb6302383538376e41f49b20b1e0ea3f1b3b143f24872a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5de794187456b3576b969069a3b839d75c7918418a97fa35dff8bb40bb76e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5de794187456b3576b969069a3b839d75c7918418a97fa35dff8bb40bb76e39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08ce42b974fba00c409c0af971f0de351e50f18b2f06f8e722150da3239bb908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08ce42b974fba00c409c0af971f0de351e50f18b2f06f8e722150da3239bb908\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cfac6be03e3cfa0b0f3450b48b67957f2d5172b9c65e27f14c2a260bda5d423f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cfac6be03e3cfa0b0f3450b48b67957f2d5172b9c65e27f14c2a260bda5d423f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:31Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:31 crc kubenswrapper[4860]: I1211 08:11:31.734681 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-pznwl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5310c1fc-66c6-40aa-b1bf-5a59a2410f9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2d6ff00ce322a28607bc3f6db05c5e4fb14acb6777ea315786ef53963ca504a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hkt48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-pznwl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:31Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:31 crc kubenswrapper[4860]: I1211 08:11:31.749108 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-26nc6" event={"ID":"a1faa733-f408-41fd-b7aa-0cd86283976a","Type":"ContainerStarted","Data":"c6b35296a5eb17dd2bb5e72c86c1cafdb30b7712cb330264f935d3b4fff05cd1"} Dec 11 08:11:31 crc kubenswrapper[4860]: I1211 08:11:31.757013 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-gsx9l" event={"ID":"aba5b6b9-f88d-4763-9562-6b4dec606df9","Type":"ContainerStarted","Data":"735ef012c523ddc952d46a431f27b2582a2c301d9e9da0c697a20ff6e2fb5d86"} Dec 11 08:11:31 crc kubenswrapper[4860]: I1211 08:11:31.757069 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-gsx9l" event={"ID":"aba5b6b9-f88d-4763-9562-6b4dec606df9","Type":"ContainerStarted","Data":"3b3d9c7c0eb7abaac2f563d3033ba47bbc3e2d38b00fd78c25463caff4f178a4"} Dec 11 08:11:31 crc kubenswrapper[4860]: I1211 08:11:31.761887 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:31 crc kubenswrapper[4860]: I1211 08:11:31.761926 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:31 crc kubenswrapper[4860]: I1211 08:11:31.761941 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:31 crc kubenswrapper[4860]: I1211 08:11:31.761957 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:31 crc kubenswrapper[4860]: I1211 08:11:31.761968 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:31Z","lastTransitionTime":"2025-12-11T08:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:31 crc kubenswrapper[4860]: I1211 08:11:31.763242 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" event={"ID":"0589a204-c98c-417d-8256-bfe3bf747660","Type":"ContainerStarted","Data":"3333b4346f59b04017e32d61f457ba7ff457166731128f339d7987ae6b56b06d"} Dec 11 08:11:31 crc kubenswrapper[4860]: I1211 08:11:31.763278 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" event={"ID":"0589a204-c98c-417d-8256-bfe3bf747660","Type":"ContainerStarted","Data":"1426084ac580895c42354630910dd8c8e3a48ddd25f22abce1d28dc3de332a63"} Dec 11 08:11:31 crc kubenswrapper[4860]: I1211 08:11:31.763299 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" event={"ID":"0589a204-c98c-417d-8256-bfe3bf747660","Type":"ContainerStarted","Data":"c9984d5d242bb258405efe91fc9caf9e7ea592834445df2e78add9553a921c76"} Dec 11 08:11:31 crc kubenswrapper[4860]: I1211 08:11:31.777257 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9872be1e-b995-4a00-904b-875dff96ad56\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f74a32a8fefea704cc73ffd1114083dab68aa44dc33f8d9d432d2d24216546c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbe00b46063d3968d59c393b87178e135f982abb7d0f32fa798b212d4b83df6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eedc9988e1a4a3c8cae730dc4a352103a6d665535b825e157d0f87c331b137c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2137fee4b7abc9765f9517d7b3b6ea796601a31d03151bf167d17136ced8b25a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f683e48f3577a68862eb6302383538376e41f49b20b1e0ea3f1b3b143f24872a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5de794187456b3576b969069a3b839d75c7918418a97fa35dff8bb40bb76e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5de794187456b3576b969069a3b839d75c7918418a97fa35dff8bb40bb76e39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08ce42b974fba00c409c0af971f0de351e50f18b2f06f8e722150da3239bb908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08ce42b974fba00c409c0af971f0de351e50f18b2f06f8e722150da3239bb908\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cfac6be03e3cfa0b0f3450b48b67957f2d5172b9c65e27f14c2a260bda5d423f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cfac6be03e3cfa0b0f3450b48b67957f2d5172b9c65e27f14c2a260bda5d423f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:31Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:31 crc kubenswrapper[4860]: I1211 08:11:31.814410 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-pznwl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5310c1fc-66c6-40aa-b1bf-5a59a2410f9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2d6ff00ce322a28607bc3f6db05c5e4fb14acb6777ea315786ef53963ca504a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hkt48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-pznwl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:31Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:31 crc kubenswrapper[4860]: I1211 08:11:31.851401 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d8c2f28628dc2f1ffbdd98a55670e22cf6cdc8cf2c6442e9d1f8e5466b2f12e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://074ea63beac52bd196507ee7cfb1c3d7fc2d37d5ec6faa4d47dde62b922e1c43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:31Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:31 crc kubenswrapper[4860]: I1211 08:11:31.864216 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:31 crc kubenswrapper[4860]: I1211 08:11:31.864241 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:31 crc kubenswrapper[4860]: I1211 08:11:31.864249 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:31 crc kubenswrapper[4860]: I1211 08:11:31.864261 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:31 crc kubenswrapper[4860]: I1211 08:11:31.864270 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:31Z","lastTransitionTime":"2025-12-11T08:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:31 crc kubenswrapper[4860]: I1211 08:11:31.891445 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:31Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:31 crc kubenswrapper[4860]: I1211 08:11:31.932051 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25a6a154da2b74952951ae94efc914ba078ea4bbe5893360b3c84d46be6bf973\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:31Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:31 crc kubenswrapper[4860]: I1211 08:11:31.966051 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:31 crc kubenswrapper[4860]: I1211 08:11:31.966085 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:31 crc kubenswrapper[4860]: I1211 08:11:31.966095 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:31 crc kubenswrapper[4860]: I1211 08:11:31.966109 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:31 crc kubenswrapper[4860]: I1211 08:11:31.966119 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:31Z","lastTransitionTime":"2025-12-11T08:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:31 crc kubenswrapper[4860]: I1211 08:11:31.974214 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6xq7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a75ce3c-5e1e-4b9f-bbde-7a3b0cce5eae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe8d4d675567b41728e476351811a93c47b19868ccbadab8c613a4eb8eca8562\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5c6nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6xq7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:31Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:32 crc kubenswrapper[4860]: I1211 08:11:32.017074 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:32Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:32 crc kubenswrapper[4860]: I1211 08:11:32.060297 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a41f35db087418d1be16953f10377042abd3f4488e5ac1bc29ca5dcfc00a37e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:32Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:32 crc kubenswrapper[4860]: I1211 08:11:32.075685 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:32 crc kubenswrapper[4860]: I1211 08:11:32.075733 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:32 crc kubenswrapper[4860]: I1211 08:11:32.075745 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:32 crc kubenswrapper[4860]: I1211 08:11:32.075762 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:32 crc kubenswrapper[4860]: I1211 08:11:32.075772 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:32Z","lastTransitionTime":"2025-12-11T08:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:32 crc kubenswrapper[4860]: I1211 08:11:32.092526 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ec3758-85f1-4a22-a4ec-c805472607a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e831db1a15976d3967d56479689572c26f2599a766baa0402897a3008428c56f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a498aa90f914e1f0370df01dc5ef174df113a2a16d86b2e3f24b19a1b99b4c0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8caebd2f4dfcbb89c597029a07c335171889761d2d200206987537fb0a228db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34e91f615b76445ff3241a304bb2964fde59de7e8a8dfff6841b2e628a31e86d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2090cab02ad74ddfd6bb57ff50eba35d18f56b0cde1c230fe011c01994b693bc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 08:11:21.171011 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 08:11:21.177674 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-644352688/tls.crt::/tmp/serving-cert-644352688/tls.key\\\\\\\"\\\\nI1211 08:11:26.508606 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 08:11:26.512741 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 08:11:26.512778 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 08:11:26.512839 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 08:11:26.512858 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 08:11:26.521684 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1211 08:11:26.521700 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1211 08:11:26.521722 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:11:26.521732 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:11:26.521738 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 08:11:26.521746 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 08:11:26.521749 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 08:11:26.521753 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1211 08:11:26.524999 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2274fb298ca7d2b37b060b4dbf9b166d39ba77a108b08edeeb4d67196e72cb3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf0aff8a6ab816b6617b1feeca9146d41bb5e8808bf1db96f8d493500c135d85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf0aff8a6ab816b6617b1feeca9146d41bb5e8808bf1db96f8d493500c135d85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:32Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:32 crc kubenswrapper[4860]: I1211 08:11:32.131863 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:32Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:32 crc kubenswrapper[4860]: I1211 08:11:32.175293 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31c30642-6e60-41b4-a477-0d802424e0aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79969402f5c33852a26fc839f37a45febf4f2377d77823284fbb01769e5585fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2d4d5de7106175087d91acb1f0c7552014f0c08bfcd65cbbf184d3f8d7826ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-99qgp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:32Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:32 crc kubenswrapper[4860]: I1211 08:11:32.185127 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:32 crc kubenswrapper[4860]: I1211 08:11:32.185158 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:32 crc kubenswrapper[4860]: I1211 08:11:32.185173 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:32 crc kubenswrapper[4860]: I1211 08:11:32.185193 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:32 crc kubenswrapper[4860]: I1211 08:11:32.185204 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:32Z","lastTransitionTime":"2025-12-11T08:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:32 crc kubenswrapper[4860]: I1211 08:11:32.216720 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-26nc6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1faa733-f408-41fd-b7aa-0cd86283976a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ac45db4f858a1c18a6b87472fc1cea2300f5845f1ad62c22260374995548a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ac45db4f858a1c18a6b87472fc1cea2300f5845f1ad62c22260374995548a0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b35296a5eb17dd2bb5e72c86c1cafdb30b7712cb330264f935d3b4fff05cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-26nc6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:32Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:32 crc kubenswrapper[4860]: I1211 08:11:32.256853 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0589a204-c98c-417d-8256-bfe3bf747660\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17cbcb49c06f3a9059cd5a0f05eedb8c785da30a3e502e31bae2f57b3486cb14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17cbcb49c06f3a9059cd5a0f05eedb8c785da30a3e502e31bae2f57b3486cb14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xvnqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:32Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:32 crc kubenswrapper[4860]: I1211 08:11:32.287582 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:32 crc kubenswrapper[4860]: I1211 08:11:32.287608 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:32 crc kubenswrapper[4860]: I1211 08:11:32.287615 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:32 crc kubenswrapper[4860]: I1211 08:11:32.287628 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:32 crc kubenswrapper[4860]: I1211 08:11:32.287636 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:32Z","lastTransitionTime":"2025-12-11T08:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:32 crc kubenswrapper[4860]: I1211 08:11:32.290844 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gsx9l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aba5b6b9-f88d-4763-9562-6b4dec606df9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6zw9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gsx9l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:32Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:32 crc kubenswrapper[4860]: I1211 08:11:32.338051 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9aa0aed-62b3-4ea2-82ae-9a653c1101a5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ad4ba1b769c6cb2b34073798f8e3c25eb0177e6d3059484bf365e6981cab710\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://077aa1dd2fe5189ff60a0f2365e47a6a545cca0de7aa6379370908db4ea5838a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d7c919899b74a2b0fcd9221f3f769a9d013267354b200a282669d342d6413d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5483384f1d2ebc1fefa9706c9e67a3dc5a00ff643f803e3f2707c252e27ad94\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:32Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:32 crc kubenswrapper[4860]: I1211 08:11:32.375140 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-pznwl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5310c1fc-66c6-40aa-b1bf-5a59a2410f9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2d6ff00ce322a28607bc3f6db05c5e4fb14acb6777ea315786ef53963ca504a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hkt48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-pznwl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:32Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:32 crc kubenswrapper[4860]: I1211 08:11:32.389806 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:32 crc kubenswrapper[4860]: I1211 08:11:32.389837 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:32 crc kubenswrapper[4860]: I1211 08:11:32.389847 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:32 crc kubenswrapper[4860]: I1211 08:11:32.389864 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:32 crc kubenswrapper[4860]: I1211 08:11:32.389876 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:32Z","lastTransitionTime":"2025-12-11T08:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:32 crc kubenswrapper[4860]: I1211 08:11:32.425701 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9872be1e-b995-4a00-904b-875dff96ad56\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f74a32a8fefea704cc73ffd1114083dab68aa44dc33f8d9d432d2d24216546c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbe00b46063d3968d59c393b87178e135f982abb7d0f32fa798b212d4b83df6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eedc9988e1a4a3c8cae730dc4a352103a6d665535b825e157d0f87c331b137c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2137fee4b7abc9765f9517d7b3b6ea796601a31d03151bf167d17136ced8b25a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f683e48f3577a68862eb6302383538376e41f49b20b1e0ea3f1b3b143f24872a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5de794187456b3576b969069a3b839d75c7918418a97fa35dff8bb40bb76e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5de794187456b3576b969069a3b839d75c7918418a97fa35dff8bb40bb76e39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08ce42b974fba00c409c0af971f0de351e50f18b2f06f8e722150da3239bb908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08ce42b974fba00c409c0af971f0de351e50f18b2f06f8e722150da3239bb908\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cfac6be03e3cfa0b0f3450b48b67957f2d5172b9c65e27f14c2a260bda5d423f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cfac6be03e3cfa0b0f3450b48b67957f2d5172b9c65e27f14c2a260bda5d423f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:32Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:32 crc kubenswrapper[4860]: I1211 08:11:32.453187 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:32Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:32 crc kubenswrapper[4860]: I1211 08:11:32.491885 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d8c2f28628dc2f1ffbdd98a55670e22cf6cdc8cf2c6442e9d1f8e5466b2f12e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://074ea63beac52bd196507ee7cfb1c3d7fc2d37d5ec6faa4d47dde62b922e1c43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:32Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:32 crc kubenswrapper[4860]: I1211 08:11:32.493561 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:32 crc kubenswrapper[4860]: I1211 08:11:32.493602 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:32 crc kubenswrapper[4860]: I1211 08:11:32.493613 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:32 crc kubenswrapper[4860]: I1211 08:11:32.493626 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:32 crc kubenswrapper[4860]: I1211 08:11:32.493635 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:32Z","lastTransitionTime":"2025-12-11T08:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:32 crc kubenswrapper[4860]: I1211 08:11:32.530504 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:32Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:32 crc kubenswrapper[4860]: I1211 08:11:32.573913 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25a6a154da2b74952951ae94efc914ba078ea4bbe5893360b3c84d46be6bf973\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:32Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:32 crc kubenswrapper[4860]: I1211 08:11:32.578083 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:11:32 crc kubenswrapper[4860]: I1211 08:11:32.578118 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:11:32 crc kubenswrapper[4860]: I1211 08:11:32.578142 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:11:32 crc kubenswrapper[4860]: E1211 08:11:32.578227 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:11:32 crc kubenswrapper[4860]: E1211 08:11:32.578305 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:11:32 crc kubenswrapper[4860]: E1211 08:11:32.578407 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:11:32 crc kubenswrapper[4860]: I1211 08:11:32.596149 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:32 crc kubenswrapper[4860]: I1211 08:11:32.596183 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:32 crc kubenswrapper[4860]: I1211 08:11:32.596191 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:32 crc kubenswrapper[4860]: I1211 08:11:32.596206 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:32 crc kubenswrapper[4860]: I1211 08:11:32.596214 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:32Z","lastTransitionTime":"2025-12-11T08:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:32 crc kubenswrapper[4860]: I1211 08:11:32.610612 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6xq7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a75ce3c-5e1e-4b9f-bbde-7a3b0cce5eae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe8d4d675567b41728e476351811a93c47b19868ccbadab8c613a4eb8eca8562\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5c6nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6xq7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:32Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:32 crc kubenswrapper[4860]: I1211 08:11:32.652051 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:32Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:32 crc kubenswrapper[4860]: I1211 08:11:32.694292 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a41f35db087418d1be16953f10377042abd3f4488e5ac1bc29ca5dcfc00a37e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:32Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:32 crc kubenswrapper[4860]: I1211 08:11:32.699068 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:32 crc kubenswrapper[4860]: I1211 08:11:32.699095 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:32 crc kubenswrapper[4860]: I1211 08:11:32.699103 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:32 crc kubenswrapper[4860]: I1211 08:11:32.699118 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:32 crc kubenswrapper[4860]: I1211 08:11:32.699126 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:32Z","lastTransitionTime":"2025-12-11T08:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:32 crc kubenswrapper[4860]: I1211 08:11:32.731905 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ec3758-85f1-4a22-a4ec-c805472607a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e831db1a15976d3967d56479689572c26f2599a766baa0402897a3008428c56f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a498aa90f914e1f0370df01dc5ef174df113a2a16d86b2e3f24b19a1b99b4c0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8caebd2f4dfcbb89c597029a07c335171889761d2d200206987537fb0a228db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34e91f615b76445ff3241a304bb2964fde59de7e8a8dfff6841b2e628a31e86d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2090cab02ad74ddfd6bb57ff50eba35d18f56b0cde1c230fe011c01994b693bc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 08:11:21.171011 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 08:11:21.177674 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-644352688/tls.crt::/tmp/serving-cert-644352688/tls.key\\\\\\\"\\\\nI1211 08:11:26.508606 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 08:11:26.512741 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 08:11:26.512778 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 08:11:26.512839 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 08:11:26.512858 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 08:11:26.521684 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1211 08:11:26.521700 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1211 08:11:26.521722 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:11:26.521732 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:11:26.521738 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 08:11:26.521746 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 08:11:26.521749 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 08:11:26.521753 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1211 08:11:26.524999 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2274fb298ca7d2b37b060b4dbf9b166d39ba77a108b08edeeb4d67196e72cb3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf0aff8a6ab816b6617b1feeca9146d41bb5e8808bf1db96f8d493500c135d85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf0aff8a6ab816b6617b1feeca9146d41bb5e8808bf1db96f8d493500c135d85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:32Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:32 crc kubenswrapper[4860]: I1211 08:11:32.770900 4860 generic.go:334] "Generic (PLEG): container finished" podID="a1faa733-f408-41fd-b7aa-0cd86283976a" containerID="c6b35296a5eb17dd2bb5e72c86c1cafdb30b7712cb330264f935d3b4fff05cd1" exitCode=0 Dec 11 08:11:32 crc kubenswrapper[4860]: I1211 08:11:32.770955 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-26nc6" event={"ID":"a1faa733-f408-41fd-b7aa-0cd86283976a","Type":"ContainerDied","Data":"c6b35296a5eb17dd2bb5e72c86c1cafdb30b7712cb330264f935d3b4fff05cd1"} Dec 11 08:11:32 crc kubenswrapper[4860]: I1211 08:11:32.774902 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" event={"ID":"0589a204-c98c-417d-8256-bfe3bf747660","Type":"ContainerStarted","Data":"6dfcc4370edaae36a5a32b0a9c29290c50f8e4f427e4e147dc3c779beaa8b780"} Dec 11 08:11:32 crc kubenswrapper[4860]: I1211 08:11:32.774941 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" event={"ID":"0589a204-c98c-417d-8256-bfe3bf747660","Type":"ContainerStarted","Data":"d856ae95e9077524c518d82d10fb6b93b7fbcaaf2b720ffafe6f7fae58bd7caa"} Dec 11 08:11:32 crc kubenswrapper[4860]: I1211 08:11:32.774957 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" event={"ID":"0589a204-c98c-417d-8256-bfe3bf747660","Type":"ContainerStarted","Data":"52bb8ade314afb9494f4e026d0af2aea7b1970d64dc66102f55e257ae4678154"} Dec 11 08:11:32 crc kubenswrapper[4860]: I1211 08:11:32.776035 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9aa0aed-62b3-4ea2-82ae-9a653c1101a5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ad4ba1b769c6cb2b34073798f8e3c25eb0177e6d3059484bf365e6981cab710\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://077aa1dd2fe5189ff60a0f2365e47a6a545cca0de7aa6379370908db4ea5838a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d7c919899b74a2b0fcd9221f3f769a9d013267354b200a282669d342d6413d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5483384f1d2ebc1fefa9706c9e67a3dc5a00ff643f803e3f2707c252e27ad94\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:32Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:32 crc kubenswrapper[4860]: I1211 08:11:32.802134 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:32 crc kubenswrapper[4860]: I1211 08:11:32.802174 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:32 crc kubenswrapper[4860]: I1211 08:11:32.802183 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:32 crc kubenswrapper[4860]: I1211 08:11:32.802199 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:32 crc kubenswrapper[4860]: I1211 08:11:32.802208 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:32Z","lastTransitionTime":"2025-12-11T08:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:32 crc kubenswrapper[4860]: I1211 08:11:32.810165 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31c30642-6e60-41b4-a477-0d802424e0aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79969402f5c33852a26fc839f37a45febf4f2377d77823284fbb01769e5585fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2d4d5de7106175087d91acb1f0c7552014f0c08bfcd65cbbf184d3f8d7826ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-99qgp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:32Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:32 crc kubenswrapper[4860]: I1211 08:11:32.852511 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-26nc6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1faa733-f408-41fd-b7aa-0cd86283976a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ac45db4f858a1c18a6b87472fc1cea2300f5845f1ad62c22260374995548a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ac45db4f858a1c18a6b87472fc1cea2300f5845f1ad62c22260374995548a0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b35296a5eb17dd2bb5e72c86c1cafdb30b7712cb330264f935d3b4fff05cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-26nc6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:32Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:32 crc kubenswrapper[4860]: I1211 08:11:32.907097 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:32 crc kubenswrapper[4860]: I1211 08:11:32.907130 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:32 crc kubenswrapper[4860]: I1211 08:11:32.907138 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:32 crc kubenswrapper[4860]: I1211 08:11:32.907151 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:32 crc kubenswrapper[4860]: I1211 08:11:32.907160 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:32Z","lastTransitionTime":"2025-12-11T08:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:32 crc kubenswrapper[4860]: I1211 08:11:32.913616 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0589a204-c98c-417d-8256-bfe3bf747660\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17cbcb49c06f3a9059cd5a0f05eedb8c785da30a3e502e31bae2f57b3486cb14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17cbcb49c06f3a9059cd5a0f05eedb8c785da30a3e502e31bae2f57b3486cb14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xvnqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:32Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:32 crc kubenswrapper[4860]: I1211 08:11:32.931577 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gsx9l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aba5b6b9-f88d-4763-9562-6b4dec606df9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://735ef012c523ddc952d46a431f27b2582a2c301d9e9da0c697a20ff6e2fb5d86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6zw9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gsx9l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:32Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:32 crc kubenswrapper[4860]: I1211 08:11:32.973442 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ec3758-85f1-4a22-a4ec-c805472607a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e831db1a15976d3967d56479689572c26f2599a766baa0402897a3008428c56f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a498aa90f914e1f0370df01dc5ef174df113a2a16d86b2e3f24b19a1b99b4c0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8caebd2f4dfcbb89c597029a07c335171889761d2d200206987537fb0a228db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34e91f615b76445ff3241a304bb2964fde59de7e8a8dfff6841b2e628a31e86d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2090cab02ad74ddfd6bb57ff50eba35d18f56b0cde1c230fe011c01994b693bc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 08:11:21.171011 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 08:11:21.177674 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-644352688/tls.crt::/tmp/serving-cert-644352688/tls.key\\\\\\\"\\\\nI1211 08:11:26.508606 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 08:11:26.512741 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 08:11:26.512778 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 08:11:26.512839 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 08:11:26.512858 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 08:11:26.521684 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1211 08:11:26.521700 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1211 08:11:26.521722 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:11:26.521732 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:11:26.521738 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 08:11:26.521746 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 08:11:26.521749 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 08:11:26.521753 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1211 08:11:26.524999 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2274fb298ca7d2b37b060b4dbf9b166d39ba77a108b08edeeb4d67196e72cb3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf0aff8a6ab816b6617b1feeca9146d41bb5e8808bf1db96f8d493500c135d85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf0aff8a6ab816b6617b1feeca9146d41bb5e8808bf1db96f8d493500c135d85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:32Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:33 crc kubenswrapper[4860]: I1211 08:11:33.009553 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:33 crc kubenswrapper[4860]: I1211 08:11:33.009586 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:33 crc kubenswrapper[4860]: I1211 08:11:33.009594 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:33 crc kubenswrapper[4860]: I1211 08:11:33.009608 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:33 crc kubenswrapper[4860]: I1211 08:11:33.009617 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:33Z","lastTransitionTime":"2025-12-11T08:11:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:33 crc kubenswrapper[4860]: I1211 08:11:33.017612 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:33Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:33 crc kubenswrapper[4860]: I1211 08:11:33.052533 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a41f35db087418d1be16953f10377042abd3f4488e5ac1bc29ca5dcfc00a37e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:33Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:33 crc kubenswrapper[4860]: I1211 08:11:33.099217 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-26nc6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1faa733-f408-41fd-b7aa-0cd86283976a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ac45db4f858a1c18a6b87472fc1cea2300f5845f1ad62c22260374995548a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ac45db4f858a1c18a6b87472fc1cea2300f5845f1ad62c22260374995548a0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b35296a5eb17dd2bb5e72c86c1cafdb30b7712cb330264f935d3b4fff05cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c6b35296a5eb17dd2bb5e72c86c1cafdb30b7712cb330264f935d3b4fff05cd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-26nc6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:33Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:33 crc kubenswrapper[4860]: I1211 08:11:33.111416 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:33 crc kubenswrapper[4860]: I1211 08:11:33.111479 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:33 crc kubenswrapper[4860]: I1211 08:11:33.111498 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:33 crc kubenswrapper[4860]: I1211 08:11:33.111527 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:33 crc kubenswrapper[4860]: I1211 08:11:33.111549 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:33Z","lastTransitionTime":"2025-12-11T08:11:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:33 crc kubenswrapper[4860]: I1211 08:11:33.136623 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0589a204-c98c-417d-8256-bfe3bf747660\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17cbcb49c06f3a9059cd5a0f05eedb8c785da30a3e502e31bae2f57b3486cb14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17cbcb49c06f3a9059cd5a0f05eedb8c785da30a3e502e31bae2f57b3486cb14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xvnqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:33Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:33 crc kubenswrapper[4860]: I1211 08:11:33.172228 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gsx9l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aba5b6b9-f88d-4763-9562-6b4dec606df9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://735ef012c523ddc952d46a431f27b2582a2c301d9e9da0c697a20ff6e2fb5d86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6zw9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gsx9l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:33Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:33 crc kubenswrapper[4860]: I1211 08:11:33.213004 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9aa0aed-62b3-4ea2-82ae-9a653c1101a5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ad4ba1b769c6cb2b34073798f8e3c25eb0177e6d3059484bf365e6981cab710\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://077aa1dd2fe5189ff60a0f2365e47a6a545cca0de7aa6379370908db4ea5838a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d7c919899b74a2b0fcd9221f3f769a9d013267354b200a282669d342d6413d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5483384f1d2ebc1fefa9706c9e67a3dc5a00ff643f803e3f2707c252e27ad94\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:33Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:33 crc kubenswrapper[4860]: I1211 08:11:33.214188 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:33 crc kubenswrapper[4860]: I1211 08:11:33.214240 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:33 crc kubenswrapper[4860]: I1211 08:11:33.214251 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:33 crc kubenswrapper[4860]: I1211 08:11:33.214269 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:33 crc kubenswrapper[4860]: I1211 08:11:33.214280 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:33Z","lastTransitionTime":"2025-12-11T08:11:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:33 crc kubenswrapper[4860]: I1211 08:11:33.256278 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31c30642-6e60-41b4-a477-0d802424e0aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79969402f5c33852a26fc839f37a45febf4f2377d77823284fbb01769e5585fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2d4d5de7106175087d91acb1f0c7552014f0c08bfcd65cbbf184d3f8d7826ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-99qgp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:33Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:33 crc kubenswrapper[4860]: I1211 08:11:33.301533 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9872be1e-b995-4a00-904b-875dff96ad56\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f74a32a8fefea704cc73ffd1114083dab68aa44dc33f8d9d432d2d24216546c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbe00b46063d3968d59c393b87178e135f982abb7d0f32fa798b212d4b83df6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eedc9988e1a4a3c8cae730dc4a352103a6d665535b825e157d0f87c331b137c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2137fee4b7abc9765f9517d7b3b6ea796601a31d03151bf167d17136ced8b25a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f683e48f3577a68862eb6302383538376e41f49b20b1e0ea3f1b3b143f24872a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5de794187456b3576b969069a3b839d75c7918418a97fa35dff8bb40bb76e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5de794187456b3576b969069a3b839d75c7918418a97fa35dff8bb40bb76e39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08ce42b974fba00c409c0af971f0de351e50f18b2f06f8e722150da3239bb908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08ce42b974fba00c409c0af971f0de351e50f18b2f06f8e722150da3239bb908\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cfac6be03e3cfa0b0f3450b48b67957f2d5172b9c65e27f14c2a260bda5d423f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cfac6be03e3cfa0b0f3450b48b67957f2d5172b9c65e27f14c2a260bda5d423f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:33Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:33 crc kubenswrapper[4860]: I1211 08:11:33.316911 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:33 crc kubenswrapper[4860]: I1211 08:11:33.317001 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:33 crc kubenswrapper[4860]: I1211 08:11:33.317019 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:33 crc kubenswrapper[4860]: I1211 08:11:33.317038 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:33 crc kubenswrapper[4860]: I1211 08:11:33.317055 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:33Z","lastTransitionTime":"2025-12-11T08:11:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:33 crc kubenswrapper[4860]: I1211 08:11:33.335125 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-pznwl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5310c1fc-66c6-40aa-b1bf-5a59a2410f9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2d6ff00ce322a28607bc3f6db05c5e4fb14acb6777ea315786ef53963ca504a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hkt48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-pznwl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:33Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:33 crc kubenswrapper[4860]: I1211 08:11:33.374030 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:33Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:33 crc kubenswrapper[4860]: I1211 08:11:33.413275 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25a6a154da2b74952951ae94efc914ba078ea4bbe5893360b3c84d46be6bf973\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:33Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:33 crc kubenswrapper[4860]: I1211 08:11:33.418753 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:33 crc kubenswrapper[4860]: I1211 08:11:33.418795 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:33 crc kubenswrapper[4860]: I1211 08:11:33.418805 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:33 crc kubenswrapper[4860]: I1211 08:11:33.418822 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:33 crc kubenswrapper[4860]: I1211 08:11:33.418834 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:33Z","lastTransitionTime":"2025-12-11T08:11:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:33 crc kubenswrapper[4860]: I1211 08:11:33.453239 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6xq7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a75ce3c-5e1e-4b9f-bbde-7a3b0cce5eae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe8d4d675567b41728e476351811a93c47b19868ccbadab8c613a4eb8eca8562\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5c6nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6xq7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:33Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:33 crc kubenswrapper[4860]: I1211 08:11:33.493689 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:33Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:33 crc kubenswrapper[4860]: I1211 08:11:33.520961 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:33 crc kubenswrapper[4860]: I1211 08:11:33.521011 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:33 crc kubenswrapper[4860]: I1211 08:11:33.521060 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:33 crc kubenswrapper[4860]: I1211 08:11:33.521082 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:33 crc kubenswrapper[4860]: I1211 08:11:33.521093 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:33Z","lastTransitionTime":"2025-12-11T08:11:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:33 crc kubenswrapper[4860]: I1211 08:11:33.536778 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d8c2f28628dc2f1ffbdd98a55670e22cf6cdc8cf2c6442e9d1f8e5466b2f12e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://074ea63beac52bd196507ee7cfb1c3d7fc2d37d5ec6faa4d47dde62b922e1c43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:33Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:33 crc kubenswrapper[4860]: I1211 08:11:33.623717 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:33 crc kubenswrapper[4860]: I1211 08:11:33.623810 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:33 crc kubenswrapper[4860]: I1211 08:11:33.623833 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:33 crc kubenswrapper[4860]: I1211 08:11:33.623859 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:33 crc kubenswrapper[4860]: I1211 08:11:33.623886 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:33Z","lastTransitionTime":"2025-12-11T08:11:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:33 crc kubenswrapper[4860]: I1211 08:11:33.726982 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:33 crc kubenswrapper[4860]: I1211 08:11:33.727032 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:33 crc kubenswrapper[4860]: I1211 08:11:33.727044 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:33 crc kubenswrapper[4860]: I1211 08:11:33.727065 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:33 crc kubenswrapper[4860]: I1211 08:11:33.727082 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:33Z","lastTransitionTime":"2025-12-11T08:11:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:33 crc kubenswrapper[4860]: I1211 08:11:33.778756 4860 generic.go:334] "Generic (PLEG): container finished" podID="a1faa733-f408-41fd-b7aa-0cd86283976a" containerID="8b9e82d6d3a247309ef5bdbf5b3a86ec083add2a138b0b237801e77b09ee96a1" exitCode=0 Dec 11 08:11:33 crc kubenswrapper[4860]: I1211 08:11:33.778791 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-26nc6" event={"ID":"a1faa733-f408-41fd-b7aa-0cd86283976a","Type":"ContainerDied","Data":"8b9e82d6d3a247309ef5bdbf5b3a86ec083add2a138b0b237801e77b09ee96a1"} Dec 11 08:11:33 crc kubenswrapper[4860]: I1211 08:11:33.797762 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:33Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:33 crc kubenswrapper[4860]: I1211 08:11:33.811545 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d8c2f28628dc2f1ffbdd98a55670e22cf6cdc8cf2c6442e9d1f8e5466b2f12e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://074ea63beac52bd196507ee7cfb1c3d7fc2d37d5ec6faa4d47dde62b922e1c43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:33Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:33 crc kubenswrapper[4860]: I1211 08:11:33.826211 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:33Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:33 crc kubenswrapper[4860]: I1211 08:11:33.837701 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25a6a154da2b74952951ae94efc914ba078ea4bbe5893360b3c84d46be6bf973\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:33Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:33 crc kubenswrapper[4860]: I1211 08:11:33.838242 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:33 crc kubenswrapper[4860]: I1211 08:11:33.838326 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:33 crc kubenswrapper[4860]: I1211 08:11:33.838353 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:33 crc kubenswrapper[4860]: I1211 08:11:33.838383 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:33 crc kubenswrapper[4860]: I1211 08:11:33.838404 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:33Z","lastTransitionTime":"2025-12-11T08:11:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:33 crc kubenswrapper[4860]: I1211 08:11:33.851682 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6xq7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a75ce3c-5e1e-4b9f-bbde-7a3b0cce5eae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe8d4d675567b41728e476351811a93c47b19868ccbadab8c613a4eb8eca8562\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5c6nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6xq7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:33Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:33 crc kubenswrapper[4860]: I1211 08:11:33.866979 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ec3758-85f1-4a22-a4ec-c805472607a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e831db1a15976d3967d56479689572c26f2599a766baa0402897a3008428c56f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a498aa90f914e1f0370df01dc5ef174df113a2a16d86b2e3f24b19a1b99b4c0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8caebd2f4dfcbb89c597029a07c335171889761d2d200206987537fb0a228db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34e91f615b76445ff3241a304bb2964fde59de7e8a8dfff6841b2e628a31e86d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2090cab02ad74ddfd6bb57ff50eba35d18f56b0cde1c230fe011c01994b693bc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 08:11:21.171011 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 08:11:21.177674 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-644352688/tls.crt::/tmp/serving-cert-644352688/tls.key\\\\\\\"\\\\nI1211 08:11:26.508606 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 08:11:26.512741 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 08:11:26.512778 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 08:11:26.512839 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 08:11:26.512858 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 08:11:26.521684 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1211 08:11:26.521700 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1211 08:11:26.521722 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:11:26.521732 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:11:26.521738 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 08:11:26.521746 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 08:11:26.521749 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 08:11:26.521753 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1211 08:11:26.524999 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2274fb298ca7d2b37b060b4dbf9b166d39ba77a108b08edeeb4d67196e72cb3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf0aff8a6ab816b6617b1feeca9146d41bb5e8808bf1db96f8d493500c135d85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf0aff8a6ab816b6617b1feeca9146d41bb5e8808bf1db96f8d493500c135d85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:33Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:33 crc kubenswrapper[4860]: I1211 08:11:33.879867 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:33Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:33 crc kubenswrapper[4860]: I1211 08:11:33.892986 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a41f35db087418d1be16953f10377042abd3f4488e5ac1bc29ca5dcfc00a37e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:33Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:33 crc kubenswrapper[4860]: I1211 08:11:33.907830 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9aa0aed-62b3-4ea2-82ae-9a653c1101a5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ad4ba1b769c6cb2b34073798f8e3c25eb0177e6d3059484bf365e6981cab710\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://077aa1dd2fe5189ff60a0f2365e47a6a545cca0de7aa6379370908db4ea5838a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d7c919899b74a2b0fcd9221f3f769a9d013267354b200a282669d342d6413d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5483384f1d2ebc1fefa9706c9e67a3dc5a00ff643f803e3f2707c252e27ad94\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:33Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:33 crc kubenswrapper[4860]: I1211 08:11:33.931743 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31c30642-6e60-41b4-a477-0d802424e0aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79969402f5c33852a26fc839f37a45febf4f2377d77823284fbb01769e5585fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2d4d5de7106175087d91acb1f0c7552014f0c08bfcd65cbbf184d3f8d7826ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-99qgp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:33Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:33 crc kubenswrapper[4860]: I1211 08:11:33.942045 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:33 crc kubenswrapper[4860]: I1211 08:11:33.942067 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:33 crc kubenswrapper[4860]: I1211 08:11:33.942075 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:33 crc kubenswrapper[4860]: I1211 08:11:33.942088 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:33 crc kubenswrapper[4860]: I1211 08:11:33.942097 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:33Z","lastTransitionTime":"2025-12-11T08:11:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:33 crc kubenswrapper[4860]: I1211 08:11:33.974168 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-26nc6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1faa733-f408-41fd-b7aa-0cd86283976a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ac45db4f858a1c18a6b87472fc1cea2300f5845f1ad62c22260374995548a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ac45db4f858a1c18a6b87472fc1cea2300f5845f1ad62c22260374995548a0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b35296a5eb17dd2bb5e72c86c1cafdb30b7712cb330264f935d3b4fff05cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c6b35296a5eb17dd2bb5e72c86c1cafdb30b7712cb330264f935d3b4fff05cd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b9e82d6d3a247309ef5bdbf5b3a86ec083add2a138b0b237801e77b09ee96a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b9e82d6d3a247309ef5bdbf5b3a86ec083add2a138b0b237801e77b09ee96a1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-26nc6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:33Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:34 crc kubenswrapper[4860]: I1211 08:11:34.019535 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0589a204-c98c-417d-8256-bfe3bf747660\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17cbcb49c06f3a9059cd5a0f05eedb8c785da30a3e502e31bae2f57b3486cb14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17cbcb49c06f3a9059cd5a0f05eedb8c785da30a3e502e31bae2f57b3486cb14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xvnqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:34Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:34 crc kubenswrapper[4860]: I1211 08:11:34.043936 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:34 crc kubenswrapper[4860]: I1211 08:11:34.043969 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:34 crc kubenswrapper[4860]: I1211 08:11:34.043982 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:34 crc kubenswrapper[4860]: I1211 08:11:34.043999 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:34 crc kubenswrapper[4860]: I1211 08:11:34.044011 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:34Z","lastTransitionTime":"2025-12-11T08:11:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:34 crc kubenswrapper[4860]: I1211 08:11:34.052435 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gsx9l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aba5b6b9-f88d-4763-9562-6b4dec606df9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://735ef012c523ddc952d46a431f27b2582a2c301d9e9da0c697a20ff6e2fb5d86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6zw9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gsx9l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:34Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:34 crc kubenswrapper[4860]: I1211 08:11:34.104098 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9872be1e-b995-4a00-904b-875dff96ad56\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f74a32a8fefea704cc73ffd1114083dab68aa44dc33f8d9d432d2d24216546c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbe00b46063d3968d59c393b87178e135f982abb7d0f32fa798b212d4b83df6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eedc9988e1a4a3c8cae730dc4a352103a6d665535b825e157d0f87c331b137c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2137fee4b7abc9765f9517d7b3b6ea796601a31d03151bf167d17136ced8b25a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f683e48f3577a68862eb6302383538376e41f49b20b1e0ea3f1b3b143f24872a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5de794187456b3576b969069a3b839d75c7918418a97fa35dff8bb40bb76e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5de794187456b3576b969069a3b839d75c7918418a97fa35dff8bb40bb76e39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08ce42b974fba00c409c0af971f0de351e50f18b2f06f8e722150da3239bb908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08ce42b974fba00c409c0af971f0de351e50f18b2f06f8e722150da3239bb908\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cfac6be03e3cfa0b0f3450b48b67957f2d5172b9c65e27f14c2a260bda5d423f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cfac6be03e3cfa0b0f3450b48b67957f2d5172b9c65e27f14c2a260bda5d423f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:34Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:34 crc kubenswrapper[4860]: I1211 08:11:34.133672 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-pznwl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5310c1fc-66c6-40aa-b1bf-5a59a2410f9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2d6ff00ce322a28607bc3f6db05c5e4fb14acb6777ea315786ef53963ca504a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hkt48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-pznwl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:34Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:34 crc kubenswrapper[4860]: I1211 08:11:34.145989 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:34 crc kubenswrapper[4860]: I1211 08:11:34.146019 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:34 crc kubenswrapper[4860]: I1211 08:11:34.146029 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:34 crc kubenswrapper[4860]: I1211 08:11:34.146043 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:34 crc kubenswrapper[4860]: I1211 08:11:34.146078 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:34Z","lastTransitionTime":"2025-12-11T08:11:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:34 crc kubenswrapper[4860]: I1211 08:11:34.248409 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:34 crc kubenswrapper[4860]: I1211 08:11:34.248439 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:34 crc kubenswrapper[4860]: I1211 08:11:34.248450 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:34 crc kubenswrapper[4860]: I1211 08:11:34.248464 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:34 crc kubenswrapper[4860]: I1211 08:11:34.248472 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:34Z","lastTransitionTime":"2025-12-11T08:11:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:34 crc kubenswrapper[4860]: I1211 08:11:34.263914 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:11:34 crc kubenswrapper[4860]: I1211 08:11:34.264048 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:11:34 crc kubenswrapper[4860]: E1211 08:11:34.264114 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:11:42.264083749 +0000 UTC m=+34.992602834 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:11:34 crc kubenswrapper[4860]: E1211 08:11:34.264199 4860 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 11 08:11:34 crc kubenswrapper[4860]: E1211 08:11:34.264361 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-11 08:11:42.264332347 +0000 UTC m=+34.992851452 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 11 08:11:34 crc kubenswrapper[4860]: I1211 08:11:34.350831 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:34 crc kubenswrapper[4860]: I1211 08:11:34.350870 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:34 crc kubenswrapper[4860]: I1211 08:11:34.350879 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:34 crc kubenswrapper[4860]: I1211 08:11:34.350891 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:34 crc kubenswrapper[4860]: I1211 08:11:34.350901 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:34Z","lastTransitionTime":"2025-12-11T08:11:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:34 crc kubenswrapper[4860]: I1211 08:11:34.365588 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:11:34 crc kubenswrapper[4860]: I1211 08:11:34.365624 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:11:34 crc kubenswrapper[4860]: I1211 08:11:34.365674 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:11:34 crc kubenswrapper[4860]: E1211 08:11:34.365798 4860 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 11 08:11:34 crc kubenswrapper[4860]: E1211 08:11:34.365812 4860 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 11 08:11:34 crc kubenswrapper[4860]: E1211 08:11:34.365823 4860 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 08:11:34 crc kubenswrapper[4860]: E1211 08:11:34.365863 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-11 08:11:42.36584961 +0000 UTC m=+35.094368665 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 08:11:34 crc kubenswrapper[4860]: E1211 08:11:34.366050 4860 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 11 08:11:34 crc kubenswrapper[4860]: E1211 08:11:34.366097 4860 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 11 08:11:34 crc kubenswrapper[4860]: E1211 08:11:34.366118 4860 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 11 08:11:34 crc kubenswrapper[4860]: E1211 08:11:34.366132 4860 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 08:11:34 crc kubenswrapper[4860]: E1211 08:11:34.366149 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-11 08:11:42.366126479 +0000 UTC m=+35.094645594 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 11 08:11:34 crc kubenswrapper[4860]: E1211 08:11:34.366169 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-11 08:11:42.36615794 +0000 UTC m=+35.094677065 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 08:11:34 crc kubenswrapper[4860]: I1211 08:11:34.452861 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:34 crc kubenswrapper[4860]: I1211 08:11:34.452903 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:34 crc kubenswrapper[4860]: I1211 08:11:34.452919 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:34 crc kubenswrapper[4860]: I1211 08:11:34.452936 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:34 crc kubenswrapper[4860]: I1211 08:11:34.452947 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:34Z","lastTransitionTime":"2025-12-11T08:11:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:34 crc kubenswrapper[4860]: I1211 08:11:34.555839 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:34 crc kubenswrapper[4860]: I1211 08:11:34.555894 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:34 crc kubenswrapper[4860]: I1211 08:11:34.555911 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:34 crc kubenswrapper[4860]: I1211 08:11:34.555933 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:34 crc kubenswrapper[4860]: I1211 08:11:34.555953 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:34Z","lastTransitionTime":"2025-12-11T08:11:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:34 crc kubenswrapper[4860]: I1211 08:11:34.578360 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:11:34 crc kubenswrapper[4860]: I1211 08:11:34.578441 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:11:34 crc kubenswrapper[4860]: E1211 08:11:34.578520 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:11:34 crc kubenswrapper[4860]: I1211 08:11:34.578579 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:11:34 crc kubenswrapper[4860]: E1211 08:11:34.578785 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:11:34 crc kubenswrapper[4860]: E1211 08:11:34.578880 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:11:34 crc kubenswrapper[4860]: I1211 08:11:34.659556 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:34 crc kubenswrapper[4860]: I1211 08:11:34.659616 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:34 crc kubenswrapper[4860]: I1211 08:11:34.659631 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:34 crc kubenswrapper[4860]: I1211 08:11:34.659675 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:34 crc kubenswrapper[4860]: I1211 08:11:34.659699 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:34Z","lastTransitionTime":"2025-12-11T08:11:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:34 crc kubenswrapper[4860]: I1211 08:11:34.763615 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:34 crc kubenswrapper[4860]: I1211 08:11:34.763704 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:34 crc kubenswrapper[4860]: I1211 08:11:34.763720 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:34 crc kubenswrapper[4860]: I1211 08:11:34.763742 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:34 crc kubenswrapper[4860]: I1211 08:11:34.763780 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:34Z","lastTransitionTime":"2025-12-11T08:11:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:34 crc kubenswrapper[4860]: I1211 08:11:34.797619 4860 generic.go:334] "Generic (PLEG): container finished" podID="a1faa733-f408-41fd-b7aa-0cd86283976a" containerID="908ee695e78145ef4375c4097d9f1e772502922d02a2fcb4566172a6ad9caaf9" exitCode=0 Dec 11 08:11:34 crc kubenswrapper[4860]: I1211 08:11:34.797685 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-26nc6" event={"ID":"a1faa733-f408-41fd-b7aa-0cd86283976a","Type":"ContainerDied","Data":"908ee695e78145ef4375c4097d9f1e772502922d02a2fcb4566172a6ad9caaf9"} Dec 11 08:11:34 crc kubenswrapper[4860]: I1211 08:11:34.803452 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" event={"ID":"0589a204-c98c-417d-8256-bfe3bf747660","Type":"ContainerStarted","Data":"bb80d9928e8349fc5963227d135355ad7ba820d4fd89877af601d005ca03334b"} Dec 11 08:11:34 crc kubenswrapper[4860]: I1211 08:11:34.812110 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gsx9l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aba5b6b9-f88d-4763-9562-6b4dec606df9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://735ef012c523ddc952d46a431f27b2582a2c301d9e9da0c697a20ff6e2fb5d86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6zw9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gsx9l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:34Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:34 crc kubenswrapper[4860]: I1211 08:11:34.830737 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9aa0aed-62b3-4ea2-82ae-9a653c1101a5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ad4ba1b769c6cb2b34073798f8e3c25eb0177e6d3059484bf365e6981cab710\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://077aa1dd2fe5189ff60a0f2365e47a6a545cca0de7aa6379370908db4ea5838a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d7c919899b74a2b0fcd9221f3f769a9d013267354b200a282669d342d6413d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5483384f1d2ebc1fefa9706c9e67a3dc5a00ff643f803e3f2707c252e27ad94\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:34Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:34 crc kubenswrapper[4860]: I1211 08:11:34.843908 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31c30642-6e60-41b4-a477-0d802424e0aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79969402f5c33852a26fc839f37a45febf4f2377d77823284fbb01769e5585fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2d4d5de7106175087d91acb1f0c7552014f0c08bfcd65cbbf184d3f8d7826ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-99qgp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:34Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:34 crc kubenswrapper[4860]: I1211 08:11:34.863171 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-26nc6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1faa733-f408-41fd-b7aa-0cd86283976a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ac45db4f858a1c18a6b87472fc1cea2300f5845f1ad62c22260374995548a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ac45db4f858a1c18a6b87472fc1cea2300f5845f1ad62c22260374995548a0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b35296a5eb17dd2bb5e72c86c1cafdb30b7712cb330264f935d3b4fff05cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c6b35296a5eb17dd2bb5e72c86c1cafdb30b7712cb330264f935d3b4fff05cd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b9e82d6d3a247309ef5bdbf5b3a86ec083add2a138b0b237801e77b09ee96a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b9e82d6d3a247309ef5bdbf5b3a86ec083add2a138b0b237801e77b09ee96a1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://908ee695e78145ef4375c4097d9f1e772502922d02a2fcb4566172a6ad9caaf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://908ee695e78145ef4375c4097d9f1e772502922d02a2fcb4566172a6ad9caaf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-26nc6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:34Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:34 crc kubenswrapper[4860]: I1211 08:11:34.866793 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:34 crc kubenswrapper[4860]: I1211 08:11:34.866841 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:34 crc kubenswrapper[4860]: I1211 08:11:34.866853 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:34 crc kubenswrapper[4860]: I1211 08:11:34.866875 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:34 crc kubenswrapper[4860]: I1211 08:11:34.866888 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:34Z","lastTransitionTime":"2025-12-11T08:11:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:34 crc kubenswrapper[4860]: I1211 08:11:34.887453 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0589a204-c98c-417d-8256-bfe3bf747660\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17cbcb49c06f3a9059cd5a0f05eedb8c785da30a3e502e31bae2f57b3486cb14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17cbcb49c06f3a9059cd5a0f05eedb8c785da30a3e502e31bae2f57b3486cb14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xvnqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:34Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:34 crc kubenswrapper[4860]: I1211 08:11:34.906352 4860 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 11 08:11:34 crc kubenswrapper[4860]: I1211 08:11:34.910835 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9872be1e-b995-4a00-904b-875dff96ad56\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f74a32a8fefea704cc73ffd1114083dab68aa44dc33f8d9d432d2d24216546c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbe00b46063d3968d59c393b87178e135f982abb7d0f32fa798b212d4b83df6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eedc9988e1a4a3c8cae730dc4a352103a6d665535b825e157d0f87c331b137c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2137fee4b7abc9765f9517d7b3b6ea796601a31d03151bf167d17136ced8b25a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f683e48f3577a68862eb6302383538376e41f49b20b1e0ea3f1b3b143f24872a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5de794187456b3576b969069a3b839d75c7918418a97fa35dff8bb40bb76e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5de794187456b3576b969069a3b839d75c7918418a97fa35dff8bb40bb76e39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08ce42b974fba00c409c0af971f0de351e50f18b2f06f8e722150da3239bb908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08ce42b974fba00c409c0af971f0de351e50f18b2f06f8e722150da3239bb908\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cfac6be03e3cfa0b0f3450b48b67957f2d5172b9c65e27f14c2a260bda5d423f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cfac6be03e3cfa0b0f3450b48b67957f2d5172b9c65e27f14c2a260bda5d423f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:34Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:34 crc kubenswrapper[4860]: I1211 08:11:34.930839 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-pznwl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5310c1fc-66c6-40aa-b1bf-5a59a2410f9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2d6ff00ce322a28607bc3f6db05c5e4fb14acb6777ea315786ef53963ca504a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hkt48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-pznwl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:34Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:34 crc kubenswrapper[4860]: I1211 08:11:34.941579 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6xq7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a75ce3c-5e1e-4b9f-bbde-7a3b0cce5eae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe8d4d675567b41728e476351811a93c47b19868ccbadab8c613a4eb8eca8562\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5c6nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6xq7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:34Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:34 crc kubenswrapper[4860]: I1211 08:11:34.955123 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:34Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:34 crc kubenswrapper[4860]: I1211 08:11:34.966952 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d8c2f28628dc2f1ffbdd98a55670e22cf6cdc8cf2c6442e9d1f8e5466b2f12e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://074ea63beac52bd196507ee7cfb1c3d7fc2d37d5ec6faa4d47dde62b922e1c43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:34Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:34 crc kubenswrapper[4860]: I1211 08:11:34.968949 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:34 crc kubenswrapper[4860]: I1211 08:11:34.968985 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:34 crc kubenswrapper[4860]: I1211 08:11:34.968994 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:34 crc kubenswrapper[4860]: I1211 08:11:34.969008 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:34 crc kubenswrapper[4860]: I1211 08:11:34.969017 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:34Z","lastTransitionTime":"2025-12-11T08:11:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:34 crc kubenswrapper[4860]: I1211 08:11:34.977292 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:34Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:34 crc kubenswrapper[4860]: I1211 08:11:34.986945 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25a6a154da2b74952951ae94efc914ba078ea4bbe5893360b3c84d46be6bf973\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:34Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:34 crc kubenswrapper[4860]: I1211 08:11:34.996660 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ec3758-85f1-4a22-a4ec-c805472607a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e831db1a15976d3967d56479689572c26f2599a766baa0402897a3008428c56f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a498aa90f914e1f0370df01dc5ef174df113a2a16d86b2e3f24b19a1b99b4c0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8caebd2f4dfcbb89c597029a07c335171889761d2d200206987537fb0a228db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34e91f615b76445ff3241a304bb2964fde59de7e8a8dfff6841b2e628a31e86d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2090cab02ad74ddfd6bb57ff50eba35d18f56b0cde1c230fe011c01994b693bc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 08:11:21.171011 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 08:11:21.177674 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-644352688/tls.crt::/tmp/serving-cert-644352688/tls.key\\\\\\\"\\\\nI1211 08:11:26.508606 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 08:11:26.512741 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 08:11:26.512778 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 08:11:26.512839 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 08:11:26.512858 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 08:11:26.521684 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1211 08:11:26.521700 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1211 08:11:26.521722 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:11:26.521732 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:11:26.521738 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 08:11:26.521746 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 08:11:26.521749 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 08:11:26.521753 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1211 08:11:26.524999 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2274fb298ca7d2b37b060b4dbf9b166d39ba77a108b08edeeb4d67196e72cb3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf0aff8a6ab816b6617b1feeca9146d41bb5e8808bf1db96f8d493500c135d85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf0aff8a6ab816b6617b1feeca9146d41bb5e8808bf1db96f8d493500c135d85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:34Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:35 crc kubenswrapper[4860]: I1211 08:11:35.005636 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:35Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:35 crc kubenswrapper[4860]: I1211 08:11:35.014947 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a41f35db087418d1be16953f10377042abd3f4488e5ac1bc29ca5dcfc00a37e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:35Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:35 crc kubenswrapper[4860]: I1211 08:11:35.071352 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:35 crc kubenswrapper[4860]: I1211 08:11:35.071388 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:35 crc kubenswrapper[4860]: I1211 08:11:35.071399 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:35 crc kubenswrapper[4860]: I1211 08:11:35.071414 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:35 crc kubenswrapper[4860]: I1211 08:11:35.071426 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:35Z","lastTransitionTime":"2025-12-11T08:11:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:35 crc kubenswrapper[4860]: I1211 08:11:35.180632 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:35 crc kubenswrapper[4860]: I1211 08:11:35.180964 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:35 crc kubenswrapper[4860]: I1211 08:11:35.181079 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:35 crc kubenswrapper[4860]: I1211 08:11:35.181108 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:35 crc kubenswrapper[4860]: I1211 08:11:35.181526 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:35Z","lastTransitionTime":"2025-12-11T08:11:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:35 crc kubenswrapper[4860]: I1211 08:11:35.284047 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:35 crc kubenswrapper[4860]: I1211 08:11:35.284131 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:35 crc kubenswrapper[4860]: I1211 08:11:35.284154 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:35 crc kubenswrapper[4860]: I1211 08:11:35.284201 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:35 crc kubenswrapper[4860]: I1211 08:11:35.284235 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:35Z","lastTransitionTime":"2025-12-11T08:11:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:35 crc kubenswrapper[4860]: I1211 08:11:35.386950 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:35 crc kubenswrapper[4860]: I1211 08:11:35.387016 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:35 crc kubenswrapper[4860]: I1211 08:11:35.387042 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:35 crc kubenswrapper[4860]: I1211 08:11:35.387074 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:35 crc kubenswrapper[4860]: I1211 08:11:35.387096 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:35Z","lastTransitionTime":"2025-12-11T08:11:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:35 crc kubenswrapper[4860]: I1211 08:11:35.490223 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:35 crc kubenswrapper[4860]: I1211 08:11:35.490287 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:35 crc kubenswrapper[4860]: I1211 08:11:35.490305 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:35 crc kubenswrapper[4860]: I1211 08:11:35.490329 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:35 crc kubenswrapper[4860]: I1211 08:11:35.490344 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:35Z","lastTransitionTime":"2025-12-11T08:11:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:35 crc kubenswrapper[4860]: I1211 08:11:35.592448 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:35 crc kubenswrapper[4860]: I1211 08:11:35.592474 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:35 crc kubenswrapper[4860]: I1211 08:11:35.592482 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:35 crc kubenswrapper[4860]: I1211 08:11:35.592493 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:35 crc kubenswrapper[4860]: I1211 08:11:35.592501 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:35Z","lastTransitionTime":"2025-12-11T08:11:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:35 crc kubenswrapper[4860]: I1211 08:11:35.694599 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:35 crc kubenswrapper[4860]: I1211 08:11:35.694708 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:35 crc kubenswrapper[4860]: I1211 08:11:35.694730 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:35 crc kubenswrapper[4860]: I1211 08:11:35.694755 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:35 crc kubenswrapper[4860]: I1211 08:11:35.694772 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:35Z","lastTransitionTime":"2025-12-11T08:11:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:35 crc kubenswrapper[4860]: I1211 08:11:35.797578 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:35 crc kubenswrapper[4860]: I1211 08:11:35.797624 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:35 crc kubenswrapper[4860]: I1211 08:11:35.797678 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:35 crc kubenswrapper[4860]: I1211 08:11:35.797703 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:35 crc kubenswrapper[4860]: I1211 08:11:35.797722 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:35Z","lastTransitionTime":"2025-12-11T08:11:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:35 crc kubenswrapper[4860]: I1211 08:11:35.823523 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-26nc6" event={"ID":"a1faa733-f408-41fd-b7aa-0cd86283976a","Type":"ContainerStarted","Data":"afb0a097f3556a8f9ed7cf8f762cb99c082fe0a36b8d8d3038bf188b08fe9a84"} Dec 11 08:11:35 crc kubenswrapper[4860]: I1211 08:11:35.843099 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0589a204-c98c-417d-8256-bfe3bf747660\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17cbcb49c06f3a9059cd5a0f05eedb8c785da30a3e502e31bae2f57b3486cb14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17cbcb49c06f3a9059cd5a0f05eedb8c785da30a3e502e31bae2f57b3486cb14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xvnqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:35Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:35 crc kubenswrapper[4860]: I1211 08:11:35.857583 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gsx9l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aba5b6b9-f88d-4763-9562-6b4dec606df9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://735ef012c523ddc952d46a431f27b2582a2c301d9e9da0c697a20ff6e2fb5d86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6zw9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gsx9l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:35Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:35 crc kubenswrapper[4860]: I1211 08:11:35.876975 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9aa0aed-62b3-4ea2-82ae-9a653c1101a5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ad4ba1b769c6cb2b34073798f8e3c25eb0177e6d3059484bf365e6981cab710\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://077aa1dd2fe5189ff60a0f2365e47a6a545cca0de7aa6379370908db4ea5838a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d7c919899b74a2b0fcd9221f3f769a9d013267354b200a282669d342d6413d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5483384f1d2ebc1fefa9706c9e67a3dc5a00ff643f803e3f2707c252e27ad94\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:35Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:35 crc kubenswrapper[4860]: I1211 08:11:35.892748 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31c30642-6e60-41b4-a477-0d802424e0aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79969402f5c33852a26fc839f37a45febf4f2377d77823284fbb01769e5585fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2d4d5de7106175087d91acb1f0c7552014f0c08bfcd65cbbf184d3f8d7826ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-99qgp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:35Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:35 crc kubenswrapper[4860]: I1211 08:11:35.900110 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:35 crc kubenswrapper[4860]: I1211 08:11:35.900139 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:35 crc kubenswrapper[4860]: I1211 08:11:35.900148 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:35 crc kubenswrapper[4860]: I1211 08:11:35.900159 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:35 crc kubenswrapper[4860]: I1211 08:11:35.900167 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:35Z","lastTransitionTime":"2025-12-11T08:11:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:35 crc kubenswrapper[4860]: I1211 08:11:35.911189 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-26nc6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1faa733-f408-41fd-b7aa-0cd86283976a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ac45db4f858a1c18a6b87472fc1cea2300f5845f1ad62c22260374995548a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ac45db4f858a1c18a6b87472fc1cea2300f5845f1ad62c22260374995548a0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b35296a5eb17dd2bb5e72c86c1cafdb30b7712cb330264f935d3b4fff05cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c6b35296a5eb17dd2bb5e72c86c1cafdb30b7712cb330264f935d3b4fff05cd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b9e82d6d3a247309ef5bdbf5b3a86ec083add2a138b0b237801e77b09ee96a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b9e82d6d3a247309ef5bdbf5b3a86ec083add2a138b0b237801e77b09ee96a1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://908ee695e78145ef4375c4097d9f1e772502922d02a2fcb4566172a6ad9caaf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://908ee695e78145ef4375c4097d9f1e772502922d02a2fcb4566172a6ad9caaf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afb0a097f3556a8f9ed7cf8f762cb99c082fe0a36b8d8d3038bf188b08fe9a84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-26nc6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:35Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:35 crc kubenswrapper[4860]: I1211 08:11:35.939247 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9872be1e-b995-4a00-904b-875dff96ad56\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f74a32a8fefea704cc73ffd1114083dab68aa44dc33f8d9d432d2d24216546c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbe00b46063d3968d59c393b87178e135f982abb7d0f32fa798b212d4b83df6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eedc9988e1a4a3c8cae730dc4a352103a6d665535b825e157d0f87c331b137c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2137fee4b7abc9765f9517d7b3b6ea796601a31d03151bf167d17136ced8b25a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f683e48f3577a68862eb6302383538376e41f49b20b1e0ea3f1b3b143f24872a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5de794187456b3576b969069a3b839d75c7918418a97fa35dff8bb40bb76e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5de794187456b3576b969069a3b839d75c7918418a97fa35dff8bb40bb76e39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08ce42b974fba00c409c0af971f0de351e50f18b2f06f8e722150da3239bb908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08ce42b974fba00c409c0af971f0de351e50f18b2f06f8e722150da3239bb908\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cfac6be03e3cfa0b0f3450b48b67957f2d5172b9c65e27f14c2a260bda5d423f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cfac6be03e3cfa0b0f3450b48b67957f2d5172b9c65e27f14c2a260bda5d423f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:35Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:35 crc kubenswrapper[4860]: I1211 08:11:35.957356 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-pznwl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5310c1fc-66c6-40aa-b1bf-5a59a2410f9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2d6ff00ce322a28607bc3f6db05c5e4fb14acb6777ea315786ef53963ca504a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hkt48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-pznwl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:35Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:35 crc kubenswrapper[4860]: I1211 08:11:35.971384 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25a6a154da2b74952951ae94efc914ba078ea4bbe5893360b3c84d46be6bf973\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:35Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:35 crc kubenswrapper[4860]: I1211 08:11:35.983975 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6xq7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a75ce3c-5e1e-4b9f-bbde-7a3b0cce5eae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe8d4d675567b41728e476351811a93c47b19868ccbadab8c613a4eb8eca8562\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5c6nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6xq7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:35Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:35 crc kubenswrapper[4860]: I1211 08:11:35.996899 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:35Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:36 crc kubenswrapper[4860]: I1211 08:11:36.002205 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:36 crc kubenswrapper[4860]: I1211 08:11:36.002242 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:36 crc kubenswrapper[4860]: I1211 08:11:36.002251 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:36 crc kubenswrapper[4860]: I1211 08:11:36.002264 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:36 crc kubenswrapper[4860]: I1211 08:11:36.002274 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:36Z","lastTransitionTime":"2025-12-11T08:11:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:36 crc kubenswrapper[4860]: I1211 08:11:36.011337 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d8c2f28628dc2f1ffbdd98a55670e22cf6cdc8cf2c6442e9d1f8e5466b2f12e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://074ea63beac52bd196507ee7cfb1c3d7fc2d37d5ec6faa4d47dde62b922e1c43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:36Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:36 crc kubenswrapper[4860]: I1211 08:11:36.023766 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:36Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:36 crc kubenswrapper[4860]: I1211 08:11:36.038748 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ec3758-85f1-4a22-a4ec-c805472607a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e831db1a15976d3967d56479689572c26f2599a766baa0402897a3008428c56f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a498aa90f914e1f0370df01dc5ef174df113a2a16d86b2e3f24b19a1b99b4c0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8caebd2f4dfcbb89c597029a07c335171889761d2d200206987537fb0a228db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34e91f615b76445ff3241a304bb2964fde59de7e8a8dfff6841b2e628a31e86d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2090cab02ad74ddfd6bb57ff50eba35d18f56b0cde1c230fe011c01994b693bc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 08:11:21.171011 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 08:11:21.177674 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-644352688/tls.crt::/tmp/serving-cert-644352688/tls.key\\\\\\\"\\\\nI1211 08:11:26.508606 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 08:11:26.512741 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 08:11:26.512778 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 08:11:26.512839 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 08:11:26.512858 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 08:11:26.521684 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1211 08:11:26.521700 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1211 08:11:26.521722 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:11:26.521732 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:11:26.521738 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 08:11:26.521746 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 08:11:26.521749 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 08:11:26.521753 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1211 08:11:26.524999 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2274fb298ca7d2b37b060b4dbf9b166d39ba77a108b08edeeb4d67196e72cb3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf0aff8a6ab816b6617b1feeca9146d41bb5e8808bf1db96f8d493500c135d85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf0aff8a6ab816b6617b1feeca9146d41bb5e8808bf1db96f8d493500c135d85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:36Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:36 crc kubenswrapper[4860]: I1211 08:11:36.055888 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:36Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:36 crc kubenswrapper[4860]: I1211 08:11:36.070008 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a41f35db087418d1be16953f10377042abd3f4488e5ac1bc29ca5dcfc00a37e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:36Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:36 crc kubenswrapper[4860]: I1211 08:11:36.104239 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:36 crc kubenswrapper[4860]: I1211 08:11:36.104279 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:36 crc kubenswrapper[4860]: I1211 08:11:36.104289 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:36 crc kubenswrapper[4860]: I1211 08:11:36.104302 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:36 crc kubenswrapper[4860]: I1211 08:11:36.104321 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:36Z","lastTransitionTime":"2025-12-11T08:11:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:36 crc kubenswrapper[4860]: I1211 08:11:36.206767 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:36 crc kubenswrapper[4860]: I1211 08:11:36.206811 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:36 crc kubenswrapper[4860]: I1211 08:11:36.206825 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:36 crc kubenswrapper[4860]: I1211 08:11:36.206841 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:36 crc kubenswrapper[4860]: I1211 08:11:36.206853 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:36Z","lastTransitionTime":"2025-12-11T08:11:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:36 crc kubenswrapper[4860]: I1211 08:11:36.309406 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:36 crc kubenswrapper[4860]: I1211 08:11:36.309466 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:36 crc kubenswrapper[4860]: I1211 08:11:36.309486 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:36 crc kubenswrapper[4860]: I1211 08:11:36.309510 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:36 crc kubenswrapper[4860]: I1211 08:11:36.309529 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:36Z","lastTransitionTime":"2025-12-11T08:11:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:36 crc kubenswrapper[4860]: I1211 08:11:36.411556 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:36 crc kubenswrapper[4860]: I1211 08:11:36.411587 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:36 crc kubenswrapper[4860]: I1211 08:11:36.411597 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:36 crc kubenswrapper[4860]: I1211 08:11:36.411609 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:36 crc kubenswrapper[4860]: I1211 08:11:36.411617 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:36Z","lastTransitionTime":"2025-12-11T08:11:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:36 crc kubenswrapper[4860]: I1211 08:11:36.514908 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:36 crc kubenswrapper[4860]: I1211 08:11:36.514945 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:36 crc kubenswrapper[4860]: I1211 08:11:36.514956 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:36 crc kubenswrapper[4860]: I1211 08:11:36.514974 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:36 crc kubenswrapper[4860]: I1211 08:11:36.514985 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:36Z","lastTransitionTime":"2025-12-11T08:11:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:36 crc kubenswrapper[4860]: I1211 08:11:36.578105 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:11:36 crc kubenswrapper[4860]: I1211 08:11:36.578105 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:11:36 crc kubenswrapper[4860]: E1211 08:11:36.578237 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:11:36 crc kubenswrapper[4860]: E1211 08:11:36.578315 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:11:36 crc kubenswrapper[4860]: I1211 08:11:36.578105 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:11:36 crc kubenswrapper[4860]: E1211 08:11:36.578420 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:11:36 crc kubenswrapper[4860]: I1211 08:11:36.616874 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:36 crc kubenswrapper[4860]: I1211 08:11:36.616949 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:36 crc kubenswrapper[4860]: I1211 08:11:36.616965 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:36 crc kubenswrapper[4860]: I1211 08:11:36.616995 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:36 crc kubenswrapper[4860]: I1211 08:11:36.617012 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:36Z","lastTransitionTime":"2025-12-11T08:11:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:36 crc kubenswrapper[4860]: I1211 08:11:36.720840 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:36 crc kubenswrapper[4860]: I1211 08:11:36.721218 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:36 crc kubenswrapper[4860]: I1211 08:11:36.721235 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:36 crc kubenswrapper[4860]: I1211 08:11:36.721256 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:36 crc kubenswrapper[4860]: I1211 08:11:36.721273 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:36Z","lastTransitionTime":"2025-12-11T08:11:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:36 crc kubenswrapper[4860]: I1211 08:11:36.824466 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:36 crc kubenswrapper[4860]: I1211 08:11:36.824520 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:36 crc kubenswrapper[4860]: I1211 08:11:36.824535 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:36 crc kubenswrapper[4860]: I1211 08:11:36.824556 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:36 crc kubenswrapper[4860]: I1211 08:11:36.824570 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:36Z","lastTransitionTime":"2025-12-11T08:11:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:36 crc kubenswrapper[4860]: I1211 08:11:36.846350 4860 generic.go:334] "Generic (PLEG): container finished" podID="a1faa733-f408-41fd-b7aa-0cd86283976a" containerID="afb0a097f3556a8f9ed7cf8f762cb99c082fe0a36b8d8d3038bf188b08fe9a84" exitCode=0 Dec 11 08:11:36 crc kubenswrapper[4860]: I1211 08:11:36.846414 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-26nc6" event={"ID":"a1faa733-f408-41fd-b7aa-0cd86283976a","Type":"ContainerDied","Data":"afb0a097f3556a8f9ed7cf8f762cb99c082fe0a36b8d8d3038bf188b08fe9a84"} Dec 11 08:11:36 crc kubenswrapper[4860]: I1211 08:11:36.882037 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9872be1e-b995-4a00-904b-875dff96ad56\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f74a32a8fefea704cc73ffd1114083dab68aa44dc33f8d9d432d2d24216546c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbe00b46063d3968d59c393b87178e135f982abb7d0f32fa798b212d4b83df6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eedc9988e1a4a3c8cae730dc4a352103a6d665535b825e157d0f87c331b137c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2137fee4b7abc9765f9517d7b3b6ea796601a31d03151bf167d17136ced8b25a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f683e48f3577a68862eb6302383538376e41f49b20b1e0ea3f1b3b143f24872a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5de794187456b3576b969069a3b839d75c7918418a97fa35dff8bb40bb76e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5de794187456b3576b969069a3b839d75c7918418a97fa35dff8bb40bb76e39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08ce42b974fba00c409c0af971f0de351e50f18b2f06f8e722150da3239bb908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08ce42b974fba00c409c0af971f0de351e50f18b2f06f8e722150da3239bb908\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cfac6be03e3cfa0b0f3450b48b67957f2d5172b9c65e27f14c2a260bda5d423f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cfac6be03e3cfa0b0f3450b48b67957f2d5172b9c65e27f14c2a260bda5d423f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:36Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:36 crc kubenswrapper[4860]: I1211 08:11:36.901966 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-pznwl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5310c1fc-66c6-40aa-b1bf-5a59a2410f9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2d6ff00ce322a28607bc3f6db05c5e4fb14acb6777ea315786ef53963ca504a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hkt48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-pznwl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:36Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:36 crc kubenswrapper[4860]: I1211 08:11:36.915216 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6xq7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a75ce3c-5e1e-4b9f-bbde-7a3b0cce5eae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe8d4d675567b41728e476351811a93c47b19868ccbadab8c613a4eb8eca8562\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5c6nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6xq7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:36Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:36 crc kubenswrapper[4860]: I1211 08:11:36.927274 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:36 crc kubenswrapper[4860]: I1211 08:11:36.927304 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:36 crc kubenswrapper[4860]: I1211 08:11:36.927315 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:36 crc kubenswrapper[4860]: I1211 08:11:36.927330 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:36 crc kubenswrapper[4860]: I1211 08:11:36.927342 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:36Z","lastTransitionTime":"2025-12-11T08:11:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:36 crc kubenswrapper[4860]: I1211 08:11:36.933703 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:36Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:36 crc kubenswrapper[4860]: I1211 08:11:36.954470 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d8c2f28628dc2f1ffbdd98a55670e22cf6cdc8cf2c6442e9d1f8e5466b2f12e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://074ea63beac52bd196507ee7cfb1c3d7fc2d37d5ec6faa4d47dde62b922e1c43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:36Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:36 crc kubenswrapper[4860]: I1211 08:11:36.984073 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:36Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:36 crc kubenswrapper[4860]: I1211 08:11:36.999361 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25a6a154da2b74952951ae94efc914ba078ea4bbe5893360b3c84d46be6bf973\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:36Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:37 crc kubenswrapper[4860]: I1211 08:11:37.018588 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ec3758-85f1-4a22-a4ec-c805472607a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e831db1a15976d3967d56479689572c26f2599a766baa0402897a3008428c56f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a498aa90f914e1f0370df01dc5ef174df113a2a16d86b2e3f24b19a1b99b4c0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8caebd2f4dfcbb89c597029a07c335171889761d2d200206987537fb0a228db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34e91f615b76445ff3241a304bb2964fde59de7e8a8dfff6841b2e628a31e86d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2090cab02ad74ddfd6bb57ff50eba35d18f56b0cde1c230fe011c01994b693bc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 08:11:21.171011 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 08:11:21.177674 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-644352688/tls.crt::/tmp/serving-cert-644352688/tls.key\\\\\\\"\\\\nI1211 08:11:26.508606 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 08:11:26.512741 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 08:11:26.512778 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 08:11:26.512839 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 08:11:26.512858 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 08:11:26.521684 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1211 08:11:26.521700 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1211 08:11:26.521722 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:11:26.521732 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:11:26.521738 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 08:11:26.521746 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 08:11:26.521749 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 08:11:26.521753 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1211 08:11:26.524999 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2274fb298ca7d2b37b060b4dbf9b166d39ba77a108b08edeeb4d67196e72cb3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf0aff8a6ab816b6617b1feeca9146d41bb5e8808bf1db96f8d493500c135d85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf0aff8a6ab816b6617b1feeca9146d41bb5e8808bf1db96f8d493500c135d85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:37Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:37 crc kubenswrapper[4860]: I1211 08:11:37.029852 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:37 crc kubenswrapper[4860]: I1211 08:11:37.029879 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:37 crc kubenswrapper[4860]: I1211 08:11:37.029889 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:37 crc kubenswrapper[4860]: I1211 08:11:37.029905 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:37 crc kubenswrapper[4860]: I1211 08:11:37.029916 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:37Z","lastTransitionTime":"2025-12-11T08:11:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:37 crc kubenswrapper[4860]: I1211 08:11:37.034366 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:37Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:37 crc kubenswrapper[4860]: I1211 08:11:37.051359 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a41f35db087418d1be16953f10377042abd3f4488e5ac1bc29ca5dcfc00a37e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:37Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:37 crc kubenswrapper[4860]: I1211 08:11:37.062715 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gsx9l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aba5b6b9-f88d-4763-9562-6b4dec606df9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://735ef012c523ddc952d46a431f27b2582a2c301d9e9da0c697a20ff6e2fb5d86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6zw9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gsx9l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:37Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:37 crc kubenswrapper[4860]: I1211 08:11:37.080696 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9aa0aed-62b3-4ea2-82ae-9a653c1101a5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ad4ba1b769c6cb2b34073798f8e3c25eb0177e6d3059484bf365e6981cab710\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://077aa1dd2fe5189ff60a0f2365e47a6a545cca0de7aa6379370908db4ea5838a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d7c919899b74a2b0fcd9221f3f769a9d013267354b200a282669d342d6413d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5483384f1d2ebc1fefa9706c9e67a3dc5a00ff643f803e3f2707c252e27ad94\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:37Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:37 crc kubenswrapper[4860]: I1211 08:11:37.091898 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31c30642-6e60-41b4-a477-0d802424e0aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79969402f5c33852a26fc839f37a45febf4f2377d77823284fbb01769e5585fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2d4d5de7106175087d91acb1f0c7552014f0c08bfcd65cbbf184d3f8d7826ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-99qgp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:37Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:37 crc kubenswrapper[4860]: I1211 08:11:37.107367 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-26nc6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1faa733-f408-41fd-b7aa-0cd86283976a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ac45db4f858a1c18a6b87472fc1cea2300f5845f1ad62c22260374995548a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ac45db4f858a1c18a6b87472fc1cea2300f5845f1ad62c22260374995548a0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b35296a5eb17dd2bb5e72c86c1cafdb30b7712cb330264f935d3b4fff05cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c6b35296a5eb17dd2bb5e72c86c1cafdb30b7712cb330264f935d3b4fff05cd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b9e82d6d3a247309ef5bdbf5b3a86ec083add2a138b0b237801e77b09ee96a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b9e82d6d3a247309ef5bdbf5b3a86ec083add2a138b0b237801e77b09ee96a1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://908ee695e78145ef4375c4097d9f1e772502922d02a2fcb4566172a6ad9caaf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://908ee695e78145ef4375c4097d9f1e772502922d02a2fcb4566172a6ad9caaf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afb0a097f3556a8f9ed7cf8f762cb99c082fe0a36b8d8d3038bf188b08fe9a84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afb0a097f3556a8f9ed7cf8f762cb99c082fe0a36b8d8d3038bf188b08fe9a84\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-26nc6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:37Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:37 crc kubenswrapper[4860]: I1211 08:11:37.130961 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0589a204-c98c-417d-8256-bfe3bf747660\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17cbcb49c06f3a9059cd5a0f05eedb8c785da30a3e502e31bae2f57b3486cb14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17cbcb49c06f3a9059cd5a0f05eedb8c785da30a3e502e31bae2f57b3486cb14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xvnqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:37Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:37 crc kubenswrapper[4860]: I1211 08:11:37.132435 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:37 crc kubenswrapper[4860]: I1211 08:11:37.132522 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:37 crc kubenswrapper[4860]: I1211 08:11:37.132543 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:37 crc kubenswrapper[4860]: I1211 08:11:37.132571 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:37 crc kubenswrapper[4860]: I1211 08:11:37.132589 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:37Z","lastTransitionTime":"2025-12-11T08:11:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:37 crc kubenswrapper[4860]: I1211 08:11:37.234917 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:37 crc kubenswrapper[4860]: I1211 08:11:37.234973 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:37 crc kubenswrapper[4860]: I1211 08:11:37.234990 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:37 crc kubenswrapper[4860]: I1211 08:11:37.235013 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:37 crc kubenswrapper[4860]: I1211 08:11:37.235030 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:37Z","lastTransitionTime":"2025-12-11T08:11:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:37 crc kubenswrapper[4860]: I1211 08:11:37.337502 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:37 crc kubenswrapper[4860]: I1211 08:11:37.337558 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:37 crc kubenswrapper[4860]: I1211 08:11:37.337578 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:37 crc kubenswrapper[4860]: I1211 08:11:37.337605 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:37 crc kubenswrapper[4860]: I1211 08:11:37.337620 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:37Z","lastTransitionTime":"2025-12-11T08:11:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:37 crc kubenswrapper[4860]: I1211 08:11:37.440769 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:37 crc kubenswrapper[4860]: I1211 08:11:37.440865 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:37 crc kubenswrapper[4860]: I1211 08:11:37.440892 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:37 crc kubenswrapper[4860]: I1211 08:11:37.440920 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:37 crc kubenswrapper[4860]: I1211 08:11:37.440938 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:37Z","lastTransitionTime":"2025-12-11T08:11:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:37 crc kubenswrapper[4860]: I1211 08:11:37.552264 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:37 crc kubenswrapper[4860]: I1211 08:11:37.552314 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:37 crc kubenswrapper[4860]: I1211 08:11:37.552332 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:37 crc kubenswrapper[4860]: I1211 08:11:37.552351 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:37 crc kubenswrapper[4860]: I1211 08:11:37.552363 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:37Z","lastTransitionTime":"2025-12-11T08:11:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:37 crc kubenswrapper[4860]: I1211 08:11:37.595716 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:37Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:37 crc kubenswrapper[4860]: I1211 08:11:37.614199 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d8c2f28628dc2f1ffbdd98a55670e22cf6cdc8cf2c6442e9d1f8e5466b2f12e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://074ea63beac52bd196507ee7cfb1c3d7fc2d37d5ec6faa4d47dde62b922e1c43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:37Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:37 crc kubenswrapper[4860]: I1211 08:11:37.631254 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:37Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:37 crc kubenswrapper[4860]: I1211 08:11:37.645963 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25a6a154da2b74952951ae94efc914ba078ea4bbe5893360b3c84d46be6bf973\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:37Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:37 crc kubenswrapper[4860]: I1211 08:11:37.657718 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:37 crc kubenswrapper[4860]: I1211 08:11:37.657762 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:37 crc kubenswrapper[4860]: I1211 08:11:37.657777 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:37 crc kubenswrapper[4860]: I1211 08:11:37.657797 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:37 crc kubenswrapper[4860]: I1211 08:11:37.657808 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:37Z","lastTransitionTime":"2025-12-11T08:11:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:37 crc kubenswrapper[4860]: I1211 08:11:37.664220 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6xq7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a75ce3c-5e1e-4b9f-bbde-7a3b0cce5eae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe8d4d675567b41728e476351811a93c47b19868ccbadab8c613a4eb8eca8562\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5c6nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6xq7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:37Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:37 crc kubenswrapper[4860]: I1211 08:11:37.680430 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ec3758-85f1-4a22-a4ec-c805472607a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e831db1a15976d3967d56479689572c26f2599a766baa0402897a3008428c56f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a498aa90f914e1f0370df01dc5ef174df113a2a16d86b2e3f24b19a1b99b4c0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8caebd2f4dfcbb89c597029a07c335171889761d2d200206987537fb0a228db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34e91f615b76445ff3241a304bb2964fde59de7e8a8dfff6841b2e628a31e86d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2090cab02ad74ddfd6bb57ff50eba35d18f56b0cde1c230fe011c01994b693bc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 08:11:21.171011 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 08:11:21.177674 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-644352688/tls.crt::/tmp/serving-cert-644352688/tls.key\\\\\\\"\\\\nI1211 08:11:26.508606 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 08:11:26.512741 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 08:11:26.512778 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 08:11:26.512839 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 08:11:26.512858 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 08:11:26.521684 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1211 08:11:26.521700 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1211 08:11:26.521722 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:11:26.521732 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:11:26.521738 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 08:11:26.521746 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 08:11:26.521749 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 08:11:26.521753 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1211 08:11:26.524999 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2274fb298ca7d2b37b060b4dbf9b166d39ba77a108b08edeeb4d67196e72cb3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf0aff8a6ab816b6617b1feeca9146d41bb5e8808bf1db96f8d493500c135d85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf0aff8a6ab816b6617b1feeca9146d41bb5e8808bf1db96f8d493500c135d85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:37Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:37 crc kubenswrapper[4860]: I1211 08:11:37.693749 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:37Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:37 crc kubenswrapper[4860]: I1211 08:11:37.708595 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a41f35db087418d1be16953f10377042abd3f4488e5ac1bc29ca5dcfc00a37e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:37Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:37 crc kubenswrapper[4860]: I1211 08:11:37.724529 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9aa0aed-62b3-4ea2-82ae-9a653c1101a5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ad4ba1b769c6cb2b34073798f8e3c25eb0177e6d3059484bf365e6981cab710\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://077aa1dd2fe5189ff60a0f2365e47a6a545cca0de7aa6379370908db4ea5838a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d7c919899b74a2b0fcd9221f3f769a9d013267354b200a282669d342d6413d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5483384f1d2ebc1fefa9706c9e67a3dc5a00ff643f803e3f2707c252e27ad94\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:37Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:37 crc kubenswrapper[4860]: I1211 08:11:37.735219 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31c30642-6e60-41b4-a477-0d802424e0aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79969402f5c33852a26fc839f37a45febf4f2377d77823284fbb01769e5585fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2d4d5de7106175087d91acb1f0c7552014f0c08bfcd65cbbf184d3f8d7826ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-99qgp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:37Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:37 crc kubenswrapper[4860]: I1211 08:11:37.753972 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-26nc6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1faa733-f408-41fd-b7aa-0cd86283976a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ac45db4f858a1c18a6b87472fc1cea2300f5845f1ad62c22260374995548a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ac45db4f858a1c18a6b87472fc1cea2300f5845f1ad62c22260374995548a0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b35296a5eb17dd2bb5e72c86c1cafdb30b7712cb330264f935d3b4fff05cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c6b35296a5eb17dd2bb5e72c86c1cafdb30b7712cb330264f935d3b4fff05cd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b9e82d6d3a247309ef5bdbf5b3a86ec083add2a138b0b237801e77b09ee96a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b9e82d6d3a247309ef5bdbf5b3a86ec083add2a138b0b237801e77b09ee96a1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://908ee695e78145ef4375c4097d9f1e772502922d02a2fcb4566172a6ad9caaf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://908ee695e78145ef4375c4097d9f1e772502922d02a2fcb4566172a6ad9caaf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afb0a097f3556a8f9ed7cf8f762cb99c082fe0a36b8d8d3038bf188b08fe9a84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afb0a097f3556a8f9ed7cf8f762cb99c082fe0a36b8d8d3038bf188b08fe9a84\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-26nc6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:37Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:37 crc kubenswrapper[4860]: I1211 08:11:37.762772 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:37 crc kubenswrapper[4860]: I1211 08:11:37.762799 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:37 crc kubenswrapper[4860]: I1211 08:11:37.762808 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:37 crc kubenswrapper[4860]: I1211 08:11:37.762821 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:37 crc kubenswrapper[4860]: I1211 08:11:37.762831 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:37Z","lastTransitionTime":"2025-12-11T08:11:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:37 crc kubenswrapper[4860]: I1211 08:11:37.773223 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0589a204-c98c-417d-8256-bfe3bf747660\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17cbcb49c06f3a9059cd5a0f05eedb8c785da30a3e502e31bae2f57b3486cb14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17cbcb49c06f3a9059cd5a0f05eedb8c785da30a3e502e31bae2f57b3486cb14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xvnqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:37Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:37 crc kubenswrapper[4860]: I1211 08:11:37.782882 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gsx9l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aba5b6b9-f88d-4763-9562-6b4dec606df9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://735ef012c523ddc952d46a431f27b2582a2c301d9e9da0c697a20ff6e2fb5d86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6zw9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gsx9l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:37Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:37 crc kubenswrapper[4860]: I1211 08:11:37.799971 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9872be1e-b995-4a00-904b-875dff96ad56\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f74a32a8fefea704cc73ffd1114083dab68aa44dc33f8d9d432d2d24216546c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbe00b46063d3968d59c393b87178e135f982abb7d0f32fa798b212d4b83df6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eedc9988e1a4a3c8cae730dc4a352103a6d665535b825e157d0f87c331b137c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2137fee4b7abc9765f9517d7b3b6ea796601a31d03151bf167d17136ced8b25a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f683e48f3577a68862eb6302383538376e41f49b20b1e0ea3f1b3b143f24872a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5de794187456b3576b969069a3b839d75c7918418a97fa35dff8bb40bb76e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5de794187456b3576b969069a3b839d75c7918418a97fa35dff8bb40bb76e39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08ce42b974fba00c409c0af971f0de351e50f18b2f06f8e722150da3239bb908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08ce42b974fba00c409c0af971f0de351e50f18b2f06f8e722150da3239bb908\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cfac6be03e3cfa0b0f3450b48b67957f2d5172b9c65e27f14c2a260bda5d423f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cfac6be03e3cfa0b0f3450b48b67957f2d5172b9c65e27f14c2a260bda5d423f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:37Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:37 crc kubenswrapper[4860]: I1211 08:11:37.815809 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-pznwl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5310c1fc-66c6-40aa-b1bf-5a59a2410f9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2d6ff00ce322a28607bc3f6db05c5e4fb14acb6777ea315786ef53963ca504a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hkt48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-pznwl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:37Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:37 crc kubenswrapper[4860]: I1211 08:11:37.853706 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" event={"ID":"0589a204-c98c-417d-8256-bfe3bf747660","Type":"ContainerStarted","Data":"526c2b9214d19b9ebe7a7078fd0459642a53e732d034602b2b8d2b8a33f4d18b"} Dec 11 08:11:37 crc kubenswrapper[4860]: I1211 08:11:37.854060 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" Dec 11 08:11:37 crc kubenswrapper[4860]: I1211 08:11:37.865913 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:37 crc kubenswrapper[4860]: I1211 08:11:37.865950 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:37 crc kubenswrapper[4860]: I1211 08:11:37.865958 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:37 crc kubenswrapper[4860]: I1211 08:11:37.865971 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:37 crc kubenswrapper[4860]: I1211 08:11:37.865981 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:37Z","lastTransitionTime":"2025-12-11T08:11:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:37 crc kubenswrapper[4860]: I1211 08:11:37.866842 4860 generic.go:334] "Generic (PLEG): container finished" podID="a1faa733-f408-41fd-b7aa-0cd86283976a" containerID="ce18f4f06db091e658684fa0bbc858e19b9f03010acf13b3ffccb1a6e3f88021" exitCode=0 Dec 11 08:11:37 crc kubenswrapper[4860]: I1211 08:11:37.866879 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-26nc6" event={"ID":"a1faa733-f408-41fd-b7aa-0cd86283976a","Type":"ContainerDied","Data":"ce18f4f06db091e658684fa0bbc858e19b9f03010acf13b3ffccb1a6e3f88021"} Dec 11 08:11:37 crc kubenswrapper[4860]: I1211 08:11:37.867556 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9aa0aed-62b3-4ea2-82ae-9a653c1101a5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ad4ba1b769c6cb2b34073798f8e3c25eb0177e6d3059484bf365e6981cab710\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://077aa1dd2fe5189ff60a0f2365e47a6a545cca0de7aa6379370908db4ea5838a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d7c919899b74a2b0fcd9221f3f769a9d013267354b200a282669d342d6413d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5483384f1d2ebc1fefa9706c9e67a3dc5a00ff643f803e3f2707c252e27ad94\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:37Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:37 crc kubenswrapper[4860]: I1211 08:11:37.880831 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31c30642-6e60-41b4-a477-0d802424e0aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79969402f5c33852a26fc839f37a45febf4f2377d77823284fbb01769e5585fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2d4d5de7106175087d91acb1f0c7552014f0c08bfcd65cbbf184d3f8d7826ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-99qgp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:37Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:37 crc kubenswrapper[4860]: I1211 08:11:37.903031 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-26nc6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1faa733-f408-41fd-b7aa-0cd86283976a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ac45db4f858a1c18a6b87472fc1cea2300f5845f1ad62c22260374995548a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ac45db4f858a1c18a6b87472fc1cea2300f5845f1ad62c22260374995548a0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b35296a5eb17dd2bb5e72c86c1cafdb30b7712cb330264f935d3b4fff05cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c6b35296a5eb17dd2bb5e72c86c1cafdb30b7712cb330264f935d3b4fff05cd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b9e82d6d3a247309ef5bdbf5b3a86ec083add2a138b0b237801e77b09ee96a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b9e82d6d3a247309ef5bdbf5b3a86ec083add2a138b0b237801e77b09ee96a1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://908ee695e78145ef4375c4097d9f1e772502922d02a2fcb4566172a6ad9caaf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://908ee695e78145ef4375c4097d9f1e772502922d02a2fcb4566172a6ad9caaf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afb0a097f3556a8f9ed7cf8f762cb99c082fe0a36b8d8d3038bf188b08fe9a84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afb0a097f3556a8f9ed7cf8f762cb99c082fe0a36b8d8d3038bf188b08fe9a84\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-26nc6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:37Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:37 crc kubenswrapper[4860]: I1211 08:11:37.908861 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" Dec 11 08:11:37 crc kubenswrapper[4860]: I1211 08:11:37.930338 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0589a204-c98c-417d-8256-bfe3bf747660\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3333b4346f59b04017e32d61f457ba7ff457166731128f339d7987ae6b56b06d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52bb8ade314afb9494f4e026d0af2aea7b1970d64dc66102f55e257ae4678154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dfcc4370edaae36a5a32b0a9c29290c50f8e4f427e4e147dc3c779beaa8b780\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d856ae95e9077524c518d82d10fb6b93b7fbcaaf2b720ffafe6f7fae58bd7caa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1426084ac580895c42354630910dd8c8e3a48ddd25f22abce1d28dc3de332a63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9984d5d242bb258405efe91fc9caf9e7ea592834445df2e78add9553a921c76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://526c2b9214d19b9ebe7a7078fd0459642a53e732d034602b2b8d2b8a33f4d18b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb80d9928e8349fc5963227d135355ad7ba820d4fd89877af601d005ca03334b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17cbcb49c06f3a9059cd5a0f05eedb8c785da30a3e502e31bae2f57b3486cb14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17cbcb49c06f3a9059cd5a0f05eedb8c785da30a3e502e31bae2f57b3486cb14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xvnqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:37Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:37 crc kubenswrapper[4860]: I1211 08:11:37.941524 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gsx9l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aba5b6b9-f88d-4763-9562-6b4dec606df9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://735ef012c523ddc952d46a431f27b2582a2c301d9e9da0c697a20ff6e2fb5d86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6zw9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gsx9l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:37Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:37 crc kubenswrapper[4860]: I1211 08:11:37.960564 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9872be1e-b995-4a00-904b-875dff96ad56\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f74a32a8fefea704cc73ffd1114083dab68aa44dc33f8d9d432d2d24216546c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbe00b46063d3968d59c393b87178e135f982abb7d0f32fa798b212d4b83df6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eedc9988e1a4a3c8cae730dc4a352103a6d665535b825e157d0f87c331b137c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2137fee4b7abc9765f9517d7b3b6ea796601a31d03151bf167d17136ced8b25a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f683e48f3577a68862eb6302383538376e41f49b20b1e0ea3f1b3b143f24872a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5de794187456b3576b969069a3b839d75c7918418a97fa35dff8bb40bb76e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5de794187456b3576b969069a3b839d75c7918418a97fa35dff8bb40bb76e39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08ce42b974fba00c409c0af971f0de351e50f18b2f06f8e722150da3239bb908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08ce42b974fba00c409c0af971f0de351e50f18b2f06f8e722150da3239bb908\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cfac6be03e3cfa0b0f3450b48b67957f2d5172b9c65e27f14c2a260bda5d423f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cfac6be03e3cfa0b0f3450b48b67957f2d5172b9c65e27f14c2a260bda5d423f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:37Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:37 crc kubenswrapper[4860]: I1211 08:11:37.970015 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:37 crc kubenswrapper[4860]: I1211 08:11:37.970052 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:37 crc kubenswrapper[4860]: I1211 08:11:37.970061 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:37 crc kubenswrapper[4860]: I1211 08:11:37.970074 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:37 crc kubenswrapper[4860]: I1211 08:11:37.970083 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:37Z","lastTransitionTime":"2025-12-11T08:11:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:37 crc kubenswrapper[4860]: I1211 08:11:37.975016 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-pznwl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5310c1fc-66c6-40aa-b1bf-5a59a2410f9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2d6ff00ce322a28607bc3f6db05c5e4fb14acb6777ea315786ef53963ca504a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hkt48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-pznwl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:37Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:37 crc kubenswrapper[4860]: I1211 08:11:37.988628 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:37Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:38 crc kubenswrapper[4860]: I1211 08:11:38.019919 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d8c2f28628dc2f1ffbdd98a55670e22cf6cdc8cf2c6442e9d1f8e5466b2f12e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://074ea63beac52bd196507ee7cfb1c3d7fc2d37d5ec6faa4d47dde62b922e1c43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:38Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:38 crc kubenswrapper[4860]: I1211 08:11:38.044404 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:38Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:38 crc kubenswrapper[4860]: I1211 08:11:38.065695 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25a6a154da2b74952951ae94efc914ba078ea4bbe5893360b3c84d46be6bf973\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:38Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:38 crc kubenswrapper[4860]: I1211 08:11:38.072521 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:38 crc kubenswrapper[4860]: I1211 08:11:38.072572 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:38 crc kubenswrapper[4860]: I1211 08:11:38.072584 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:38 crc kubenswrapper[4860]: I1211 08:11:38.072603 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:38 crc kubenswrapper[4860]: I1211 08:11:38.072615 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:38Z","lastTransitionTime":"2025-12-11T08:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:38 crc kubenswrapper[4860]: I1211 08:11:38.076994 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6xq7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a75ce3c-5e1e-4b9f-bbde-7a3b0cce5eae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe8d4d675567b41728e476351811a93c47b19868ccbadab8c613a4eb8eca8562\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5c6nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6xq7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:38Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:38 crc kubenswrapper[4860]: I1211 08:11:38.090076 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ec3758-85f1-4a22-a4ec-c805472607a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e831db1a15976d3967d56479689572c26f2599a766baa0402897a3008428c56f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a498aa90f914e1f0370df01dc5ef174df113a2a16d86b2e3f24b19a1b99b4c0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8caebd2f4dfcbb89c597029a07c335171889761d2d200206987537fb0a228db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34e91f615b76445ff3241a304bb2964fde59de7e8a8dfff6841b2e628a31e86d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2090cab02ad74ddfd6bb57ff50eba35d18f56b0cde1c230fe011c01994b693bc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 08:11:21.171011 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 08:11:21.177674 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-644352688/tls.crt::/tmp/serving-cert-644352688/tls.key\\\\\\\"\\\\nI1211 08:11:26.508606 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 08:11:26.512741 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 08:11:26.512778 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 08:11:26.512839 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 08:11:26.512858 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 08:11:26.521684 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1211 08:11:26.521700 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1211 08:11:26.521722 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:11:26.521732 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:11:26.521738 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 08:11:26.521746 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 08:11:26.521749 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 08:11:26.521753 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1211 08:11:26.524999 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2274fb298ca7d2b37b060b4dbf9b166d39ba77a108b08edeeb4d67196e72cb3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf0aff8a6ab816b6617b1feeca9146d41bb5e8808bf1db96f8d493500c135d85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf0aff8a6ab816b6617b1feeca9146d41bb5e8808bf1db96f8d493500c135d85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:38Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:38 crc kubenswrapper[4860]: I1211 08:11:38.102851 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:38Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:38 crc kubenswrapper[4860]: I1211 08:11:38.114248 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a41f35db087418d1be16953f10377042abd3f4488e5ac1bc29ca5dcfc00a37e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:38Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:38 crc kubenswrapper[4860]: I1211 08:11:38.135865 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9872be1e-b995-4a00-904b-875dff96ad56\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f74a32a8fefea704cc73ffd1114083dab68aa44dc33f8d9d432d2d24216546c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbe00b46063d3968d59c393b87178e135f982abb7d0f32fa798b212d4b83df6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eedc9988e1a4a3c8cae730dc4a352103a6d665535b825e157d0f87c331b137c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2137fee4b7abc9765f9517d7b3b6ea796601a31d03151bf167d17136ced8b25a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f683e48f3577a68862eb6302383538376e41f49b20b1e0ea3f1b3b143f24872a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5de794187456b3576b969069a3b839d75c7918418a97fa35dff8bb40bb76e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5de794187456b3576b969069a3b839d75c7918418a97fa35dff8bb40bb76e39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08ce42b974fba00c409c0af971f0de351e50f18b2f06f8e722150da3239bb908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08ce42b974fba00c409c0af971f0de351e50f18b2f06f8e722150da3239bb908\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cfac6be03e3cfa0b0f3450b48b67957f2d5172b9c65e27f14c2a260bda5d423f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cfac6be03e3cfa0b0f3450b48b67957f2d5172b9c65e27f14c2a260bda5d423f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:38Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:38 crc kubenswrapper[4860]: I1211 08:11:38.150436 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-pznwl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5310c1fc-66c6-40aa-b1bf-5a59a2410f9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2d6ff00ce322a28607bc3f6db05c5e4fb14acb6777ea315786ef53963ca504a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hkt48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-pznwl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:38Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:38 crc kubenswrapper[4860]: I1211 08:11:38.165379 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d8c2f28628dc2f1ffbdd98a55670e22cf6cdc8cf2c6442e9d1f8e5466b2f12e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://074ea63beac52bd196507ee7cfb1c3d7fc2d37d5ec6faa4d47dde62b922e1c43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:38Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:38 crc kubenswrapper[4860]: I1211 08:11:38.175208 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:38 crc kubenswrapper[4860]: I1211 08:11:38.175267 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:38 crc kubenswrapper[4860]: I1211 08:11:38.175283 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:38 crc kubenswrapper[4860]: I1211 08:11:38.175305 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:38 crc kubenswrapper[4860]: I1211 08:11:38.175317 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:38Z","lastTransitionTime":"2025-12-11T08:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:38 crc kubenswrapper[4860]: I1211 08:11:38.176633 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:38Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:38 crc kubenswrapper[4860]: I1211 08:11:38.188927 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25a6a154da2b74952951ae94efc914ba078ea4bbe5893360b3c84d46be6bf973\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:38Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:38 crc kubenswrapper[4860]: I1211 08:11:38.198887 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6xq7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a75ce3c-5e1e-4b9f-bbde-7a3b0cce5eae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe8d4d675567b41728e476351811a93c47b19868ccbadab8c613a4eb8eca8562\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5c6nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6xq7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:38Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:38 crc kubenswrapper[4860]: I1211 08:11:38.209430 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:38Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:38 crc kubenswrapper[4860]: I1211 08:11:38.222025 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a41f35db087418d1be16953f10377042abd3f4488e5ac1bc29ca5dcfc00a37e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:38Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:38 crc kubenswrapper[4860]: I1211 08:11:38.235542 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ec3758-85f1-4a22-a4ec-c805472607a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e831db1a15976d3967d56479689572c26f2599a766baa0402897a3008428c56f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a498aa90f914e1f0370df01dc5ef174df113a2a16d86b2e3f24b19a1b99b4c0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8caebd2f4dfcbb89c597029a07c335171889761d2d200206987537fb0a228db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34e91f615b76445ff3241a304bb2964fde59de7e8a8dfff6841b2e628a31e86d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2090cab02ad74ddfd6bb57ff50eba35d18f56b0cde1c230fe011c01994b693bc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 08:11:21.171011 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 08:11:21.177674 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-644352688/tls.crt::/tmp/serving-cert-644352688/tls.key\\\\\\\"\\\\nI1211 08:11:26.508606 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 08:11:26.512741 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 08:11:26.512778 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 08:11:26.512839 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 08:11:26.512858 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 08:11:26.521684 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1211 08:11:26.521700 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1211 08:11:26.521722 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:11:26.521732 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:11:26.521738 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 08:11:26.521746 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 08:11:26.521749 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 08:11:26.521753 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1211 08:11:26.524999 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2274fb298ca7d2b37b060b4dbf9b166d39ba77a108b08edeeb4d67196e72cb3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf0aff8a6ab816b6617b1feeca9146d41bb5e8808bf1db96f8d493500c135d85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf0aff8a6ab816b6617b1feeca9146d41bb5e8808bf1db96f8d493500c135d85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:38Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:38 crc kubenswrapper[4860]: I1211 08:11:38.247882 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:38Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:38 crc kubenswrapper[4860]: I1211 08:11:38.268776 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31c30642-6e60-41b4-a477-0d802424e0aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79969402f5c33852a26fc839f37a45febf4f2377d77823284fbb01769e5585fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2d4d5de7106175087d91acb1f0c7552014f0c08bfcd65cbbf184d3f8d7826ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-99qgp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:38Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:38 crc kubenswrapper[4860]: I1211 08:11:38.277259 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:38 crc kubenswrapper[4860]: I1211 08:11:38.277386 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:38 crc kubenswrapper[4860]: I1211 08:11:38.277448 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:38 crc kubenswrapper[4860]: I1211 08:11:38.277513 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:38 crc kubenswrapper[4860]: I1211 08:11:38.277570 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:38Z","lastTransitionTime":"2025-12-11T08:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:38 crc kubenswrapper[4860]: I1211 08:11:38.285252 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-26nc6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1faa733-f408-41fd-b7aa-0cd86283976a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ac45db4f858a1c18a6b87472fc1cea2300f5845f1ad62c22260374995548a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ac45db4f858a1c18a6b87472fc1cea2300f5845f1ad62c22260374995548a0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b35296a5eb17dd2bb5e72c86c1cafdb30b7712cb330264f935d3b4fff05cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c6b35296a5eb17dd2bb5e72c86c1cafdb30b7712cb330264f935d3b4fff05cd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b9e82d6d3a247309ef5bdbf5b3a86ec083add2a138b0b237801e77b09ee96a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b9e82d6d3a247309ef5bdbf5b3a86ec083add2a138b0b237801e77b09ee96a1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://908ee695e78145ef4375c4097d9f1e772502922d02a2fcb4566172a6ad9caaf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://908ee695e78145ef4375c4097d9f1e772502922d02a2fcb4566172a6ad9caaf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afb0a097f3556a8f9ed7cf8f762cb99c082fe0a36b8d8d3038bf188b08fe9a84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afb0a097f3556a8f9ed7cf8f762cb99c082fe0a36b8d8d3038bf188b08fe9a84\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce18f4f06db091e658684fa0bbc858e19b9f03010acf13b3ffccb1a6e3f88021\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce18f4f06db091e658684fa0bbc858e19b9f03010acf13b3ffccb1a6e3f88021\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-26nc6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:38Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:38 crc kubenswrapper[4860]: I1211 08:11:38.307409 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0589a204-c98c-417d-8256-bfe3bf747660\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3333b4346f59b04017e32d61f457ba7ff457166731128f339d7987ae6b56b06d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52bb8ade314afb9494f4e026d0af2aea7b1970d64dc66102f55e257ae4678154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dfcc4370edaae36a5a32b0a9c29290c50f8e4f427e4e147dc3c779beaa8b780\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d856ae95e9077524c518d82d10fb6b93b7fbcaaf2b720ffafe6f7fae58bd7caa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1426084ac580895c42354630910dd8c8e3a48ddd25f22abce1d28dc3de332a63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9984d5d242bb258405efe91fc9caf9e7ea592834445df2e78add9553a921c76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://526c2b9214d19b9ebe7a7078fd0459642a53e732d034602b2b8d2b8a33f4d18b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb80d9928e8349fc5963227d135355ad7ba820d4fd89877af601d005ca03334b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17cbcb49c06f3a9059cd5a0f05eedb8c785da30a3e502e31bae2f57b3486cb14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17cbcb49c06f3a9059cd5a0f05eedb8c785da30a3e502e31bae2f57b3486cb14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xvnqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:38Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:38 crc kubenswrapper[4860]: I1211 08:11:38.321923 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gsx9l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aba5b6b9-f88d-4763-9562-6b4dec606df9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://735ef012c523ddc952d46a431f27b2582a2c301d9e9da0c697a20ff6e2fb5d86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6zw9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gsx9l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:38Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:38 crc kubenswrapper[4860]: I1211 08:11:38.338787 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9aa0aed-62b3-4ea2-82ae-9a653c1101a5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ad4ba1b769c6cb2b34073798f8e3c25eb0177e6d3059484bf365e6981cab710\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://077aa1dd2fe5189ff60a0f2365e47a6a545cca0de7aa6379370908db4ea5838a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d7c919899b74a2b0fcd9221f3f769a9d013267354b200a282669d342d6413d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5483384f1d2ebc1fefa9706c9e67a3dc5a00ff643f803e3f2707c252e27ad94\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:38Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:38 crc kubenswrapper[4860]: I1211 08:11:38.380298 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:38 crc kubenswrapper[4860]: I1211 08:11:38.380447 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:38 crc kubenswrapper[4860]: I1211 08:11:38.380515 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:38 crc kubenswrapper[4860]: I1211 08:11:38.380584 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:38 crc kubenswrapper[4860]: I1211 08:11:38.380661 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:38Z","lastTransitionTime":"2025-12-11T08:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:38 crc kubenswrapper[4860]: I1211 08:11:38.484107 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:38 crc kubenswrapper[4860]: I1211 08:11:38.484139 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:38 crc kubenswrapper[4860]: I1211 08:11:38.484149 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:38 crc kubenswrapper[4860]: I1211 08:11:38.484161 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:38 crc kubenswrapper[4860]: I1211 08:11:38.484170 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:38Z","lastTransitionTime":"2025-12-11T08:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:38 crc kubenswrapper[4860]: I1211 08:11:38.578384 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:11:38 crc kubenswrapper[4860]: I1211 08:11:38.578440 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:11:38 crc kubenswrapper[4860]: I1211 08:11:38.578447 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:11:38 crc kubenswrapper[4860]: E1211 08:11:38.578556 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:11:38 crc kubenswrapper[4860]: E1211 08:11:38.578716 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:11:38 crc kubenswrapper[4860]: E1211 08:11:38.578921 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:11:38 crc kubenswrapper[4860]: I1211 08:11:38.585971 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:38 crc kubenswrapper[4860]: I1211 08:11:38.586172 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:38 crc kubenswrapper[4860]: I1211 08:11:38.586253 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:38 crc kubenswrapper[4860]: I1211 08:11:38.586341 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:38 crc kubenswrapper[4860]: I1211 08:11:38.586420 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:38Z","lastTransitionTime":"2025-12-11T08:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:38 crc kubenswrapper[4860]: I1211 08:11:38.689551 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:38 crc kubenswrapper[4860]: I1211 08:11:38.689888 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:38 crc kubenswrapper[4860]: I1211 08:11:38.689897 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:38 crc kubenswrapper[4860]: I1211 08:11:38.689922 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:38 crc kubenswrapper[4860]: I1211 08:11:38.689933 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:38Z","lastTransitionTime":"2025-12-11T08:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:38 crc kubenswrapper[4860]: I1211 08:11:38.792209 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:38 crc kubenswrapper[4860]: I1211 08:11:38.792276 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:38 crc kubenswrapper[4860]: I1211 08:11:38.792291 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:38 crc kubenswrapper[4860]: I1211 08:11:38.792313 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:38 crc kubenswrapper[4860]: I1211 08:11:38.792328 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:38Z","lastTransitionTime":"2025-12-11T08:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:38 crc kubenswrapper[4860]: I1211 08:11:38.875339 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-26nc6" event={"ID":"a1faa733-f408-41fd-b7aa-0cd86283976a","Type":"ContainerStarted","Data":"fce5b642b843ed30823c94e0b0c3d8bb1e5d2d93d691f7b8ef901ece03b0ac6c"} Dec 11 08:11:38 crc kubenswrapper[4860]: I1211 08:11:38.875451 4860 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 11 08:11:38 crc kubenswrapper[4860]: I1211 08:11:38.876268 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" Dec 11 08:11:38 crc kubenswrapper[4860]: I1211 08:11:38.896154 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25a6a154da2b74952951ae94efc914ba078ea4bbe5893360b3c84d46be6bf973\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:38Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:38 crc kubenswrapper[4860]: I1211 08:11:38.896794 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:38 crc kubenswrapper[4860]: I1211 08:11:38.896830 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:38 crc kubenswrapper[4860]: I1211 08:11:38.896847 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:38 crc kubenswrapper[4860]: I1211 08:11:38.896863 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:38 crc kubenswrapper[4860]: I1211 08:11:38.896875 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:38Z","lastTransitionTime":"2025-12-11T08:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:38 crc kubenswrapper[4860]: I1211 08:11:38.914878 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6xq7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a75ce3c-5e1e-4b9f-bbde-7a3b0cce5eae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe8d4d675567b41728e476351811a93c47b19868ccbadab8c613a4eb8eca8562\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5c6nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6xq7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:38Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:38 crc kubenswrapper[4860]: I1211 08:11:38.927803 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" Dec 11 08:11:38 crc kubenswrapper[4860]: I1211 08:11:38.934167 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:38Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:38 crc kubenswrapper[4860]: I1211 08:11:38.952388 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d8c2f28628dc2f1ffbdd98a55670e22cf6cdc8cf2c6442e9d1f8e5466b2f12e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://074ea63beac52bd196507ee7cfb1c3d7fc2d37d5ec6faa4d47dde62b922e1c43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:38Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:38 crc kubenswrapper[4860]: I1211 08:11:38.970531 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:38Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:38 crc kubenswrapper[4860]: I1211 08:11:38.992701 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ec3758-85f1-4a22-a4ec-c805472607a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e831db1a15976d3967d56479689572c26f2599a766baa0402897a3008428c56f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a498aa90f914e1f0370df01dc5ef174df113a2a16d86b2e3f24b19a1b99b4c0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8caebd2f4dfcbb89c597029a07c335171889761d2d200206987537fb0a228db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34e91f615b76445ff3241a304bb2964fde59de7e8a8dfff6841b2e628a31e86d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2090cab02ad74ddfd6bb57ff50eba35d18f56b0cde1c230fe011c01994b693bc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 08:11:21.171011 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 08:11:21.177674 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-644352688/tls.crt::/tmp/serving-cert-644352688/tls.key\\\\\\\"\\\\nI1211 08:11:26.508606 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 08:11:26.512741 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 08:11:26.512778 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 08:11:26.512839 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 08:11:26.512858 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 08:11:26.521684 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1211 08:11:26.521700 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1211 08:11:26.521722 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:11:26.521732 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:11:26.521738 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 08:11:26.521746 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 08:11:26.521749 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 08:11:26.521753 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1211 08:11:26.524999 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2274fb298ca7d2b37b060b4dbf9b166d39ba77a108b08edeeb4d67196e72cb3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf0aff8a6ab816b6617b1feeca9146d41bb5e8808bf1db96f8d493500c135d85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf0aff8a6ab816b6617b1feeca9146d41bb5e8808bf1db96f8d493500c135d85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:38Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:39 crc kubenswrapper[4860]: I1211 08:11:39.003327 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:39 crc kubenswrapper[4860]: I1211 08:11:39.003380 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:39 crc kubenswrapper[4860]: I1211 08:11:39.003408 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:39 crc kubenswrapper[4860]: I1211 08:11:39.003433 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:39 crc kubenswrapper[4860]: I1211 08:11:39.003451 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:39Z","lastTransitionTime":"2025-12-11T08:11:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:39 crc kubenswrapper[4860]: I1211 08:11:39.006920 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:39Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:39 crc kubenswrapper[4860]: I1211 08:11:39.022088 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a41f35db087418d1be16953f10377042abd3f4488e5ac1bc29ca5dcfc00a37e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:39Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:39 crc kubenswrapper[4860]: I1211 08:11:39.048799 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0589a204-c98c-417d-8256-bfe3bf747660\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3333b4346f59b04017e32d61f457ba7ff457166731128f339d7987ae6b56b06d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52bb8ade314afb9494f4e026d0af2aea7b1970d64dc66102f55e257ae4678154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dfcc4370edaae36a5a32b0a9c29290c50f8e4f427e4e147dc3c779beaa8b780\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d856ae95e9077524c518d82d10fb6b93b7fbcaaf2b720ffafe6f7fae58bd7caa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1426084ac580895c42354630910dd8c8e3a48ddd25f22abce1d28dc3de332a63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9984d5d242bb258405efe91fc9caf9e7ea592834445df2e78add9553a921c76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://526c2b9214d19b9ebe7a7078fd0459642a53e732d034602b2b8d2b8a33f4d18b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb80d9928e8349fc5963227d135355ad7ba820d4fd89877af601d005ca03334b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17cbcb49c06f3a9059cd5a0f05eedb8c785da30a3e502e31bae2f57b3486cb14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17cbcb49c06f3a9059cd5a0f05eedb8c785da30a3e502e31bae2f57b3486cb14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xvnqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:39Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:39 crc kubenswrapper[4860]: I1211 08:11:39.061815 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gsx9l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aba5b6b9-f88d-4763-9562-6b4dec606df9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://735ef012c523ddc952d46a431f27b2582a2c301d9e9da0c697a20ff6e2fb5d86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6zw9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gsx9l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:39Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:39 crc kubenswrapper[4860]: I1211 08:11:39.078146 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9aa0aed-62b3-4ea2-82ae-9a653c1101a5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ad4ba1b769c6cb2b34073798f8e3c25eb0177e6d3059484bf365e6981cab710\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://077aa1dd2fe5189ff60a0f2365e47a6a545cca0de7aa6379370908db4ea5838a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d7c919899b74a2b0fcd9221f3f769a9d013267354b200a282669d342d6413d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5483384f1d2ebc1fefa9706c9e67a3dc5a00ff643f803e3f2707c252e27ad94\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:39Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:39 crc kubenswrapper[4860]: I1211 08:11:39.094221 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31c30642-6e60-41b4-a477-0d802424e0aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79969402f5c33852a26fc839f37a45febf4f2377d77823284fbb01769e5585fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2d4d5de7106175087d91acb1f0c7552014f0c08bfcd65cbbf184d3f8d7826ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-99qgp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:39Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:39 crc kubenswrapper[4860]: I1211 08:11:39.111849 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:39 crc kubenswrapper[4860]: I1211 08:11:39.111910 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:39 crc kubenswrapper[4860]: I1211 08:11:39.111932 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:39 crc kubenswrapper[4860]: I1211 08:11:39.111957 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:39 crc kubenswrapper[4860]: I1211 08:11:39.111978 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:39Z","lastTransitionTime":"2025-12-11T08:11:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:39 crc kubenswrapper[4860]: I1211 08:11:39.118890 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-26nc6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1faa733-f408-41fd-b7aa-0cd86283976a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fce5b642b843ed30823c94e0b0c3d8bb1e5d2d93d691f7b8ef901ece03b0ac6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ac45db4f858a1c18a6b87472fc1cea2300f5845f1ad62c22260374995548a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ac45db4f858a1c18a6b87472fc1cea2300f5845f1ad62c22260374995548a0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b35296a5eb17dd2bb5e72c86c1cafdb30b7712cb330264f935d3b4fff05cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c6b35296a5eb17dd2bb5e72c86c1cafdb30b7712cb330264f935d3b4fff05cd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b9e82d6d3a247309ef5bdbf5b3a86ec083add2a138b0b237801e77b09ee96a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b9e82d6d3a247309ef5bdbf5b3a86ec083add2a138b0b237801e77b09ee96a1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://908ee695e78145ef4375c4097d9f1e772502922d02a2fcb4566172a6ad9caaf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://908ee695e78145ef4375c4097d9f1e772502922d02a2fcb4566172a6ad9caaf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afb0a097f3556a8f9ed7cf8f762cb99c082fe0a36b8d8d3038bf188b08fe9a84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afb0a097f3556a8f9ed7cf8f762cb99c082fe0a36b8d8d3038bf188b08fe9a84\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce18f4f06db091e658684fa0bbc858e19b9f03010acf13b3ffccb1a6e3f88021\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce18f4f06db091e658684fa0bbc858e19b9f03010acf13b3ffccb1a6e3f88021\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-26nc6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:39Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:39 crc kubenswrapper[4860]: I1211 08:11:39.147439 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9872be1e-b995-4a00-904b-875dff96ad56\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f74a32a8fefea704cc73ffd1114083dab68aa44dc33f8d9d432d2d24216546c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbe00b46063d3968d59c393b87178e135f982abb7d0f32fa798b212d4b83df6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eedc9988e1a4a3c8cae730dc4a352103a6d665535b825e157d0f87c331b137c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2137fee4b7abc9765f9517d7b3b6ea796601a31d03151bf167d17136ced8b25a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f683e48f3577a68862eb6302383538376e41f49b20b1e0ea3f1b3b143f24872a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5de794187456b3576b969069a3b839d75c7918418a97fa35dff8bb40bb76e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5de794187456b3576b969069a3b839d75c7918418a97fa35dff8bb40bb76e39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08ce42b974fba00c409c0af971f0de351e50f18b2f06f8e722150da3239bb908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08ce42b974fba00c409c0af971f0de351e50f18b2f06f8e722150da3239bb908\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cfac6be03e3cfa0b0f3450b48b67957f2d5172b9c65e27f14c2a260bda5d423f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cfac6be03e3cfa0b0f3450b48b67957f2d5172b9c65e27f14c2a260bda5d423f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:39Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:39 crc kubenswrapper[4860]: I1211 08:11:39.161561 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-pznwl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5310c1fc-66c6-40aa-b1bf-5a59a2410f9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2d6ff00ce322a28607bc3f6db05c5e4fb14acb6777ea315786ef53963ca504a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hkt48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-pznwl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:39Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:39 crc kubenswrapper[4860]: I1211 08:11:39.173117 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31c30642-6e60-41b4-a477-0d802424e0aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79969402f5c33852a26fc839f37a45febf4f2377d77823284fbb01769e5585fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2d4d5de7106175087d91acb1f0c7552014f0c08bfcd65cbbf184d3f8d7826ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-99qgp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:39Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:39 crc kubenswrapper[4860]: I1211 08:11:39.190988 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-26nc6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1faa733-f408-41fd-b7aa-0cd86283976a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fce5b642b843ed30823c94e0b0c3d8bb1e5d2d93d691f7b8ef901ece03b0ac6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ac45db4f858a1c18a6b87472fc1cea2300f5845f1ad62c22260374995548a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ac45db4f858a1c18a6b87472fc1cea2300f5845f1ad62c22260374995548a0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b35296a5eb17dd2bb5e72c86c1cafdb30b7712cb330264f935d3b4fff05cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c6b35296a5eb17dd2bb5e72c86c1cafdb30b7712cb330264f935d3b4fff05cd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b9e82d6d3a247309ef5bdbf5b3a86ec083add2a138b0b237801e77b09ee96a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b9e82d6d3a247309ef5bdbf5b3a86ec083add2a138b0b237801e77b09ee96a1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://908ee695e78145ef4375c4097d9f1e772502922d02a2fcb4566172a6ad9caaf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://908ee695e78145ef4375c4097d9f1e772502922d02a2fcb4566172a6ad9caaf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afb0a097f3556a8f9ed7cf8f762cb99c082fe0a36b8d8d3038bf188b08fe9a84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afb0a097f3556a8f9ed7cf8f762cb99c082fe0a36b8d8d3038bf188b08fe9a84\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce18f4f06db091e658684fa0bbc858e19b9f03010acf13b3ffccb1a6e3f88021\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce18f4f06db091e658684fa0bbc858e19b9f03010acf13b3ffccb1a6e3f88021\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-26nc6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:39Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:39 crc kubenswrapper[4860]: I1211 08:11:39.211105 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0589a204-c98c-417d-8256-bfe3bf747660\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3333b4346f59b04017e32d61f457ba7ff457166731128f339d7987ae6b56b06d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52bb8ade314afb9494f4e026d0af2aea7b1970d64dc66102f55e257ae4678154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dfcc4370edaae36a5a32b0a9c29290c50f8e4f427e4e147dc3c779beaa8b780\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d856ae95e9077524c518d82d10fb6b93b7fbcaaf2b720ffafe6f7fae58bd7caa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1426084ac580895c42354630910dd8c8e3a48ddd25f22abce1d28dc3de332a63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9984d5d242bb258405efe91fc9caf9e7ea592834445df2e78add9553a921c76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://526c2b9214d19b9ebe7a7078fd0459642a53e732d034602b2b8d2b8a33f4d18b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb80d9928e8349fc5963227d135355ad7ba820d4fd89877af601d005ca03334b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17cbcb49c06f3a9059cd5a0f05eedb8c785da30a3e502e31bae2f57b3486cb14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17cbcb49c06f3a9059cd5a0f05eedb8c785da30a3e502e31bae2f57b3486cb14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xvnqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:39Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:39 crc kubenswrapper[4860]: I1211 08:11:39.214303 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:39 crc kubenswrapper[4860]: I1211 08:11:39.214335 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:39 crc kubenswrapper[4860]: I1211 08:11:39.214347 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:39 crc kubenswrapper[4860]: I1211 08:11:39.214363 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:39 crc kubenswrapper[4860]: I1211 08:11:39.214375 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:39Z","lastTransitionTime":"2025-12-11T08:11:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:39 crc kubenswrapper[4860]: I1211 08:11:39.223142 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gsx9l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aba5b6b9-f88d-4763-9562-6b4dec606df9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://735ef012c523ddc952d46a431f27b2582a2c301d9e9da0c697a20ff6e2fb5d86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6zw9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gsx9l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:39Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:39 crc kubenswrapper[4860]: I1211 08:11:39.278252 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9aa0aed-62b3-4ea2-82ae-9a653c1101a5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ad4ba1b769c6cb2b34073798f8e3c25eb0177e6d3059484bf365e6981cab710\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://077aa1dd2fe5189ff60a0f2365e47a6a545cca0de7aa6379370908db4ea5838a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d7c919899b74a2b0fcd9221f3f769a9d013267354b200a282669d342d6413d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5483384f1d2ebc1fefa9706c9e67a3dc5a00ff643f803e3f2707c252e27ad94\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:39Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:39 crc kubenswrapper[4860]: I1211 08:11:39.300596 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9872be1e-b995-4a00-904b-875dff96ad56\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f74a32a8fefea704cc73ffd1114083dab68aa44dc33f8d9d432d2d24216546c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbe00b46063d3968d59c393b87178e135f982abb7d0f32fa798b212d4b83df6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eedc9988e1a4a3c8cae730dc4a352103a6d665535b825e157d0f87c331b137c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2137fee4b7abc9765f9517d7b3b6ea796601a31d03151bf167d17136ced8b25a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f683e48f3577a68862eb6302383538376e41f49b20b1e0ea3f1b3b143f24872a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5de794187456b3576b969069a3b839d75c7918418a97fa35dff8bb40bb76e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5de794187456b3576b969069a3b839d75c7918418a97fa35dff8bb40bb76e39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08ce42b974fba00c409c0af971f0de351e50f18b2f06f8e722150da3239bb908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08ce42b974fba00c409c0af971f0de351e50f18b2f06f8e722150da3239bb908\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cfac6be03e3cfa0b0f3450b48b67957f2d5172b9c65e27f14c2a260bda5d423f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cfac6be03e3cfa0b0f3450b48b67957f2d5172b9c65e27f14c2a260bda5d423f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:39Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:39 crc kubenswrapper[4860]: I1211 08:11:39.312055 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-pznwl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5310c1fc-66c6-40aa-b1bf-5a59a2410f9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2d6ff00ce322a28607bc3f6db05c5e4fb14acb6777ea315786ef53963ca504a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hkt48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-pznwl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:39Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:39 crc kubenswrapper[4860]: I1211 08:11:39.316180 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:39 crc kubenswrapper[4860]: I1211 08:11:39.316215 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:39 crc kubenswrapper[4860]: I1211 08:11:39.316226 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:39 crc kubenswrapper[4860]: I1211 08:11:39.316240 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:39 crc kubenswrapper[4860]: I1211 08:11:39.316251 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:39Z","lastTransitionTime":"2025-12-11T08:11:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:39 crc kubenswrapper[4860]: I1211 08:11:39.324700 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d8c2f28628dc2f1ffbdd98a55670e22cf6cdc8cf2c6442e9d1f8e5466b2f12e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://074ea63beac52bd196507ee7cfb1c3d7fc2d37d5ec6faa4d47dde62b922e1c43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:39Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:39 crc kubenswrapper[4860]: I1211 08:11:39.337730 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:39Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:39 crc kubenswrapper[4860]: I1211 08:11:39.347455 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25a6a154da2b74952951ae94efc914ba078ea4bbe5893360b3c84d46be6bf973\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:39Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:39 crc kubenswrapper[4860]: I1211 08:11:39.355689 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6xq7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a75ce3c-5e1e-4b9f-bbde-7a3b0cce5eae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe8d4d675567b41728e476351811a93c47b19868ccbadab8c613a4eb8eca8562\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5c6nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6xq7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:39Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:39 crc kubenswrapper[4860]: I1211 08:11:39.365751 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:39Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:39 crc kubenswrapper[4860]: I1211 08:11:39.376777 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a41f35db087418d1be16953f10377042abd3f4488e5ac1bc29ca5dcfc00a37e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:39Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:39 crc kubenswrapper[4860]: I1211 08:11:39.387705 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ec3758-85f1-4a22-a4ec-c805472607a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e831db1a15976d3967d56479689572c26f2599a766baa0402897a3008428c56f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a498aa90f914e1f0370df01dc5ef174df113a2a16d86b2e3f24b19a1b99b4c0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8caebd2f4dfcbb89c597029a07c335171889761d2d200206987537fb0a228db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34e91f615b76445ff3241a304bb2964fde59de7e8a8dfff6841b2e628a31e86d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2090cab02ad74ddfd6bb57ff50eba35d18f56b0cde1c230fe011c01994b693bc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 08:11:21.171011 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 08:11:21.177674 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-644352688/tls.crt::/tmp/serving-cert-644352688/tls.key\\\\\\\"\\\\nI1211 08:11:26.508606 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 08:11:26.512741 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 08:11:26.512778 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 08:11:26.512839 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 08:11:26.512858 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 08:11:26.521684 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1211 08:11:26.521700 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1211 08:11:26.521722 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:11:26.521732 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:11:26.521738 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 08:11:26.521746 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 08:11:26.521749 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 08:11:26.521753 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1211 08:11:26.524999 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2274fb298ca7d2b37b060b4dbf9b166d39ba77a108b08edeeb4d67196e72cb3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf0aff8a6ab816b6617b1feeca9146d41bb5e8808bf1db96f8d493500c135d85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf0aff8a6ab816b6617b1feeca9146d41bb5e8808bf1db96f8d493500c135d85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:39Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:39 crc kubenswrapper[4860]: I1211 08:11:39.398787 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:39Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:39 crc kubenswrapper[4860]: I1211 08:11:39.417800 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:39 crc kubenswrapper[4860]: I1211 08:11:39.417826 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:39 crc kubenswrapper[4860]: I1211 08:11:39.417835 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:39 crc kubenswrapper[4860]: I1211 08:11:39.417847 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:39 crc kubenswrapper[4860]: I1211 08:11:39.417855 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:39Z","lastTransitionTime":"2025-12-11T08:11:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:39 crc kubenswrapper[4860]: I1211 08:11:39.530093 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:39 crc kubenswrapper[4860]: I1211 08:11:39.530440 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:39 crc kubenswrapper[4860]: I1211 08:11:39.530581 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:39 crc kubenswrapper[4860]: I1211 08:11:39.530744 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:39 crc kubenswrapper[4860]: I1211 08:11:39.530852 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:39Z","lastTransitionTime":"2025-12-11T08:11:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:39 crc kubenswrapper[4860]: I1211 08:11:39.633920 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:39 crc kubenswrapper[4860]: I1211 08:11:39.634585 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:39 crc kubenswrapper[4860]: I1211 08:11:39.634697 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:39 crc kubenswrapper[4860]: I1211 08:11:39.634780 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:39 crc kubenswrapper[4860]: I1211 08:11:39.634852 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:39Z","lastTransitionTime":"2025-12-11T08:11:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:39 crc kubenswrapper[4860]: I1211 08:11:39.744586 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:39 crc kubenswrapper[4860]: I1211 08:11:39.744909 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:39 crc kubenswrapper[4860]: I1211 08:11:39.745031 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:39 crc kubenswrapper[4860]: I1211 08:11:39.745131 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:39 crc kubenswrapper[4860]: I1211 08:11:39.745222 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:39Z","lastTransitionTime":"2025-12-11T08:11:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:39 crc kubenswrapper[4860]: I1211 08:11:39.848570 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:39 crc kubenswrapper[4860]: I1211 08:11:39.848633 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:39 crc kubenswrapper[4860]: I1211 08:11:39.848698 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:39 crc kubenswrapper[4860]: I1211 08:11:39.848729 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:39 crc kubenswrapper[4860]: I1211 08:11:39.848753 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:39Z","lastTransitionTime":"2025-12-11T08:11:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:39 crc kubenswrapper[4860]: I1211 08:11:39.877925 4860 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 11 08:11:39 crc kubenswrapper[4860]: I1211 08:11:39.950971 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:39 crc kubenswrapper[4860]: I1211 08:11:39.951006 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:39 crc kubenswrapper[4860]: I1211 08:11:39.951014 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:39 crc kubenswrapper[4860]: I1211 08:11:39.951027 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:39 crc kubenswrapper[4860]: I1211 08:11:39.951037 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:39Z","lastTransitionTime":"2025-12-11T08:11:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:40 crc kubenswrapper[4860]: I1211 08:11:40.053632 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:40 crc kubenswrapper[4860]: I1211 08:11:40.053675 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:40 crc kubenswrapper[4860]: I1211 08:11:40.053683 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:40 crc kubenswrapper[4860]: I1211 08:11:40.053697 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:40 crc kubenswrapper[4860]: I1211 08:11:40.053705 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:40Z","lastTransitionTime":"2025-12-11T08:11:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:40 crc kubenswrapper[4860]: I1211 08:11:40.156947 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:40 crc kubenswrapper[4860]: I1211 08:11:40.156995 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:40 crc kubenswrapper[4860]: I1211 08:11:40.157014 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:40 crc kubenswrapper[4860]: I1211 08:11:40.157035 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:40 crc kubenswrapper[4860]: I1211 08:11:40.157051 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:40Z","lastTransitionTime":"2025-12-11T08:11:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:40 crc kubenswrapper[4860]: I1211 08:11:40.259817 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:40 crc kubenswrapper[4860]: I1211 08:11:40.259856 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:40 crc kubenswrapper[4860]: I1211 08:11:40.259871 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:40 crc kubenswrapper[4860]: I1211 08:11:40.259891 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:40 crc kubenswrapper[4860]: I1211 08:11:40.259906 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:40Z","lastTransitionTime":"2025-12-11T08:11:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:40 crc kubenswrapper[4860]: I1211 08:11:40.362743 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:40 crc kubenswrapper[4860]: I1211 08:11:40.362781 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:40 crc kubenswrapper[4860]: I1211 08:11:40.362791 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:40 crc kubenswrapper[4860]: I1211 08:11:40.362805 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:40 crc kubenswrapper[4860]: I1211 08:11:40.362814 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:40Z","lastTransitionTime":"2025-12-11T08:11:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:40 crc kubenswrapper[4860]: I1211 08:11:40.416973 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 08:11:40 crc kubenswrapper[4860]: I1211 08:11:40.445838 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9872be1e-b995-4a00-904b-875dff96ad56\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f74a32a8fefea704cc73ffd1114083dab68aa44dc33f8d9d432d2d24216546c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbe00b46063d3968d59c393b87178e135f982abb7d0f32fa798b212d4b83df6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eedc9988e1a4a3c8cae730dc4a352103a6d665535b825e157d0f87c331b137c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2137fee4b7abc9765f9517d7b3b6ea796601a31d03151bf167d17136ced8b25a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f683e48f3577a68862eb6302383538376e41f49b20b1e0ea3f1b3b143f24872a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5de794187456b3576b969069a3b839d75c7918418a97fa35dff8bb40bb76e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5de794187456b3576b969069a3b839d75c7918418a97fa35dff8bb40bb76e39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08ce42b974fba00c409c0af971f0de351e50f18b2f06f8e722150da3239bb908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08ce42b974fba00c409c0af971f0de351e50f18b2f06f8e722150da3239bb908\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cfac6be03e3cfa0b0f3450b48b67957f2d5172b9c65e27f14c2a260bda5d423f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cfac6be03e3cfa0b0f3450b48b67957f2d5172b9c65e27f14c2a260bda5d423f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:40Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:40 crc kubenswrapper[4860]: I1211 08:11:40.465460 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:40 crc kubenswrapper[4860]: I1211 08:11:40.465500 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:40 crc kubenswrapper[4860]: I1211 08:11:40.465511 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:40 crc kubenswrapper[4860]: I1211 08:11:40.465527 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:40 crc kubenswrapper[4860]: I1211 08:11:40.465540 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:40Z","lastTransitionTime":"2025-12-11T08:11:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:40 crc kubenswrapper[4860]: I1211 08:11:40.468755 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-pznwl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5310c1fc-66c6-40aa-b1bf-5a59a2410f9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2d6ff00ce322a28607bc3f6db05c5e4fb14acb6777ea315786ef53963ca504a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hkt48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-pznwl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:40Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:40 crc kubenswrapper[4860]: I1211 08:11:40.482523 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6xq7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a75ce3c-5e1e-4b9f-bbde-7a3b0cce5eae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe8d4d675567b41728e476351811a93c47b19868ccbadab8c613a4eb8eca8562\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5c6nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6xq7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:40Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:40 crc kubenswrapper[4860]: I1211 08:11:40.495715 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:40Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:40 crc kubenswrapper[4860]: I1211 08:11:40.509050 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d8c2f28628dc2f1ffbdd98a55670e22cf6cdc8cf2c6442e9d1f8e5466b2f12e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://074ea63beac52bd196507ee7cfb1c3d7fc2d37d5ec6faa4d47dde62b922e1c43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:40Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:40 crc kubenswrapper[4860]: I1211 08:11:40.520897 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:40Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:40 crc kubenswrapper[4860]: I1211 08:11:40.532695 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25a6a154da2b74952951ae94efc914ba078ea4bbe5893360b3c84d46be6bf973\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:40Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:40 crc kubenswrapper[4860]: I1211 08:11:40.547456 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ec3758-85f1-4a22-a4ec-c805472607a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e831db1a15976d3967d56479689572c26f2599a766baa0402897a3008428c56f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a498aa90f914e1f0370df01dc5ef174df113a2a16d86b2e3f24b19a1b99b4c0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8caebd2f4dfcbb89c597029a07c335171889761d2d200206987537fb0a228db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34e91f615b76445ff3241a304bb2964fde59de7e8a8dfff6841b2e628a31e86d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2090cab02ad74ddfd6bb57ff50eba35d18f56b0cde1c230fe011c01994b693bc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 08:11:21.171011 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 08:11:21.177674 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-644352688/tls.crt::/tmp/serving-cert-644352688/tls.key\\\\\\\"\\\\nI1211 08:11:26.508606 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 08:11:26.512741 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 08:11:26.512778 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 08:11:26.512839 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 08:11:26.512858 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 08:11:26.521684 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1211 08:11:26.521700 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1211 08:11:26.521722 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:11:26.521732 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:11:26.521738 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 08:11:26.521746 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 08:11:26.521749 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 08:11:26.521753 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1211 08:11:26.524999 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2274fb298ca7d2b37b060b4dbf9b166d39ba77a108b08edeeb4d67196e72cb3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf0aff8a6ab816b6617b1feeca9146d41bb5e8808bf1db96f8d493500c135d85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf0aff8a6ab816b6617b1feeca9146d41bb5e8808bf1db96f8d493500c135d85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:40Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:40 crc kubenswrapper[4860]: I1211 08:11:40.561208 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:40Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:40 crc kubenswrapper[4860]: I1211 08:11:40.567510 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:40 crc kubenswrapper[4860]: I1211 08:11:40.567795 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:40 crc kubenswrapper[4860]: I1211 08:11:40.567911 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:40 crc kubenswrapper[4860]: I1211 08:11:40.568010 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:40 crc kubenswrapper[4860]: I1211 08:11:40.568090 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:40Z","lastTransitionTime":"2025-12-11T08:11:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:40 crc kubenswrapper[4860]: I1211 08:11:40.576752 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a41f35db087418d1be16953f10377042abd3f4488e5ac1bc29ca5dcfc00a37e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:40Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:40 crc kubenswrapper[4860]: I1211 08:11:40.577842 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:11:40 crc kubenswrapper[4860]: I1211 08:11:40.577922 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:11:40 crc kubenswrapper[4860]: I1211 08:11:40.577881 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:11:40 crc kubenswrapper[4860]: E1211 08:11:40.578069 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:11:40 crc kubenswrapper[4860]: E1211 08:11:40.578174 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:11:40 crc kubenswrapper[4860]: E1211 08:11:40.578246 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:11:40 crc kubenswrapper[4860]: I1211 08:11:40.587985 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gsx9l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aba5b6b9-f88d-4763-9562-6b4dec606df9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://735ef012c523ddc952d46a431f27b2582a2c301d9e9da0c697a20ff6e2fb5d86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6zw9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gsx9l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:40Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:40 crc kubenswrapper[4860]: I1211 08:11:40.600969 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9aa0aed-62b3-4ea2-82ae-9a653c1101a5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ad4ba1b769c6cb2b34073798f8e3c25eb0177e6d3059484bf365e6981cab710\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://077aa1dd2fe5189ff60a0f2365e47a6a545cca0de7aa6379370908db4ea5838a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d7c919899b74a2b0fcd9221f3f769a9d013267354b200a282669d342d6413d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5483384f1d2ebc1fefa9706c9e67a3dc5a00ff643f803e3f2707c252e27ad94\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:40Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:40 crc kubenswrapper[4860]: I1211 08:11:40.612960 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31c30642-6e60-41b4-a477-0d802424e0aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79969402f5c33852a26fc839f37a45febf4f2377d77823284fbb01769e5585fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2d4d5de7106175087d91acb1f0c7552014f0c08bfcd65cbbf184d3f8d7826ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-99qgp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:40Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:40 crc kubenswrapper[4860]: I1211 08:11:40.629432 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-26nc6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1faa733-f408-41fd-b7aa-0cd86283976a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fce5b642b843ed30823c94e0b0c3d8bb1e5d2d93d691f7b8ef901ece03b0ac6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ac45db4f858a1c18a6b87472fc1cea2300f5845f1ad62c22260374995548a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ac45db4f858a1c18a6b87472fc1cea2300f5845f1ad62c22260374995548a0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b35296a5eb17dd2bb5e72c86c1cafdb30b7712cb330264f935d3b4fff05cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c6b35296a5eb17dd2bb5e72c86c1cafdb30b7712cb330264f935d3b4fff05cd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b9e82d6d3a247309ef5bdbf5b3a86ec083add2a138b0b237801e77b09ee96a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b9e82d6d3a247309ef5bdbf5b3a86ec083add2a138b0b237801e77b09ee96a1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://908ee695e78145ef4375c4097d9f1e772502922d02a2fcb4566172a6ad9caaf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://908ee695e78145ef4375c4097d9f1e772502922d02a2fcb4566172a6ad9caaf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afb0a097f3556a8f9ed7cf8f762cb99c082fe0a36b8d8d3038bf188b08fe9a84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afb0a097f3556a8f9ed7cf8f762cb99c082fe0a36b8d8d3038bf188b08fe9a84\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce18f4f06db091e658684fa0bbc858e19b9f03010acf13b3ffccb1a6e3f88021\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce18f4f06db091e658684fa0bbc858e19b9f03010acf13b3ffccb1a6e3f88021\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-26nc6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:40Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:40 crc kubenswrapper[4860]: I1211 08:11:40.649161 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0589a204-c98c-417d-8256-bfe3bf747660\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3333b4346f59b04017e32d61f457ba7ff457166731128f339d7987ae6b56b06d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52bb8ade314afb9494f4e026d0af2aea7b1970d64dc66102f55e257ae4678154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dfcc4370edaae36a5a32b0a9c29290c50f8e4f427e4e147dc3c779beaa8b780\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d856ae95e9077524c518d82d10fb6b93b7fbcaaf2b720ffafe6f7fae58bd7caa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1426084ac580895c42354630910dd8c8e3a48ddd25f22abce1d28dc3de332a63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9984d5d242bb258405efe91fc9caf9e7ea592834445df2e78add9553a921c76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://526c2b9214d19b9ebe7a7078fd0459642a53e732d034602b2b8d2b8a33f4d18b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb80d9928e8349fc5963227d135355ad7ba820d4fd89877af601d005ca03334b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17cbcb49c06f3a9059cd5a0f05eedb8c785da30a3e502e31bae2f57b3486cb14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17cbcb49c06f3a9059cd5a0f05eedb8c785da30a3e502e31bae2f57b3486cb14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xvnqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:40Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:40 crc kubenswrapper[4860]: I1211 08:11:40.671181 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:40 crc kubenswrapper[4860]: I1211 08:11:40.671224 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:40 crc kubenswrapper[4860]: I1211 08:11:40.671236 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:40 crc kubenswrapper[4860]: I1211 08:11:40.671253 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:40 crc kubenswrapper[4860]: I1211 08:11:40.671263 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:40Z","lastTransitionTime":"2025-12-11T08:11:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:40 crc kubenswrapper[4860]: I1211 08:11:40.773956 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:40 crc kubenswrapper[4860]: I1211 08:11:40.774014 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:40 crc kubenswrapper[4860]: I1211 08:11:40.774031 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:40 crc kubenswrapper[4860]: I1211 08:11:40.774050 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:40 crc kubenswrapper[4860]: I1211 08:11:40.774064 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:40Z","lastTransitionTime":"2025-12-11T08:11:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:40 crc kubenswrapper[4860]: I1211 08:11:40.814016 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:40 crc kubenswrapper[4860]: I1211 08:11:40.814058 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:40 crc kubenswrapper[4860]: I1211 08:11:40.814071 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:40 crc kubenswrapper[4860]: I1211 08:11:40.814090 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:40 crc kubenswrapper[4860]: I1211 08:11:40.814102 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:40Z","lastTransitionTime":"2025-12-11T08:11:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:40 crc kubenswrapper[4860]: E1211 08:11:40.832875 4860 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:11:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:11:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:11:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:11:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"52631a1a-bc27-4dd5-8309-5dc39b5dcc41\\\",\\\"systemUUID\\\":\\\"299d3cd2-e1ad-4f5c-8af4-f12c0a5d76e7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:40Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:40 crc kubenswrapper[4860]: I1211 08:11:40.837343 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:40 crc kubenswrapper[4860]: I1211 08:11:40.837405 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:40 crc kubenswrapper[4860]: I1211 08:11:40.837417 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:40 crc kubenswrapper[4860]: I1211 08:11:40.837433 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:40 crc kubenswrapper[4860]: I1211 08:11:40.837467 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:40Z","lastTransitionTime":"2025-12-11T08:11:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:40 crc kubenswrapper[4860]: E1211 08:11:40.854487 4860 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:11:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:11:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:11:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:11:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"52631a1a-bc27-4dd5-8309-5dc39b5dcc41\\\",\\\"systemUUID\\\":\\\"299d3cd2-e1ad-4f5c-8af4-f12c0a5d76e7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:40Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:40 crc kubenswrapper[4860]: I1211 08:11:40.858859 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:40 crc kubenswrapper[4860]: I1211 08:11:40.859018 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:40 crc kubenswrapper[4860]: I1211 08:11:40.859109 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:40 crc kubenswrapper[4860]: I1211 08:11:40.859199 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:40 crc kubenswrapper[4860]: I1211 08:11:40.859286 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:40Z","lastTransitionTime":"2025-12-11T08:11:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:40 crc kubenswrapper[4860]: E1211 08:11:40.875450 4860 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:11:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:11:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:11:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:11:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"52631a1a-bc27-4dd5-8309-5dc39b5dcc41\\\",\\\"systemUUID\\\":\\\"299d3cd2-e1ad-4f5c-8af4-f12c0a5d76e7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:40Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:40 crc kubenswrapper[4860]: I1211 08:11:40.879280 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:40 crc kubenswrapper[4860]: I1211 08:11:40.879313 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:40 crc kubenswrapper[4860]: I1211 08:11:40.879321 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:40 crc kubenswrapper[4860]: I1211 08:11:40.879339 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:40 crc kubenswrapper[4860]: I1211 08:11:40.879351 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:40Z","lastTransitionTime":"2025-12-11T08:11:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:40 crc kubenswrapper[4860]: I1211 08:11:40.880792 4860 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 11 08:11:40 crc kubenswrapper[4860]: E1211 08:11:40.893756 4860 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:11:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:11:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:11:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:11:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"52631a1a-bc27-4dd5-8309-5dc39b5dcc41\\\",\\\"systemUUID\\\":\\\"299d3cd2-e1ad-4f5c-8af4-f12c0a5d76e7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:40Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:40 crc kubenswrapper[4860]: I1211 08:11:40.897989 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:40 crc kubenswrapper[4860]: I1211 08:11:40.898038 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:40 crc kubenswrapper[4860]: I1211 08:11:40.898056 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:40 crc kubenswrapper[4860]: I1211 08:11:40.898079 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:40 crc kubenswrapper[4860]: I1211 08:11:40.898096 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:40Z","lastTransitionTime":"2025-12-11T08:11:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:40 crc kubenswrapper[4860]: E1211 08:11:40.915700 4860 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:11:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:11:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:11:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:11:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"52631a1a-bc27-4dd5-8309-5dc39b5dcc41\\\",\\\"systemUUID\\\":\\\"299d3cd2-e1ad-4f5c-8af4-f12c0a5d76e7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:40Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:40 crc kubenswrapper[4860]: E1211 08:11:40.915834 4860 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 11 08:11:40 crc kubenswrapper[4860]: I1211 08:11:40.918077 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:40 crc kubenswrapper[4860]: I1211 08:11:40.918323 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:40 crc kubenswrapper[4860]: I1211 08:11:40.918538 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:40 crc kubenswrapper[4860]: I1211 08:11:40.918804 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:40 crc kubenswrapper[4860]: I1211 08:11:40.919009 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:40Z","lastTransitionTime":"2025-12-11T08:11:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:41 crc kubenswrapper[4860]: I1211 08:11:41.022231 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:41 crc kubenswrapper[4860]: I1211 08:11:41.022269 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:41 crc kubenswrapper[4860]: I1211 08:11:41.022306 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:41 crc kubenswrapper[4860]: I1211 08:11:41.022326 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:41 crc kubenswrapper[4860]: I1211 08:11:41.022337 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:41Z","lastTransitionTime":"2025-12-11T08:11:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:41 crc kubenswrapper[4860]: I1211 08:11:41.124584 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:41 crc kubenswrapper[4860]: I1211 08:11:41.124630 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:41 crc kubenswrapper[4860]: I1211 08:11:41.124701 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:41 crc kubenswrapper[4860]: I1211 08:11:41.124724 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:41 crc kubenswrapper[4860]: I1211 08:11:41.124738 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:41Z","lastTransitionTime":"2025-12-11T08:11:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:41 crc kubenswrapper[4860]: I1211 08:11:41.227343 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:41 crc kubenswrapper[4860]: I1211 08:11:41.227429 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:41 crc kubenswrapper[4860]: I1211 08:11:41.227449 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:41 crc kubenswrapper[4860]: I1211 08:11:41.227518 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:41 crc kubenswrapper[4860]: I1211 08:11:41.227540 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:41Z","lastTransitionTime":"2025-12-11T08:11:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:41 crc kubenswrapper[4860]: I1211 08:11:41.330804 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:41 crc kubenswrapper[4860]: I1211 08:11:41.330886 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:41 crc kubenswrapper[4860]: I1211 08:11:41.330909 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:41 crc kubenswrapper[4860]: I1211 08:11:41.330942 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:41 crc kubenswrapper[4860]: I1211 08:11:41.330967 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:41Z","lastTransitionTime":"2025-12-11T08:11:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:41 crc kubenswrapper[4860]: I1211 08:11:41.348149 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-854v6"] Dec 11 08:11:41 crc kubenswrapper[4860]: I1211 08:11:41.348608 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-854v6" Dec 11 08:11:41 crc kubenswrapper[4860]: I1211 08:11:41.350764 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 11 08:11:41 crc kubenswrapper[4860]: I1211 08:11:41.351565 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 11 08:11:41 crc kubenswrapper[4860]: I1211 08:11:41.364363 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:41Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:41 crc kubenswrapper[4860]: I1211 08:11:41.380137 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a41f35db087418d1be16953f10377042abd3f4488e5ac1bc29ca5dcfc00a37e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:41Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:41 crc kubenswrapper[4860]: I1211 08:11:41.396705 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ec3758-85f1-4a22-a4ec-c805472607a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e831db1a15976d3967d56479689572c26f2599a766baa0402897a3008428c56f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a498aa90f914e1f0370df01dc5ef174df113a2a16d86b2e3f24b19a1b99b4c0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8caebd2f4dfcbb89c597029a07c335171889761d2d200206987537fb0a228db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34e91f615b76445ff3241a304bb2964fde59de7e8a8dfff6841b2e628a31e86d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2090cab02ad74ddfd6bb57ff50eba35d18f56b0cde1c230fe011c01994b693bc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 08:11:21.171011 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 08:11:21.177674 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-644352688/tls.crt::/tmp/serving-cert-644352688/tls.key\\\\\\\"\\\\nI1211 08:11:26.508606 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 08:11:26.512741 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 08:11:26.512778 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 08:11:26.512839 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 08:11:26.512858 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 08:11:26.521684 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1211 08:11:26.521700 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1211 08:11:26.521722 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:11:26.521732 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:11:26.521738 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 08:11:26.521746 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 08:11:26.521749 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 08:11:26.521753 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1211 08:11:26.524999 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2274fb298ca7d2b37b060b4dbf9b166d39ba77a108b08edeeb4d67196e72cb3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf0aff8a6ab816b6617b1feeca9146d41bb5e8808bf1db96f8d493500c135d85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf0aff8a6ab816b6617b1feeca9146d41bb5e8808bf1db96f8d493500c135d85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:41Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:41 crc kubenswrapper[4860]: I1211 08:11:41.408827 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9aa0aed-62b3-4ea2-82ae-9a653c1101a5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ad4ba1b769c6cb2b34073798f8e3c25eb0177e6d3059484bf365e6981cab710\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://077aa1dd2fe5189ff60a0f2365e47a6a545cca0de7aa6379370908db4ea5838a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d7c919899b74a2b0fcd9221f3f769a9d013267354b200a282669d342d6413d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5483384f1d2ebc1fefa9706c9e67a3dc5a00ff643f803e3f2707c252e27ad94\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:41Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:41 crc kubenswrapper[4860]: I1211 08:11:41.424510 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31c30642-6e60-41b4-a477-0d802424e0aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79969402f5c33852a26fc839f37a45febf4f2377d77823284fbb01769e5585fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2d4d5de7106175087d91acb1f0c7552014f0c08bfcd65cbbf184d3f8d7826ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-99qgp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:41Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:41 crc kubenswrapper[4860]: I1211 08:11:41.440035 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:41 crc kubenswrapper[4860]: I1211 08:11:41.440098 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:41 crc kubenswrapper[4860]: I1211 08:11:41.440109 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:41 crc kubenswrapper[4860]: I1211 08:11:41.440126 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:41 crc kubenswrapper[4860]: I1211 08:11:41.440138 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:41Z","lastTransitionTime":"2025-12-11T08:11:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:41 crc kubenswrapper[4860]: I1211 08:11:41.442251 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-26nc6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1faa733-f408-41fd-b7aa-0cd86283976a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fce5b642b843ed30823c94e0b0c3d8bb1e5d2d93d691f7b8ef901ece03b0ac6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ac45db4f858a1c18a6b87472fc1cea2300f5845f1ad62c22260374995548a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ac45db4f858a1c18a6b87472fc1cea2300f5845f1ad62c22260374995548a0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b35296a5eb17dd2bb5e72c86c1cafdb30b7712cb330264f935d3b4fff05cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c6b35296a5eb17dd2bb5e72c86c1cafdb30b7712cb330264f935d3b4fff05cd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b9e82d6d3a247309ef5bdbf5b3a86ec083add2a138b0b237801e77b09ee96a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b9e82d6d3a247309ef5bdbf5b3a86ec083add2a138b0b237801e77b09ee96a1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://908ee695e78145ef4375c4097d9f1e772502922d02a2fcb4566172a6ad9caaf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://908ee695e78145ef4375c4097d9f1e772502922d02a2fcb4566172a6ad9caaf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afb0a097f3556a8f9ed7cf8f762cb99c082fe0a36b8d8d3038bf188b08fe9a84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afb0a097f3556a8f9ed7cf8f762cb99c082fe0a36b8d8d3038bf188b08fe9a84\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce18f4f06db091e658684fa0bbc858e19b9f03010acf13b3ffccb1a6e3f88021\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce18f4f06db091e658684fa0bbc858e19b9f03010acf13b3ffccb1a6e3f88021\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-26nc6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:41Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:41 crc kubenswrapper[4860]: I1211 08:11:41.449752 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/e6529ba7-2341-43ac-a187-95ab530fe775-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-854v6\" (UID: \"e6529ba7-2341-43ac-a187-95ab530fe775\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-854v6" Dec 11 08:11:41 crc kubenswrapper[4860]: I1211 08:11:41.449829 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/e6529ba7-2341-43ac-a187-95ab530fe775-env-overrides\") pod \"ovnkube-control-plane-749d76644c-854v6\" (UID: \"e6529ba7-2341-43ac-a187-95ab530fe775\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-854v6" Dec 11 08:11:41 crc kubenswrapper[4860]: I1211 08:11:41.449856 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kh22b\" (UniqueName: \"kubernetes.io/projected/e6529ba7-2341-43ac-a187-95ab530fe775-kube-api-access-kh22b\") pod \"ovnkube-control-plane-749d76644c-854v6\" (UID: \"e6529ba7-2341-43ac-a187-95ab530fe775\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-854v6" Dec 11 08:11:41 crc kubenswrapper[4860]: I1211 08:11:41.449899 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/e6529ba7-2341-43ac-a187-95ab530fe775-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-854v6\" (UID: \"e6529ba7-2341-43ac-a187-95ab530fe775\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-854v6" Dec 11 08:11:41 crc kubenswrapper[4860]: I1211 08:11:41.465226 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0589a204-c98c-417d-8256-bfe3bf747660\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3333b4346f59b04017e32d61f457ba7ff457166731128f339d7987ae6b56b06d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52bb8ade314afb9494f4e026d0af2aea7b1970d64dc66102f55e257ae4678154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dfcc4370edaae36a5a32b0a9c29290c50f8e4f427e4e147dc3c779beaa8b780\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d856ae95e9077524c518d82d10fb6b93b7fbcaaf2b720ffafe6f7fae58bd7caa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1426084ac580895c42354630910dd8c8e3a48ddd25f22abce1d28dc3de332a63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9984d5d242bb258405efe91fc9caf9e7ea592834445df2e78add9553a921c76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://526c2b9214d19b9ebe7a7078fd0459642a53e732d034602b2b8d2b8a33f4d18b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb80d9928e8349fc5963227d135355ad7ba820d4fd89877af601d005ca03334b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17cbcb49c06f3a9059cd5a0f05eedb8c785da30a3e502e31bae2f57b3486cb14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17cbcb49c06f3a9059cd5a0f05eedb8c785da30a3e502e31bae2f57b3486cb14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xvnqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:41Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:41 crc kubenswrapper[4860]: I1211 08:11:41.477751 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gsx9l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aba5b6b9-f88d-4763-9562-6b4dec606df9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://735ef012c523ddc952d46a431f27b2582a2c301d9e9da0c697a20ff6e2fb5d86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6zw9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gsx9l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:41Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:41 crc kubenswrapper[4860]: I1211 08:11:41.492335 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-pznwl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5310c1fc-66c6-40aa-b1bf-5a59a2410f9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2d6ff00ce322a28607bc3f6db05c5e4fb14acb6777ea315786ef53963ca504a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hkt48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-pznwl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:41Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:41 crc kubenswrapper[4860]: I1211 08:11:41.505013 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-854v6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e6529ba7-2341-43ac-a187-95ab530fe775\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kh22b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kh22b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-854v6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:41Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:41 crc kubenswrapper[4860]: I1211 08:11:41.523235 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9872be1e-b995-4a00-904b-875dff96ad56\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f74a32a8fefea704cc73ffd1114083dab68aa44dc33f8d9d432d2d24216546c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbe00b46063d3968d59c393b87178e135f982abb7d0f32fa798b212d4b83df6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eedc9988e1a4a3c8cae730dc4a352103a6d665535b825e157d0f87c331b137c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2137fee4b7abc9765f9517d7b3b6ea796601a31d03151bf167d17136ced8b25a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f683e48f3577a68862eb6302383538376e41f49b20b1e0ea3f1b3b143f24872a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5de794187456b3576b969069a3b839d75c7918418a97fa35dff8bb40bb76e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5de794187456b3576b969069a3b839d75c7918418a97fa35dff8bb40bb76e39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08ce42b974fba00c409c0af971f0de351e50f18b2f06f8e722150da3239bb908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08ce42b974fba00c409c0af971f0de351e50f18b2f06f8e722150da3239bb908\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cfac6be03e3cfa0b0f3450b48b67957f2d5172b9c65e27f14c2a260bda5d423f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cfac6be03e3cfa0b0f3450b48b67957f2d5172b9c65e27f14c2a260bda5d423f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:41Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:41 crc kubenswrapper[4860]: I1211 08:11:41.535755 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:41Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:41 crc kubenswrapper[4860]: I1211 08:11:41.542415 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:41 crc kubenswrapper[4860]: I1211 08:11:41.542456 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:41 crc kubenswrapper[4860]: I1211 08:11:41.542466 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:41 crc kubenswrapper[4860]: I1211 08:11:41.542482 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:41 crc kubenswrapper[4860]: I1211 08:11:41.542494 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:41Z","lastTransitionTime":"2025-12-11T08:11:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:41 crc kubenswrapper[4860]: I1211 08:11:41.550686 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/e6529ba7-2341-43ac-a187-95ab530fe775-env-overrides\") pod \"ovnkube-control-plane-749d76644c-854v6\" (UID: \"e6529ba7-2341-43ac-a187-95ab530fe775\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-854v6" Dec 11 08:11:41 crc kubenswrapper[4860]: I1211 08:11:41.550765 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kh22b\" (UniqueName: \"kubernetes.io/projected/e6529ba7-2341-43ac-a187-95ab530fe775-kube-api-access-kh22b\") pod \"ovnkube-control-plane-749d76644c-854v6\" (UID: \"e6529ba7-2341-43ac-a187-95ab530fe775\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-854v6" Dec 11 08:11:41 crc kubenswrapper[4860]: I1211 08:11:41.550817 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/e6529ba7-2341-43ac-a187-95ab530fe775-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-854v6\" (UID: \"e6529ba7-2341-43ac-a187-95ab530fe775\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-854v6" Dec 11 08:11:41 crc kubenswrapper[4860]: I1211 08:11:41.550858 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/e6529ba7-2341-43ac-a187-95ab530fe775-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-854v6\" (UID: \"e6529ba7-2341-43ac-a187-95ab530fe775\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-854v6" Dec 11 08:11:41 crc kubenswrapper[4860]: I1211 08:11:41.551953 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/e6529ba7-2341-43ac-a187-95ab530fe775-env-overrides\") pod \"ovnkube-control-plane-749d76644c-854v6\" (UID: \"e6529ba7-2341-43ac-a187-95ab530fe775\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-854v6" Dec 11 08:11:41 crc kubenswrapper[4860]: I1211 08:11:41.552051 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/e6529ba7-2341-43ac-a187-95ab530fe775-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-854v6\" (UID: \"e6529ba7-2341-43ac-a187-95ab530fe775\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-854v6" Dec 11 08:11:41 crc kubenswrapper[4860]: I1211 08:11:41.553808 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d8c2f28628dc2f1ffbdd98a55670e22cf6cdc8cf2c6442e9d1f8e5466b2f12e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://074ea63beac52bd196507ee7cfb1c3d7fc2d37d5ec6faa4d47dde62b922e1c43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:41Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:41 crc kubenswrapper[4860]: I1211 08:11:41.568529 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kh22b\" (UniqueName: \"kubernetes.io/projected/e6529ba7-2341-43ac-a187-95ab530fe775-kube-api-access-kh22b\") pod \"ovnkube-control-plane-749d76644c-854v6\" (UID: \"e6529ba7-2341-43ac-a187-95ab530fe775\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-854v6" Dec 11 08:11:41 crc kubenswrapper[4860]: I1211 08:11:41.570454 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/e6529ba7-2341-43ac-a187-95ab530fe775-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-854v6\" (UID: \"e6529ba7-2341-43ac-a187-95ab530fe775\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-854v6" Dec 11 08:11:41 crc kubenswrapper[4860]: I1211 08:11:41.595741 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:41Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:41 crc kubenswrapper[4860]: I1211 08:11:41.609964 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25a6a154da2b74952951ae94efc914ba078ea4bbe5893360b3c84d46be6bf973\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:41Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:41 crc kubenswrapper[4860]: I1211 08:11:41.620145 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6xq7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a75ce3c-5e1e-4b9f-bbde-7a3b0cce5eae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe8d4d675567b41728e476351811a93c47b19868ccbadab8c613a4eb8eca8562\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5c6nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6xq7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:41Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:41 crc kubenswrapper[4860]: I1211 08:11:41.648679 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:41 crc kubenswrapper[4860]: I1211 08:11:41.648720 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:41 crc kubenswrapper[4860]: I1211 08:11:41.648732 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:41 crc kubenswrapper[4860]: I1211 08:11:41.648749 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:41 crc kubenswrapper[4860]: I1211 08:11:41.648760 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:41Z","lastTransitionTime":"2025-12-11T08:11:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:41 crc kubenswrapper[4860]: I1211 08:11:41.673673 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-854v6" Dec 11 08:11:41 crc kubenswrapper[4860]: W1211 08:11:41.697916 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode6529ba7_2341_43ac_a187_95ab530fe775.slice/crio-50bba361320c09e7f916cf03ebe1cc9dec83bd1b7fce1f7babf3c99db75f835e WatchSource:0}: Error finding container 50bba361320c09e7f916cf03ebe1cc9dec83bd1b7fce1f7babf3c99db75f835e: Status 404 returned error can't find the container with id 50bba361320c09e7f916cf03ebe1cc9dec83bd1b7fce1f7babf3c99db75f835e Dec 11 08:11:41 crc kubenswrapper[4860]: I1211 08:11:41.750975 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:41 crc kubenswrapper[4860]: I1211 08:11:41.751006 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:41 crc kubenswrapper[4860]: I1211 08:11:41.751016 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:41 crc kubenswrapper[4860]: I1211 08:11:41.751030 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:41 crc kubenswrapper[4860]: I1211 08:11:41.751039 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:41Z","lastTransitionTime":"2025-12-11T08:11:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:41 crc kubenswrapper[4860]: I1211 08:11:41.866053 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:41 crc kubenswrapper[4860]: I1211 08:11:41.866132 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:41 crc kubenswrapper[4860]: I1211 08:11:41.866152 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:41 crc kubenswrapper[4860]: I1211 08:11:41.866177 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:41 crc kubenswrapper[4860]: I1211 08:11:41.866195 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:41Z","lastTransitionTime":"2025-12-11T08:11:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:41 crc kubenswrapper[4860]: I1211 08:11:41.885229 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-854v6" event={"ID":"e6529ba7-2341-43ac-a187-95ab530fe775","Type":"ContainerStarted","Data":"50bba361320c09e7f916cf03ebe1cc9dec83bd1b7fce1f7babf3c99db75f835e"} Dec 11 08:11:41 crc kubenswrapper[4860]: I1211 08:11:41.968273 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:41 crc kubenswrapper[4860]: I1211 08:11:41.968314 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:41 crc kubenswrapper[4860]: I1211 08:11:41.968324 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:41 crc kubenswrapper[4860]: I1211 08:11:41.968339 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:41 crc kubenswrapper[4860]: I1211 08:11:41.968349 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:41Z","lastTransitionTime":"2025-12-11T08:11:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:42 crc kubenswrapper[4860]: I1211 08:11:42.071105 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:42 crc kubenswrapper[4860]: I1211 08:11:42.071499 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:42 crc kubenswrapper[4860]: I1211 08:11:42.071521 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:42 crc kubenswrapper[4860]: I1211 08:11:42.071551 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:42 crc kubenswrapper[4860]: I1211 08:11:42.071570 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:42Z","lastTransitionTime":"2025-12-11T08:11:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:42 crc kubenswrapper[4860]: I1211 08:11:42.175200 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:42 crc kubenswrapper[4860]: I1211 08:11:42.175260 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:42 crc kubenswrapper[4860]: I1211 08:11:42.175277 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:42 crc kubenswrapper[4860]: I1211 08:11:42.175306 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:42 crc kubenswrapper[4860]: I1211 08:11:42.175323 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:42Z","lastTransitionTime":"2025-12-11T08:11:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:42 crc kubenswrapper[4860]: I1211 08:11:42.278256 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:42 crc kubenswrapper[4860]: I1211 08:11:42.278309 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:42 crc kubenswrapper[4860]: I1211 08:11:42.278327 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:42 crc kubenswrapper[4860]: I1211 08:11:42.278360 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:42 crc kubenswrapper[4860]: I1211 08:11:42.278385 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:42Z","lastTransitionTime":"2025-12-11T08:11:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:42 crc kubenswrapper[4860]: I1211 08:11:42.361540 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:11:42 crc kubenswrapper[4860]: E1211 08:11:42.361738 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:11:58.361708993 +0000 UTC m=+51.090228058 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:11:42 crc kubenswrapper[4860]: I1211 08:11:42.361868 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:11:42 crc kubenswrapper[4860]: E1211 08:11:42.362019 4860 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 11 08:11:42 crc kubenswrapper[4860]: E1211 08:11:42.362072 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-11 08:11:58.362063324 +0000 UTC m=+51.090582389 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 11 08:11:42 crc kubenswrapper[4860]: I1211 08:11:42.381541 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:42 crc kubenswrapper[4860]: I1211 08:11:42.381580 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:42 crc kubenswrapper[4860]: I1211 08:11:42.381590 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:42 crc kubenswrapper[4860]: I1211 08:11:42.381605 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:42 crc kubenswrapper[4860]: I1211 08:11:42.381615 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:42Z","lastTransitionTime":"2025-12-11T08:11:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:42 crc kubenswrapper[4860]: I1211 08:11:42.463989 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:11:42 crc kubenswrapper[4860]: I1211 08:11:42.464059 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:11:42 crc kubenswrapper[4860]: I1211 08:11:42.464106 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:11:42 crc kubenswrapper[4860]: E1211 08:11:42.464274 4860 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 11 08:11:42 crc kubenswrapper[4860]: E1211 08:11:42.464374 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-11 08:11:58.464351132 +0000 UTC m=+51.192870197 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 11 08:11:42 crc kubenswrapper[4860]: E1211 08:11:42.464384 4860 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 11 08:11:42 crc kubenswrapper[4860]: E1211 08:11:42.464418 4860 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 11 08:11:42 crc kubenswrapper[4860]: E1211 08:11:42.464436 4860 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 08:11:42 crc kubenswrapper[4860]: E1211 08:11:42.464532 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-11 08:11:58.464502027 +0000 UTC m=+51.193021122 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 08:11:42 crc kubenswrapper[4860]: E1211 08:11:42.465445 4860 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 11 08:11:42 crc kubenswrapper[4860]: E1211 08:11:42.465503 4860 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 11 08:11:42 crc kubenswrapper[4860]: E1211 08:11:42.465532 4860 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 08:11:42 crc kubenswrapper[4860]: E1211 08:11:42.465674 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-11 08:11:58.465612073 +0000 UTC m=+51.194131168 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 08:11:42 crc kubenswrapper[4860]: I1211 08:11:42.466393 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-gjhlj"] Dec 11 08:11:42 crc kubenswrapper[4860]: I1211 08:11:42.467019 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gjhlj" Dec 11 08:11:42 crc kubenswrapper[4860]: E1211 08:11:42.467119 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gjhlj" podUID="5cfd640a-f3d2-4b28-87e1-5f49f6341e21" Dec 11 08:11:42 crc kubenswrapper[4860]: I1211 08:11:42.485583 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:42 crc kubenswrapper[4860]: I1211 08:11:42.485671 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:42 crc kubenswrapper[4860]: I1211 08:11:42.485714 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:42 crc kubenswrapper[4860]: I1211 08:11:42.485739 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:42 crc kubenswrapper[4860]: I1211 08:11:42.485756 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:42Z","lastTransitionTime":"2025-12-11T08:11:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:42 crc kubenswrapper[4860]: I1211 08:11:42.494088 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ec3758-85f1-4a22-a4ec-c805472607a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e831db1a15976d3967d56479689572c26f2599a766baa0402897a3008428c56f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a498aa90f914e1f0370df01dc5ef174df113a2a16d86b2e3f24b19a1b99b4c0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8caebd2f4dfcbb89c597029a07c335171889761d2d200206987537fb0a228db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34e91f615b76445ff3241a304bb2964fde59de7e8a8dfff6841b2e628a31e86d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2090cab02ad74ddfd6bb57ff50eba35d18f56b0cde1c230fe011c01994b693bc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 08:11:21.171011 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 08:11:21.177674 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-644352688/tls.crt::/tmp/serving-cert-644352688/tls.key\\\\\\\"\\\\nI1211 08:11:26.508606 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 08:11:26.512741 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 08:11:26.512778 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 08:11:26.512839 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 08:11:26.512858 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 08:11:26.521684 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1211 08:11:26.521700 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1211 08:11:26.521722 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:11:26.521732 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:11:26.521738 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 08:11:26.521746 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 08:11:26.521749 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 08:11:26.521753 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1211 08:11:26.524999 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2274fb298ca7d2b37b060b4dbf9b166d39ba77a108b08edeeb4d67196e72cb3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf0aff8a6ab816b6617b1feeca9146d41bb5e8808bf1db96f8d493500c135d85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf0aff8a6ab816b6617b1feeca9146d41bb5e8808bf1db96f8d493500c135d85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:42Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:42 crc kubenswrapper[4860]: I1211 08:11:42.514620 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:42Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:42 crc kubenswrapper[4860]: I1211 08:11:42.539462 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a41f35db087418d1be16953f10377042abd3f4488e5ac1bc29ca5dcfc00a37e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:42Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:42 crc kubenswrapper[4860]: I1211 08:11:42.556795 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gjhlj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5cfd640a-f3d2-4b28-87e1-5f49f6341e21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knq7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knq7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:42Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gjhlj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:42Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:42 crc kubenswrapper[4860]: I1211 08:11:42.565134 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5cfd640a-f3d2-4b28-87e1-5f49f6341e21-metrics-certs\") pod \"network-metrics-daemon-gjhlj\" (UID: \"5cfd640a-f3d2-4b28-87e1-5f49f6341e21\") " pod="openshift-multus/network-metrics-daemon-gjhlj" Dec 11 08:11:42 crc kubenswrapper[4860]: I1211 08:11:42.565188 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-knq7c\" (UniqueName: \"kubernetes.io/projected/5cfd640a-f3d2-4b28-87e1-5f49f6341e21-kube-api-access-knq7c\") pod \"network-metrics-daemon-gjhlj\" (UID: \"5cfd640a-f3d2-4b28-87e1-5f49f6341e21\") " pod="openshift-multus/network-metrics-daemon-gjhlj" Dec 11 08:11:42 crc kubenswrapper[4860]: I1211 08:11:42.596492 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:11:42 crc kubenswrapper[4860]: E1211 08:11:42.596665 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:11:42 crc kubenswrapper[4860]: I1211 08:11:42.597002 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:11:42 crc kubenswrapper[4860]: E1211 08:11:42.597057 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:11:42 crc kubenswrapper[4860]: I1211 08:11:42.597091 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:11:42 crc kubenswrapper[4860]: E1211 08:11:42.597129 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:11:42 crc kubenswrapper[4860]: I1211 08:11:42.599146 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:42 crc kubenswrapper[4860]: I1211 08:11:42.599169 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:42 crc kubenswrapper[4860]: I1211 08:11:42.599177 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:42 crc kubenswrapper[4860]: I1211 08:11:42.599189 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:42 crc kubenswrapper[4860]: I1211 08:11:42.599197 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:42Z","lastTransitionTime":"2025-12-11T08:11:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:42 crc kubenswrapper[4860]: I1211 08:11:42.602394 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gsx9l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aba5b6b9-f88d-4763-9562-6b4dec606df9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://735ef012c523ddc952d46a431f27b2582a2c301d9e9da0c697a20ff6e2fb5d86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6zw9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gsx9l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:42Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:42 crc kubenswrapper[4860]: I1211 08:11:42.615254 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9aa0aed-62b3-4ea2-82ae-9a653c1101a5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ad4ba1b769c6cb2b34073798f8e3c25eb0177e6d3059484bf365e6981cab710\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://077aa1dd2fe5189ff60a0f2365e47a6a545cca0de7aa6379370908db4ea5838a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d7c919899b74a2b0fcd9221f3f769a9d013267354b200a282669d342d6413d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5483384f1d2ebc1fefa9706c9e67a3dc5a00ff643f803e3f2707c252e27ad94\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:42Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:42 crc kubenswrapper[4860]: I1211 08:11:42.631993 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31c30642-6e60-41b4-a477-0d802424e0aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79969402f5c33852a26fc839f37a45febf4f2377d77823284fbb01769e5585fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2d4d5de7106175087d91acb1f0c7552014f0c08bfcd65cbbf184d3f8d7826ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-99qgp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:42Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:42 crc kubenswrapper[4860]: I1211 08:11:42.648233 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-26nc6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1faa733-f408-41fd-b7aa-0cd86283976a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fce5b642b843ed30823c94e0b0c3d8bb1e5d2d93d691f7b8ef901ece03b0ac6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ac45db4f858a1c18a6b87472fc1cea2300f5845f1ad62c22260374995548a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ac45db4f858a1c18a6b87472fc1cea2300f5845f1ad62c22260374995548a0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b35296a5eb17dd2bb5e72c86c1cafdb30b7712cb330264f935d3b4fff05cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c6b35296a5eb17dd2bb5e72c86c1cafdb30b7712cb330264f935d3b4fff05cd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b9e82d6d3a247309ef5bdbf5b3a86ec083add2a138b0b237801e77b09ee96a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b9e82d6d3a247309ef5bdbf5b3a86ec083add2a138b0b237801e77b09ee96a1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://908ee695e78145ef4375c4097d9f1e772502922d02a2fcb4566172a6ad9caaf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://908ee695e78145ef4375c4097d9f1e772502922d02a2fcb4566172a6ad9caaf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afb0a097f3556a8f9ed7cf8f762cb99c082fe0a36b8d8d3038bf188b08fe9a84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afb0a097f3556a8f9ed7cf8f762cb99c082fe0a36b8d8d3038bf188b08fe9a84\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce18f4f06db091e658684fa0bbc858e19b9f03010acf13b3ffccb1a6e3f88021\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce18f4f06db091e658684fa0bbc858e19b9f03010acf13b3ffccb1a6e3f88021\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-26nc6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:42Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:42 crc kubenswrapper[4860]: I1211 08:11:42.666441 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5cfd640a-f3d2-4b28-87e1-5f49f6341e21-metrics-certs\") pod \"network-metrics-daemon-gjhlj\" (UID: \"5cfd640a-f3d2-4b28-87e1-5f49f6341e21\") " pod="openshift-multus/network-metrics-daemon-gjhlj" Dec 11 08:11:42 crc kubenswrapper[4860]: I1211 08:11:42.666491 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-knq7c\" (UniqueName: \"kubernetes.io/projected/5cfd640a-f3d2-4b28-87e1-5f49f6341e21-kube-api-access-knq7c\") pod \"network-metrics-daemon-gjhlj\" (UID: \"5cfd640a-f3d2-4b28-87e1-5f49f6341e21\") " pod="openshift-multus/network-metrics-daemon-gjhlj" Dec 11 08:11:42 crc kubenswrapper[4860]: E1211 08:11:42.666673 4860 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 11 08:11:42 crc kubenswrapper[4860]: E1211 08:11:42.666752 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5cfd640a-f3d2-4b28-87e1-5f49f6341e21-metrics-certs podName:5cfd640a-f3d2-4b28-87e1-5f49f6341e21 nodeName:}" failed. No retries permitted until 2025-12-11 08:11:43.166729236 +0000 UTC m=+35.895248351 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5cfd640a-f3d2-4b28-87e1-5f49f6341e21-metrics-certs") pod "network-metrics-daemon-gjhlj" (UID: "5cfd640a-f3d2-4b28-87e1-5f49f6341e21") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 11 08:11:42 crc kubenswrapper[4860]: I1211 08:11:42.670340 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0589a204-c98c-417d-8256-bfe3bf747660\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3333b4346f59b04017e32d61f457ba7ff457166731128f339d7987ae6b56b06d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52bb8ade314afb9494f4e026d0af2aea7b1970d64dc66102f55e257ae4678154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dfcc4370edaae36a5a32b0a9c29290c50f8e4f427e4e147dc3c779beaa8b780\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d856ae95e9077524c518d82d10fb6b93b7fbcaaf2b720ffafe6f7fae58bd7caa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1426084ac580895c42354630910dd8c8e3a48ddd25f22abce1d28dc3de332a63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9984d5d242bb258405efe91fc9caf9e7ea592834445df2e78add9553a921c76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://526c2b9214d19b9ebe7a7078fd0459642a53e732d034602b2b8d2b8a33f4d18b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb80d9928e8349fc5963227d135355ad7ba820d4fd89877af601d005ca03334b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17cbcb49c06f3a9059cd5a0f05eedb8c785da30a3e502e31bae2f57b3486cb14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17cbcb49c06f3a9059cd5a0f05eedb8c785da30a3e502e31bae2f57b3486cb14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xvnqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:42Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:42 crc kubenswrapper[4860]: I1211 08:11:42.684706 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-knq7c\" (UniqueName: \"kubernetes.io/projected/5cfd640a-f3d2-4b28-87e1-5f49f6341e21-kube-api-access-knq7c\") pod \"network-metrics-daemon-gjhlj\" (UID: \"5cfd640a-f3d2-4b28-87e1-5f49f6341e21\") " pod="openshift-multus/network-metrics-daemon-gjhlj" Dec 11 08:11:42 crc kubenswrapper[4860]: I1211 08:11:42.690277 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9872be1e-b995-4a00-904b-875dff96ad56\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f74a32a8fefea704cc73ffd1114083dab68aa44dc33f8d9d432d2d24216546c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbe00b46063d3968d59c393b87178e135f982abb7d0f32fa798b212d4b83df6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eedc9988e1a4a3c8cae730dc4a352103a6d665535b825e157d0f87c331b137c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2137fee4b7abc9765f9517d7b3b6ea796601a31d03151bf167d17136ced8b25a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f683e48f3577a68862eb6302383538376e41f49b20b1e0ea3f1b3b143f24872a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5de794187456b3576b969069a3b839d75c7918418a97fa35dff8bb40bb76e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5de794187456b3576b969069a3b839d75c7918418a97fa35dff8bb40bb76e39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08ce42b974fba00c409c0af971f0de351e50f18b2f06f8e722150da3239bb908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08ce42b974fba00c409c0af971f0de351e50f18b2f06f8e722150da3239bb908\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cfac6be03e3cfa0b0f3450b48b67957f2d5172b9c65e27f14c2a260bda5d423f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cfac6be03e3cfa0b0f3450b48b67957f2d5172b9c65e27f14c2a260bda5d423f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:42Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:42 crc kubenswrapper[4860]: I1211 08:11:42.701322 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:42 crc kubenswrapper[4860]: I1211 08:11:42.701372 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:42 crc kubenswrapper[4860]: I1211 08:11:42.701386 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:42 crc kubenswrapper[4860]: I1211 08:11:42.701403 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:42 crc kubenswrapper[4860]: I1211 08:11:42.701415 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:42Z","lastTransitionTime":"2025-12-11T08:11:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:42 crc kubenswrapper[4860]: I1211 08:11:42.710786 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-pznwl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5310c1fc-66c6-40aa-b1bf-5a59a2410f9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2d6ff00ce322a28607bc3f6db05c5e4fb14acb6777ea315786ef53963ca504a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hkt48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-pznwl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:42Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:42 crc kubenswrapper[4860]: I1211 08:11:42.721875 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-854v6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e6529ba7-2341-43ac-a187-95ab530fe775\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kh22b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kh22b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-854v6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:42Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:42 crc kubenswrapper[4860]: I1211 08:11:42.731740 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6xq7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a75ce3c-5e1e-4b9f-bbde-7a3b0cce5eae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe8d4d675567b41728e476351811a93c47b19868ccbadab8c613a4eb8eca8562\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5c6nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6xq7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:42Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:42 crc kubenswrapper[4860]: I1211 08:11:42.744410 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:42Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:42 crc kubenswrapper[4860]: I1211 08:11:42.757387 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d8c2f28628dc2f1ffbdd98a55670e22cf6cdc8cf2c6442e9d1f8e5466b2f12e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://074ea63beac52bd196507ee7cfb1c3d7fc2d37d5ec6faa4d47dde62b922e1c43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:42Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:42 crc kubenswrapper[4860]: I1211 08:11:42.768449 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:42Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:42 crc kubenswrapper[4860]: I1211 08:11:42.782124 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25a6a154da2b74952951ae94efc914ba078ea4bbe5893360b3c84d46be6bf973\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:42Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:42 crc kubenswrapper[4860]: I1211 08:11:42.803785 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:42 crc kubenswrapper[4860]: I1211 08:11:42.803819 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:42 crc kubenswrapper[4860]: I1211 08:11:42.803827 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:42 crc kubenswrapper[4860]: I1211 08:11:42.803840 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:42 crc kubenswrapper[4860]: I1211 08:11:42.803850 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:42Z","lastTransitionTime":"2025-12-11T08:11:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:42 crc kubenswrapper[4860]: I1211 08:11:42.891826 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-xvnqp_0589a204-c98c-417d-8256-bfe3bf747660/ovnkube-controller/0.log" Dec 11 08:11:42 crc kubenswrapper[4860]: I1211 08:11:42.894055 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" event={"ID":"0589a204-c98c-417d-8256-bfe3bf747660","Type":"ContainerDied","Data":"526c2b9214d19b9ebe7a7078fd0459642a53e732d034602b2b8d2b8a33f4d18b"} Dec 11 08:11:42 crc kubenswrapper[4860]: I1211 08:11:42.894663 4860 scope.go:117] "RemoveContainer" containerID="526c2b9214d19b9ebe7a7078fd0459642a53e732d034602b2b8d2b8a33f4d18b" Dec 11 08:11:42 crc kubenswrapper[4860]: I1211 08:11:42.894715 4860 generic.go:334] "Generic (PLEG): container finished" podID="0589a204-c98c-417d-8256-bfe3bf747660" containerID="526c2b9214d19b9ebe7a7078fd0459642a53e732d034602b2b8d2b8a33f4d18b" exitCode=1 Dec 11 08:11:42 crc kubenswrapper[4860]: I1211 08:11:42.896152 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-854v6" event={"ID":"e6529ba7-2341-43ac-a187-95ab530fe775","Type":"ContainerStarted","Data":"f0d4e3f761ad15ae1424204e40b9d3f60188ab4a9047bd4b5ac6a8a8ca1fb385"} Dec 11 08:11:42 crc kubenswrapper[4860]: I1211 08:11:42.896174 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-854v6" event={"ID":"e6529ba7-2341-43ac-a187-95ab530fe775","Type":"ContainerStarted","Data":"8986538a421f0184f52f288b0e171da695591caf2b10c67ad2dce1073a6e6016"} Dec 11 08:11:42 crc kubenswrapper[4860]: I1211 08:11:42.905693 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:42 crc kubenswrapper[4860]: I1211 08:11:42.905720 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:42 crc kubenswrapper[4860]: I1211 08:11:42.905730 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:42 crc kubenswrapper[4860]: I1211 08:11:42.905747 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:42 crc kubenswrapper[4860]: I1211 08:11:42.905758 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:42Z","lastTransitionTime":"2025-12-11T08:11:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:42 crc kubenswrapper[4860]: I1211 08:11:42.919003 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gjhlj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5cfd640a-f3d2-4b28-87e1-5f49f6341e21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knq7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knq7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:42Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gjhlj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:42Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:42 crc kubenswrapper[4860]: I1211 08:11:42.934345 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ec3758-85f1-4a22-a4ec-c805472607a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e831db1a15976d3967d56479689572c26f2599a766baa0402897a3008428c56f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a498aa90f914e1f0370df01dc5ef174df113a2a16d86b2e3f24b19a1b99b4c0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8caebd2f4dfcbb89c597029a07c335171889761d2d200206987537fb0a228db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34e91f615b76445ff3241a304bb2964fde59de7e8a8dfff6841b2e628a31e86d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2090cab02ad74ddfd6bb57ff50eba35d18f56b0cde1c230fe011c01994b693bc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 08:11:21.171011 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 08:11:21.177674 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-644352688/tls.crt::/tmp/serving-cert-644352688/tls.key\\\\\\\"\\\\nI1211 08:11:26.508606 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 08:11:26.512741 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 08:11:26.512778 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 08:11:26.512839 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 08:11:26.512858 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 08:11:26.521684 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1211 08:11:26.521700 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1211 08:11:26.521722 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:11:26.521732 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:11:26.521738 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 08:11:26.521746 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 08:11:26.521749 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 08:11:26.521753 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1211 08:11:26.524999 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2274fb298ca7d2b37b060b4dbf9b166d39ba77a108b08edeeb4d67196e72cb3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf0aff8a6ab816b6617b1feeca9146d41bb5e8808bf1db96f8d493500c135d85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf0aff8a6ab816b6617b1feeca9146d41bb5e8808bf1db96f8d493500c135d85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:42Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:42 crc kubenswrapper[4860]: I1211 08:11:42.947525 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:42Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:42 crc kubenswrapper[4860]: I1211 08:11:42.961109 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a41f35db087418d1be16953f10377042abd3f4488e5ac1bc29ca5dcfc00a37e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:42Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:42 crc kubenswrapper[4860]: I1211 08:11:42.975238 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-26nc6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1faa733-f408-41fd-b7aa-0cd86283976a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fce5b642b843ed30823c94e0b0c3d8bb1e5d2d93d691f7b8ef901ece03b0ac6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ac45db4f858a1c18a6b87472fc1cea2300f5845f1ad62c22260374995548a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ac45db4f858a1c18a6b87472fc1cea2300f5845f1ad62c22260374995548a0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b35296a5eb17dd2bb5e72c86c1cafdb30b7712cb330264f935d3b4fff05cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c6b35296a5eb17dd2bb5e72c86c1cafdb30b7712cb330264f935d3b4fff05cd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b9e82d6d3a247309ef5bdbf5b3a86ec083add2a138b0b237801e77b09ee96a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b9e82d6d3a247309ef5bdbf5b3a86ec083add2a138b0b237801e77b09ee96a1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://908ee695e78145ef4375c4097d9f1e772502922d02a2fcb4566172a6ad9caaf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://908ee695e78145ef4375c4097d9f1e772502922d02a2fcb4566172a6ad9caaf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afb0a097f3556a8f9ed7cf8f762cb99c082fe0a36b8d8d3038bf188b08fe9a84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afb0a097f3556a8f9ed7cf8f762cb99c082fe0a36b8d8d3038bf188b08fe9a84\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce18f4f06db091e658684fa0bbc858e19b9f03010acf13b3ffccb1a6e3f88021\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce18f4f06db091e658684fa0bbc858e19b9f03010acf13b3ffccb1a6e3f88021\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-26nc6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:42Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:42 crc kubenswrapper[4860]: I1211 08:11:42.994290 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0589a204-c98c-417d-8256-bfe3bf747660\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3333b4346f59b04017e32d61f457ba7ff457166731128f339d7987ae6b56b06d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52bb8ade314afb9494f4e026d0af2aea7b1970d64dc66102f55e257ae4678154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dfcc4370edaae36a5a32b0a9c29290c50f8e4f427e4e147dc3c779beaa8b780\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d856ae95e9077524c518d82d10fb6b93b7fbcaaf2b720ffafe6f7fae58bd7caa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1426084ac580895c42354630910dd8c8e3a48ddd25f22abce1d28dc3de332a63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9984d5d242bb258405efe91fc9caf9e7ea592834445df2e78add9553a921c76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://526c2b9214d19b9ebe7a7078fd0459642a53e732d034602b2b8d2b8a33f4d18b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://526c2b9214d19b9ebe7a7078fd0459642a53e732d034602b2b8d2b8a33f4d18b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"message\\\":\\\"workPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1211 08:11:42.548303 6122 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1211 08:11:42.548667 6122 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1211 08:11:42.548830 6122 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1211 08:11:42.548871 6122 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1211 08:11:42.548960 6122 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1211 08:11:42.549108 6122 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1211 08:11:42.549161 6122 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1211 08:11:42.550530 6122 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb80d9928e8349fc5963227d135355ad7ba820d4fd89877af601d005ca03334b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17cbcb49c06f3a9059cd5a0f05eedb8c785da30a3e502e31bae2f57b3486cb14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17cbcb49c06f3a9059cd5a0f05eedb8c785da30a3e502e31bae2f57b3486cb14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xvnqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:42Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:43 crc kubenswrapper[4860]: I1211 08:11:43.005544 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gsx9l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aba5b6b9-f88d-4763-9562-6b4dec606df9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://735ef012c523ddc952d46a431f27b2582a2c301d9e9da0c697a20ff6e2fb5d86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6zw9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gsx9l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:43Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:43 crc kubenswrapper[4860]: I1211 08:11:43.007780 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:43 crc kubenswrapper[4860]: I1211 08:11:43.007799 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:43 crc kubenswrapper[4860]: I1211 08:11:43.007808 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:43 crc kubenswrapper[4860]: I1211 08:11:43.007822 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:43 crc kubenswrapper[4860]: I1211 08:11:43.007836 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:43Z","lastTransitionTime":"2025-12-11T08:11:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:43 crc kubenswrapper[4860]: I1211 08:11:43.018172 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9aa0aed-62b3-4ea2-82ae-9a653c1101a5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ad4ba1b769c6cb2b34073798f8e3c25eb0177e6d3059484bf365e6981cab710\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://077aa1dd2fe5189ff60a0f2365e47a6a545cca0de7aa6379370908db4ea5838a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d7c919899b74a2b0fcd9221f3f769a9d013267354b200a282669d342d6413d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5483384f1d2ebc1fefa9706c9e67a3dc5a00ff643f803e3f2707c252e27ad94\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:43Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:43 crc kubenswrapper[4860]: I1211 08:11:43.029424 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31c30642-6e60-41b4-a477-0d802424e0aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79969402f5c33852a26fc839f37a45febf4f2377d77823284fbb01769e5585fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2d4d5de7106175087d91acb1f0c7552014f0c08bfcd65cbbf184d3f8d7826ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-99qgp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:43Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:43 crc kubenswrapper[4860]: I1211 08:11:43.048844 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9872be1e-b995-4a00-904b-875dff96ad56\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f74a32a8fefea704cc73ffd1114083dab68aa44dc33f8d9d432d2d24216546c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbe00b46063d3968d59c393b87178e135f982abb7d0f32fa798b212d4b83df6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eedc9988e1a4a3c8cae730dc4a352103a6d665535b825e157d0f87c331b137c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2137fee4b7abc9765f9517d7b3b6ea796601a31d03151bf167d17136ced8b25a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f683e48f3577a68862eb6302383538376e41f49b20b1e0ea3f1b3b143f24872a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5de794187456b3576b969069a3b839d75c7918418a97fa35dff8bb40bb76e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5de794187456b3576b969069a3b839d75c7918418a97fa35dff8bb40bb76e39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08ce42b974fba00c409c0af971f0de351e50f18b2f06f8e722150da3239bb908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08ce42b974fba00c409c0af971f0de351e50f18b2f06f8e722150da3239bb908\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cfac6be03e3cfa0b0f3450b48b67957f2d5172b9c65e27f14c2a260bda5d423f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cfac6be03e3cfa0b0f3450b48b67957f2d5172b9c65e27f14c2a260bda5d423f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:43Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:43 crc kubenswrapper[4860]: I1211 08:11:43.065604 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-pznwl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5310c1fc-66c6-40aa-b1bf-5a59a2410f9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2d6ff00ce322a28607bc3f6db05c5e4fb14acb6777ea315786ef53963ca504a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hkt48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-pznwl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:43Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:43 crc kubenswrapper[4860]: I1211 08:11:43.077163 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-854v6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e6529ba7-2341-43ac-a187-95ab530fe775\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kh22b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kh22b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-854v6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:43Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:43 crc kubenswrapper[4860]: I1211 08:11:43.088758 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:43Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:43 crc kubenswrapper[4860]: I1211 08:11:43.101364 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25a6a154da2b74952951ae94efc914ba078ea4bbe5893360b3c84d46be6bf973\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:43Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:43 crc kubenswrapper[4860]: I1211 08:11:43.117511 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:43 crc kubenswrapper[4860]: I1211 08:11:43.117573 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:43 crc kubenswrapper[4860]: I1211 08:11:43.117593 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:43 crc kubenswrapper[4860]: I1211 08:11:43.117881 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:43 crc kubenswrapper[4860]: I1211 08:11:43.117906 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:43Z","lastTransitionTime":"2025-12-11T08:11:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:43 crc kubenswrapper[4860]: I1211 08:11:43.123279 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6xq7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a75ce3c-5e1e-4b9f-bbde-7a3b0cce5eae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe8d4d675567b41728e476351811a93c47b19868ccbadab8c613a4eb8eca8562\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5c6nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6xq7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:43Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:43 crc kubenswrapper[4860]: I1211 08:11:43.146635 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:43Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:43 crc kubenswrapper[4860]: I1211 08:11:43.166063 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d8c2f28628dc2f1ffbdd98a55670e22cf6cdc8cf2c6442e9d1f8e5466b2f12e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://074ea63beac52bd196507ee7cfb1c3d7fc2d37d5ec6faa4d47dde62b922e1c43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:43Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:43 crc kubenswrapper[4860]: I1211 08:11:43.171912 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5cfd640a-f3d2-4b28-87e1-5f49f6341e21-metrics-certs\") pod \"network-metrics-daemon-gjhlj\" (UID: \"5cfd640a-f3d2-4b28-87e1-5f49f6341e21\") " pod="openshift-multus/network-metrics-daemon-gjhlj" Dec 11 08:11:43 crc kubenswrapper[4860]: E1211 08:11:43.172028 4860 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 11 08:11:43 crc kubenswrapper[4860]: E1211 08:11:43.172070 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5cfd640a-f3d2-4b28-87e1-5f49f6341e21-metrics-certs podName:5cfd640a-f3d2-4b28-87e1-5f49f6341e21 nodeName:}" failed. No retries permitted until 2025-12-11 08:11:44.172058876 +0000 UTC m=+36.900577931 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5cfd640a-f3d2-4b28-87e1-5f49f6341e21-metrics-certs") pod "network-metrics-daemon-gjhlj" (UID: "5cfd640a-f3d2-4b28-87e1-5f49f6341e21") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 11 08:11:43 crc kubenswrapper[4860]: I1211 08:11:43.178499 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31c30642-6e60-41b4-a477-0d802424e0aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79969402f5c33852a26fc839f37a45febf4f2377d77823284fbb01769e5585fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2d4d5de7106175087d91acb1f0c7552014f0c08bfcd65cbbf184d3f8d7826ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-99qgp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:43Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:43 crc kubenswrapper[4860]: I1211 08:11:43.191922 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-26nc6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1faa733-f408-41fd-b7aa-0cd86283976a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fce5b642b843ed30823c94e0b0c3d8bb1e5d2d93d691f7b8ef901ece03b0ac6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ac45db4f858a1c18a6b87472fc1cea2300f5845f1ad62c22260374995548a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ac45db4f858a1c18a6b87472fc1cea2300f5845f1ad62c22260374995548a0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b35296a5eb17dd2bb5e72c86c1cafdb30b7712cb330264f935d3b4fff05cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c6b35296a5eb17dd2bb5e72c86c1cafdb30b7712cb330264f935d3b4fff05cd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b9e82d6d3a247309ef5bdbf5b3a86ec083add2a138b0b237801e77b09ee96a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b9e82d6d3a247309ef5bdbf5b3a86ec083add2a138b0b237801e77b09ee96a1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://908ee695e78145ef4375c4097d9f1e772502922d02a2fcb4566172a6ad9caaf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://908ee695e78145ef4375c4097d9f1e772502922d02a2fcb4566172a6ad9caaf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afb0a097f3556a8f9ed7cf8f762cb99c082fe0a36b8d8d3038bf188b08fe9a84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afb0a097f3556a8f9ed7cf8f762cb99c082fe0a36b8d8d3038bf188b08fe9a84\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce18f4f06db091e658684fa0bbc858e19b9f03010acf13b3ffccb1a6e3f88021\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce18f4f06db091e658684fa0bbc858e19b9f03010acf13b3ffccb1a6e3f88021\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-26nc6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:43Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:43 crc kubenswrapper[4860]: I1211 08:11:43.211789 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0589a204-c98c-417d-8256-bfe3bf747660\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3333b4346f59b04017e32d61f457ba7ff457166731128f339d7987ae6b56b06d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52bb8ade314afb9494f4e026d0af2aea7b1970d64dc66102f55e257ae4678154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dfcc4370edaae36a5a32b0a9c29290c50f8e4f427e4e147dc3c779beaa8b780\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d856ae95e9077524c518d82d10fb6b93b7fbcaaf2b720ffafe6f7fae58bd7caa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1426084ac580895c42354630910dd8c8e3a48ddd25f22abce1d28dc3de332a63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9984d5d242bb258405efe91fc9caf9e7ea592834445df2e78add9553a921c76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://526c2b9214d19b9ebe7a7078fd0459642a53e732d034602b2b8d2b8a33f4d18b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://526c2b9214d19b9ebe7a7078fd0459642a53e732d034602b2b8d2b8a33f4d18b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"message\\\":\\\"workPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1211 08:11:42.548303 6122 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1211 08:11:42.548667 6122 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1211 08:11:42.548830 6122 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1211 08:11:42.548871 6122 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1211 08:11:42.548960 6122 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1211 08:11:42.549108 6122 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1211 08:11:42.549161 6122 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1211 08:11:42.550530 6122 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb80d9928e8349fc5963227d135355ad7ba820d4fd89877af601d005ca03334b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17cbcb49c06f3a9059cd5a0f05eedb8c785da30a3e502e31bae2f57b3486cb14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17cbcb49c06f3a9059cd5a0f05eedb8c785da30a3e502e31bae2f57b3486cb14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xvnqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:43Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:43 crc kubenswrapper[4860]: I1211 08:11:43.220484 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:43 crc kubenswrapper[4860]: I1211 08:11:43.220521 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:43 crc kubenswrapper[4860]: I1211 08:11:43.220534 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:43 crc kubenswrapper[4860]: I1211 08:11:43.220748 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:43 crc kubenswrapper[4860]: I1211 08:11:43.220759 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:43Z","lastTransitionTime":"2025-12-11T08:11:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:43 crc kubenswrapper[4860]: I1211 08:11:43.222443 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gsx9l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aba5b6b9-f88d-4763-9562-6b4dec606df9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://735ef012c523ddc952d46a431f27b2582a2c301d9e9da0c697a20ff6e2fb5d86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6zw9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gsx9l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:43Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:43 crc kubenswrapper[4860]: I1211 08:11:43.234602 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9aa0aed-62b3-4ea2-82ae-9a653c1101a5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ad4ba1b769c6cb2b34073798f8e3c25eb0177e6d3059484bf365e6981cab710\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://077aa1dd2fe5189ff60a0f2365e47a6a545cca0de7aa6379370908db4ea5838a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d7c919899b74a2b0fcd9221f3f769a9d013267354b200a282669d342d6413d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5483384f1d2ebc1fefa9706c9e67a3dc5a00ff643f803e3f2707c252e27ad94\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:43Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:43 crc kubenswrapper[4860]: I1211 08:11:43.250958 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-854v6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e6529ba7-2341-43ac-a187-95ab530fe775\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8986538a421f0184f52f288b0e171da695591caf2b10c67ad2dce1073a6e6016\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kh22b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0d4e3f761ad15ae1424204e40b9d3f60188ab4a9047bd4b5ac6a8a8ca1fb385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kh22b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-854v6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:43Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:43 crc kubenswrapper[4860]: I1211 08:11:43.270551 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9872be1e-b995-4a00-904b-875dff96ad56\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f74a32a8fefea704cc73ffd1114083dab68aa44dc33f8d9d432d2d24216546c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbe00b46063d3968d59c393b87178e135f982abb7d0f32fa798b212d4b83df6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eedc9988e1a4a3c8cae730dc4a352103a6d665535b825e157d0f87c331b137c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2137fee4b7abc9765f9517d7b3b6ea796601a31d03151bf167d17136ced8b25a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f683e48f3577a68862eb6302383538376e41f49b20b1e0ea3f1b3b143f24872a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5de794187456b3576b969069a3b839d75c7918418a97fa35dff8bb40bb76e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5de794187456b3576b969069a3b839d75c7918418a97fa35dff8bb40bb76e39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08ce42b974fba00c409c0af971f0de351e50f18b2f06f8e722150da3239bb908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08ce42b974fba00c409c0af971f0de351e50f18b2f06f8e722150da3239bb908\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cfac6be03e3cfa0b0f3450b48b67957f2d5172b9c65e27f14c2a260bda5d423f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cfac6be03e3cfa0b0f3450b48b67957f2d5172b9c65e27f14c2a260bda5d423f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:43Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:43 crc kubenswrapper[4860]: I1211 08:11:43.291839 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-pznwl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5310c1fc-66c6-40aa-b1bf-5a59a2410f9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2d6ff00ce322a28607bc3f6db05c5e4fb14acb6777ea315786ef53963ca504a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hkt48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-pznwl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:43Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:43 crc kubenswrapper[4860]: I1211 08:11:43.305417 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d8c2f28628dc2f1ffbdd98a55670e22cf6cdc8cf2c6442e9d1f8e5466b2f12e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://074ea63beac52bd196507ee7cfb1c3d7fc2d37d5ec6faa4d47dde62b922e1c43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:43Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:43 crc kubenswrapper[4860]: I1211 08:11:43.315149 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:43Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:43 crc kubenswrapper[4860]: I1211 08:11:43.323675 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:43 crc kubenswrapper[4860]: I1211 08:11:43.323715 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:43 crc kubenswrapper[4860]: I1211 08:11:43.323725 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:43 crc kubenswrapper[4860]: I1211 08:11:43.323741 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:43 crc kubenswrapper[4860]: I1211 08:11:43.323752 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:43Z","lastTransitionTime":"2025-12-11T08:11:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:43 crc kubenswrapper[4860]: I1211 08:11:43.326391 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25a6a154da2b74952951ae94efc914ba078ea4bbe5893360b3c84d46be6bf973\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:43Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:43 crc kubenswrapper[4860]: I1211 08:11:43.337406 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6xq7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a75ce3c-5e1e-4b9f-bbde-7a3b0cce5eae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe8d4d675567b41728e476351811a93c47b19868ccbadab8c613a4eb8eca8562\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5c6nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6xq7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:43Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:43 crc kubenswrapper[4860]: I1211 08:11:43.353665 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:43Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:43 crc kubenswrapper[4860]: I1211 08:11:43.365222 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a41f35db087418d1be16953f10377042abd3f4488e5ac1bc29ca5dcfc00a37e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:43Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:43 crc kubenswrapper[4860]: I1211 08:11:43.374365 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gjhlj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5cfd640a-f3d2-4b28-87e1-5f49f6341e21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knq7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knq7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:42Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gjhlj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:43Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:43 crc kubenswrapper[4860]: I1211 08:11:43.385654 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ec3758-85f1-4a22-a4ec-c805472607a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e831db1a15976d3967d56479689572c26f2599a766baa0402897a3008428c56f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a498aa90f914e1f0370df01dc5ef174df113a2a16d86b2e3f24b19a1b99b4c0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8caebd2f4dfcbb89c597029a07c335171889761d2d200206987537fb0a228db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34e91f615b76445ff3241a304bb2964fde59de7e8a8dfff6841b2e628a31e86d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2090cab02ad74ddfd6bb57ff50eba35d18f56b0cde1c230fe011c01994b693bc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 08:11:21.171011 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 08:11:21.177674 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-644352688/tls.crt::/tmp/serving-cert-644352688/tls.key\\\\\\\"\\\\nI1211 08:11:26.508606 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 08:11:26.512741 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 08:11:26.512778 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 08:11:26.512839 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 08:11:26.512858 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 08:11:26.521684 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1211 08:11:26.521700 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1211 08:11:26.521722 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:11:26.521732 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:11:26.521738 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 08:11:26.521746 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 08:11:26.521749 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 08:11:26.521753 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1211 08:11:26.524999 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2274fb298ca7d2b37b060b4dbf9b166d39ba77a108b08edeeb4d67196e72cb3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf0aff8a6ab816b6617b1feeca9146d41bb5e8808bf1db96f8d493500c135d85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf0aff8a6ab816b6617b1feeca9146d41bb5e8808bf1db96f8d493500c135d85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:43Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:43 crc kubenswrapper[4860]: I1211 08:11:43.416008 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:43Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:43 crc kubenswrapper[4860]: I1211 08:11:43.433572 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:43 crc kubenswrapper[4860]: I1211 08:11:43.433623 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:43 crc kubenswrapper[4860]: I1211 08:11:43.433634 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:43 crc kubenswrapper[4860]: I1211 08:11:43.433666 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:43 crc kubenswrapper[4860]: I1211 08:11:43.433677 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:43Z","lastTransitionTime":"2025-12-11T08:11:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:43 crc kubenswrapper[4860]: I1211 08:11:43.535996 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:43 crc kubenswrapper[4860]: I1211 08:11:43.536220 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:43 crc kubenswrapper[4860]: I1211 08:11:43.536306 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:43 crc kubenswrapper[4860]: I1211 08:11:43.536417 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:43 crc kubenswrapper[4860]: I1211 08:11:43.536491 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:43Z","lastTransitionTime":"2025-12-11T08:11:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:43 crc kubenswrapper[4860]: I1211 08:11:43.639261 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:43 crc kubenswrapper[4860]: I1211 08:11:43.639287 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:43 crc kubenswrapper[4860]: I1211 08:11:43.639295 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:43 crc kubenswrapper[4860]: I1211 08:11:43.639306 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:43 crc kubenswrapper[4860]: I1211 08:11:43.639314 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:43Z","lastTransitionTime":"2025-12-11T08:11:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:43 crc kubenswrapper[4860]: I1211 08:11:43.742218 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:43 crc kubenswrapper[4860]: I1211 08:11:43.742303 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:43 crc kubenswrapper[4860]: I1211 08:11:43.742325 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:43 crc kubenswrapper[4860]: I1211 08:11:43.742356 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:43 crc kubenswrapper[4860]: I1211 08:11:43.742381 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:43Z","lastTransitionTime":"2025-12-11T08:11:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:43 crc kubenswrapper[4860]: I1211 08:11:43.845097 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:43 crc kubenswrapper[4860]: I1211 08:11:43.845152 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:43 crc kubenswrapper[4860]: I1211 08:11:43.845168 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:43 crc kubenswrapper[4860]: I1211 08:11:43.845187 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:43 crc kubenswrapper[4860]: I1211 08:11:43.845201 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:43Z","lastTransitionTime":"2025-12-11T08:11:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:43 crc kubenswrapper[4860]: I1211 08:11:43.903039 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-xvnqp_0589a204-c98c-417d-8256-bfe3bf747660/ovnkube-controller/0.log" Dec 11 08:11:43 crc kubenswrapper[4860]: I1211 08:11:43.906753 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" event={"ID":"0589a204-c98c-417d-8256-bfe3bf747660","Type":"ContainerStarted","Data":"592dbdb2905039d89bdf741dbd0fd1e1fbd15acbabecdad6f07c758657fa610a"} Dec 11 08:11:43 crc kubenswrapper[4860]: I1211 08:11:43.907090 4860 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 11 08:11:43 crc kubenswrapper[4860]: I1211 08:11:43.928003 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9aa0aed-62b3-4ea2-82ae-9a653c1101a5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ad4ba1b769c6cb2b34073798f8e3c25eb0177e6d3059484bf365e6981cab710\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://077aa1dd2fe5189ff60a0f2365e47a6a545cca0de7aa6379370908db4ea5838a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d7c919899b74a2b0fcd9221f3f769a9d013267354b200a282669d342d6413d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5483384f1d2ebc1fefa9706c9e67a3dc5a00ff643f803e3f2707c252e27ad94\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:43Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:43 crc kubenswrapper[4860]: I1211 08:11:43.941554 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31c30642-6e60-41b4-a477-0d802424e0aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79969402f5c33852a26fc839f37a45febf4f2377d77823284fbb01769e5585fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2d4d5de7106175087d91acb1f0c7552014f0c08bfcd65cbbf184d3f8d7826ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-99qgp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:43Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:43 crc kubenswrapper[4860]: I1211 08:11:43.947119 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:43 crc kubenswrapper[4860]: I1211 08:11:43.947175 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:43 crc kubenswrapper[4860]: I1211 08:11:43.947195 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:43 crc kubenswrapper[4860]: I1211 08:11:43.947217 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:43 crc kubenswrapper[4860]: I1211 08:11:43.947241 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:43Z","lastTransitionTime":"2025-12-11T08:11:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:43 crc kubenswrapper[4860]: I1211 08:11:43.953555 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-26nc6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1faa733-f408-41fd-b7aa-0cd86283976a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fce5b642b843ed30823c94e0b0c3d8bb1e5d2d93d691f7b8ef901ece03b0ac6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ac45db4f858a1c18a6b87472fc1cea2300f5845f1ad62c22260374995548a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ac45db4f858a1c18a6b87472fc1cea2300f5845f1ad62c22260374995548a0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b35296a5eb17dd2bb5e72c86c1cafdb30b7712cb330264f935d3b4fff05cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c6b35296a5eb17dd2bb5e72c86c1cafdb30b7712cb330264f935d3b4fff05cd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b9e82d6d3a247309ef5bdbf5b3a86ec083add2a138b0b237801e77b09ee96a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b9e82d6d3a247309ef5bdbf5b3a86ec083add2a138b0b237801e77b09ee96a1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://908ee695e78145ef4375c4097d9f1e772502922d02a2fcb4566172a6ad9caaf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://908ee695e78145ef4375c4097d9f1e772502922d02a2fcb4566172a6ad9caaf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afb0a097f3556a8f9ed7cf8f762cb99c082fe0a36b8d8d3038bf188b08fe9a84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afb0a097f3556a8f9ed7cf8f762cb99c082fe0a36b8d8d3038bf188b08fe9a84\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce18f4f06db091e658684fa0bbc858e19b9f03010acf13b3ffccb1a6e3f88021\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce18f4f06db091e658684fa0bbc858e19b9f03010acf13b3ffccb1a6e3f88021\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-26nc6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:43Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:43 crc kubenswrapper[4860]: I1211 08:11:43.972857 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0589a204-c98c-417d-8256-bfe3bf747660\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3333b4346f59b04017e32d61f457ba7ff457166731128f339d7987ae6b56b06d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52bb8ade314afb9494f4e026d0af2aea7b1970d64dc66102f55e257ae4678154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dfcc4370edaae36a5a32b0a9c29290c50f8e4f427e4e147dc3c779beaa8b780\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d856ae95e9077524c518d82d10fb6b93b7fbcaaf2b720ffafe6f7fae58bd7caa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1426084ac580895c42354630910dd8c8e3a48ddd25f22abce1d28dc3de332a63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9984d5d242bb258405efe91fc9caf9e7ea592834445df2e78add9553a921c76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://592dbdb2905039d89bdf741dbd0fd1e1fbd15acbabecdad6f07c758657fa610a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://526c2b9214d19b9ebe7a7078fd0459642a53e732d034602b2b8d2b8a33f4d18b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"message\\\":\\\"workPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1211 08:11:42.548303 6122 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1211 08:11:42.548667 6122 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1211 08:11:42.548830 6122 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1211 08:11:42.548871 6122 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1211 08:11:42.548960 6122 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1211 08:11:42.549108 6122 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1211 08:11:42.549161 6122 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1211 08:11:42.550530 6122 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:36Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb80d9928e8349fc5963227d135355ad7ba820d4fd89877af601d005ca03334b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17cbcb49c06f3a9059cd5a0f05eedb8c785da30a3e502e31bae2f57b3486cb14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17cbcb49c06f3a9059cd5a0f05eedb8c785da30a3e502e31bae2f57b3486cb14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xvnqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:43Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:43 crc kubenswrapper[4860]: I1211 08:11:43.983514 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gsx9l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aba5b6b9-f88d-4763-9562-6b4dec606df9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://735ef012c523ddc952d46a431f27b2582a2c301d9e9da0c697a20ff6e2fb5d86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6zw9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gsx9l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:43Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:44 crc kubenswrapper[4860]: I1211 08:11:44.006148 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9872be1e-b995-4a00-904b-875dff96ad56\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f74a32a8fefea704cc73ffd1114083dab68aa44dc33f8d9d432d2d24216546c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbe00b46063d3968d59c393b87178e135f982abb7d0f32fa798b212d4b83df6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eedc9988e1a4a3c8cae730dc4a352103a6d665535b825e157d0f87c331b137c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2137fee4b7abc9765f9517d7b3b6ea796601a31d03151bf167d17136ced8b25a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f683e48f3577a68862eb6302383538376e41f49b20b1e0ea3f1b3b143f24872a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5de794187456b3576b969069a3b839d75c7918418a97fa35dff8bb40bb76e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5de794187456b3576b969069a3b839d75c7918418a97fa35dff8bb40bb76e39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08ce42b974fba00c409c0af971f0de351e50f18b2f06f8e722150da3239bb908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08ce42b974fba00c409c0af971f0de351e50f18b2f06f8e722150da3239bb908\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cfac6be03e3cfa0b0f3450b48b67957f2d5172b9c65e27f14c2a260bda5d423f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cfac6be03e3cfa0b0f3450b48b67957f2d5172b9c65e27f14c2a260bda5d423f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:44Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:44 crc kubenswrapper[4860]: I1211 08:11:44.019963 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-pznwl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5310c1fc-66c6-40aa-b1bf-5a59a2410f9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2d6ff00ce322a28607bc3f6db05c5e4fb14acb6777ea315786ef53963ca504a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hkt48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-pznwl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:44Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:44 crc kubenswrapper[4860]: I1211 08:11:44.030923 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-854v6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e6529ba7-2341-43ac-a187-95ab530fe775\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8986538a421f0184f52f288b0e171da695591caf2b10c67ad2dce1073a6e6016\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kh22b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0d4e3f761ad15ae1424204e40b9d3f60188ab4a9047bd4b5ac6a8a8ca1fb385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kh22b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-854v6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:44Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:44 crc kubenswrapper[4860]: I1211 08:11:44.043893 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:44Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:44 crc kubenswrapper[4860]: I1211 08:11:44.049052 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:44 crc kubenswrapper[4860]: I1211 08:11:44.049086 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:44 crc kubenswrapper[4860]: I1211 08:11:44.049098 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:44 crc kubenswrapper[4860]: I1211 08:11:44.049114 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:44 crc kubenswrapper[4860]: I1211 08:11:44.049126 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:44Z","lastTransitionTime":"2025-12-11T08:11:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:44 crc kubenswrapper[4860]: I1211 08:11:44.061576 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d8c2f28628dc2f1ffbdd98a55670e22cf6cdc8cf2c6442e9d1f8e5466b2f12e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://074ea63beac52bd196507ee7cfb1c3d7fc2d37d5ec6faa4d47dde62b922e1c43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:44Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:44 crc kubenswrapper[4860]: I1211 08:11:44.076381 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:44Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:44 crc kubenswrapper[4860]: I1211 08:11:44.089010 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25a6a154da2b74952951ae94efc914ba078ea4bbe5893360b3c84d46be6bf973\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:44Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:44 crc kubenswrapper[4860]: I1211 08:11:44.099208 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6xq7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a75ce3c-5e1e-4b9f-bbde-7a3b0cce5eae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe8d4d675567b41728e476351811a93c47b19868ccbadab8c613a4eb8eca8562\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5c6nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6xq7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:44Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:44 crc kubenswrapper[4860]: I1211 08:11:44.112219 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ec3758-85f1-4a22-a4ec-c805472607a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e831db1a15976d3967d56479689572c26f2599a766baa0402897a3008428c56f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a498aa90f914e1f0370df01dc5ef174df113a2a16d86b2e3f24b19a1b99b4c0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8caebd2f4dfcbb89c597029a07c335171889761d2d200206987537fb0a228db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34e91f615b76445ff3241a304bb2964fde59de7e8a8dfff6841b2e628a31e86d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2090cab02ad74ddfd6bb57ff50eba35d18f56b0cde1c230fe011c01994b693bc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 08:11:21.171011 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 08:11:21.177674 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-644352688/tls.crt::/tmp/serving-cert-644352688/tls.key\\\\\\\"\\\\nI1211 08:11:26.508606 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 08:11:26.512741 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 08:11:26.512778 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 08:11:26.512839 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 08:11:26.512858 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 08:11:26.521684 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1211 08:11:26.521700 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1211 08:11:26.521722 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:11:26.521732 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:11:26.521738 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 08:11:26.521746 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 08:11:26.521749 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 08:11:26.521753 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1211 08:11:26.524999 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2274fb298ca7d2b37b060b4dbf9b166d39ba77a108b08edeeb4d67196e72cb3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf0aff8a6ab816b6617b1feeca9146d41bb5e8808bf1db96f8d493500c135d85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf0aff8a6ab816b6617b1feeca9146d41bb5e8808bf1db96f8d493500c135d85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:44Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:44 crc kubenswrapper[4860]: I1211 08:11:44.123079 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:44Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:44 crc kubenswrapper[4860]: I1211 08:11:44.133112 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a41f35db087418d1be16953f10377042abd3f4488e5ac1bc29ca5dcfc00a37e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:44Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:44 crc kubenswrapper[4860]: I1211 08:11:44.141100 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gjhlj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5cfd640a-f3d2-4b28-87e1-5f49f6341e21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knq7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knq7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:42Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gjhlj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:44Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:44 crc kubenswrapper[4860]: I1211 08:11:44.151378 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:44 crc kubenswrapper[4860]: I1211 08:11:44.151403 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:44 crc kubenswrapper[4860]: I1211 08:11:44.151411 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:44 crc kubenswrapper[4860]: I1211 08:11:44.151425 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:44 crc kubenswrapper[4860]: I1211 08:11:44.151436 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:44Z","lastTransitionTime":"2025-12-11T08:11:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:44 crc kubenswrapper[4860]: I1211 08:11:44.193227 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5cfd640a-f3d2-4b28-87e1-5f49f6341e21-metrics-certs\") pod \"network-metrics-daemon-gjhlj\" (UID: \"5cfd640a-f3d2-4b28-87e1-5f49f6341e21\") " pod="openshift-multus/network-metrics-daemon-gjhlj" Dec 11 08:11:44 crc kubenswrapper[4860]: E1211 08:11:44.193339 4860 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 11 08:11:44 crc kubenswrapper[4860]: E1211 08:11:44.193389 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5cfd640a-f3d2-4b28-87e1-5f49f6341e21-metrics-certs podName:5cfd640a-f3d2-4b28-87e1-5f49f6341e21 nodeName:}" failed. No retries permitted until 2025-12-11 08:11:46.19337535 +0000 UTC m=+38.921894405 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5cfd640a-f3d2-4b28-87e1-5f49f6341e21-metrics-certs") pod "network-metrics-daemon-gjhlj" (UID: "5cfd640a-f3d2-4b28-87e1-5f49f6341e21") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 11 08:11:44 crc kubenswrapper[4860]: I1211 08:11:44.254246 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:44 crc kubenswrapper[4860]: I1211 08:11:44.254308 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:44 crc kubenswrapper[4860]: I1211 08:11:44.254324 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:44 crc kubenswrapper[4860]: I1211 08:11:44.254350 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:44 crc kubenswrapper[4860]: I1211 08:11:44.254372 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:44Z","lastTransitionTime":"2025-12-11T08:11:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:44 crc kubenswrapper[4860]: I1211 08:11:44.357894 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:44 crc kubenswrapper[4860]: I1211 08:11:44.357941 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:44 crc kubenswrapper[4860]: I1211 08:11:44.357972 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:44 crc kubenswrapper[4860]: I1211 08:11:44.357989 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:44 crc kubenswrapper[4860]: I1211 08:11:44.358014 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:44Z","lastTransitionTime":"2025-12-11T08:11:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:44 crc kubenswrapper[4860]: I1211 08:11:44.461816 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:44 crc kubenswrapper[4860]: I1211 08:11:44.461906 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:44 crc kubenswrapper[4860]: I1211 08:11:44.461927 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:44 crc kubenswrapper[4860]: I1211 08:11:44.461958 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:44 crc kubenswrapper[4860]: I1211 08:11:44.461981 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:44Z","lastTransitionTime":"2025-12-11T08:11:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:44 crc kubenswrapper[4860]: I1211 08:11:44.565504 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:44 crc kubenswrapper[4860]: I1211 08:11:44.565560 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:44 crc kubenswrapper[4860]: I1211 08:11:44.565571 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:44 crc kubenswrapper[4860]: I1211 08:11:44.565589 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:44 crc kubenswrapper[4860]: I1211 08:11:44.565601 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:44Z","lastTransitionTime":"2025-12-11T08:11:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:44 crc kubenswrapper[4860]: I1211 08:11:44.577875 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:11:44 crc kubenswrapper[4860]: I1211 08:11:44.577919 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gjhlj" Dec 11 08:11:44 crc kubenswrapper[4860]: I1211 08:11:44.577893 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:11:44 crc kubenswrapper[4860]: E1211 08:11:44.578007 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:11:44 crc kubenswrapper[4860]: E1211 08:11:44.578075 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gjhlj" podUID="5cfd640a-f3d2-4b28-87e1-5f49f6341e21" Dec 11 08:11:44 crc kubenswrapper[4860]: I1211 08:11:44.577946 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:11:44 crc kubenswrapper[4860]: E1211 08:11:44.578162 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:11:44 crc kubenswrapper[4860]: E1211 08:11:44.578324 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:11:44 crc kubenswrapper[4860]: I1211 08:11:44.668738 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:44 crc kubenswrapper[4860]: I1211 08:11:44.668843 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:44 crc kubenswrapper[4860]: I1211 08:11:44.668875 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:44 crc kubenswrapper[4860]: I1211 08:11:44.668905 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:44 crc kubenswrapper[4860]: I1211 08:11:44.668925 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:44Z","lastTransitionTime":"2025-12-11T08:11:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:44 crc kubenswrapper[4860]: I1211 08:11:44.773287 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:44 crc kubenswrapper[4860]: I1211 08:11:44.773616 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:44 crc kubenswrapper[4860]: I1211 08:11:44.773711 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:44 crc kubenswrapper[4860]: I1211 08:11:44.773743 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:44 crc kubenswrapper[4860]: I1211 08:11:44.773765 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:44Z","lastTransitionTime":"2025-12-11T08:11:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:44 crc kubenswrapper[4860]: I1211 08:11:44.876314 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:44 crc kubenswrapper[4860]: I1211 08:11:44.876381 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:44 crc kubenswrapper[4860]: I1211 08:11:44.876402 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:44 crc kubenswrapper[4860]: I1211 08:11:44.876428 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:44 crc kubenswrapper[4860]: I1211 08:11:44.876447 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:44Z","lastTransitionTime":"2025-12-11T08:11:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:44 crc kubenswrapper[4860]: I1211 08:11:44.913022 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-xvnqp_0589a204-c98c-417d-8256-bfe3bf747660/ovnkube-controller/1.log" Dec 11 08:11:44 crc kubenswrapper[4860]: I1211 08:11:44.913698 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-xvnqp_0589a204-c98c-417d-8256-bfe3bf747660/ovnkube-controller/0.log" Dec 11 08:11:44 crc kubenswrapper[4860]: I1211 08:11:44.916050 4860 generic.go:334] "Generic (PLEG): container finished" podID="0589a204-c98c-417d-8256-bfe3bf747660" containerID="592dbdb2905039d89bdf741dbd0fd1e1fbd15acbabecdad6f07c758657fa610a" exitCode=1 Dec 11 08:11:44 crc kubenswrapper[4860]: I1211 08:11:44.916102 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" event={"ID":"0589a204-c98c-417d-8256-bfe3bf747660","Type":"ContainerDied","Data":"592dbdb2905039d89bdf741dbd0fd1e1fbd15acbabecdad6f07c758657fa610a"} Dec 11 08:11:44 crc kubenswrapper[4860]: I1211 08:11:44.916147 4860 scope.go:117] "RemoveContainer" containerID="526c2b9214d19b9ebe7a7078fd0459642a53e732d034602b2b8d2b8a33f4d18b" Dec 11 08:11:44 crc kubenswrapper[4860]: I1211 08:11:44.917397 4860 scope.go:117] "RemoveContainer" containerID="592dbdb2905039d89bdf741dbd0fd1e1fbd15acbabecdad6f07c758657fa610a" Dec 11 08:11:44 crc kubenswrapper[4860]: E1211 08:11:44.917706 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-xvnqp_openshift-ovn-kubernetes(0589a204-c98c-417d-8256-bfe3bf747660)\"" pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" podUID="0589a204-c98c-417d-8256-bfe3bf747660" Dec 11 08:11:44 crc kubenswrapper[4860]: I1211 08:11:44.934878 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31c30642-6e60-41b4-a477-0d802424e0aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79969402f5c33852a26fc839f37a45febf4f2377d77823284fbb01769e5585fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2d4d5de7106175087d91acb1f0c7552014f0c08bfcd65cbbf184d3f8d7826ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-99qgp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:44Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:44 crc kubenswrapper[4860]: I1211 08:11:44.958927 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-26nc6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1faa733-f408-41fd-b7aa-0cd86283976a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fce5b642b843ed30823c94e0b0c3d8bb1e5d2d93d691f7b8ef901ece03b0ac6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ac45db4f858a1c18a6b87472fc1cea2300f5845f1ad62c22260374995548a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ac45db4f858a1c18a6b87472fc1cea2300f5845f1ad62c22260374995548a0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b35296a5eb17dd2bb5e72c86c1cafdb30b7712cb330264f935d3b4fff05cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c6b35296a5eb17dd2bb5e72c86c1cafdb30b7712cb330264f935d3b4fff05cd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b9e82d6d3a247309ef5bdbf5b3a86ec083add2a138b0b237801e77b09ee96a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b9e82d6d3a247309ef5bdbf5b3a86ec083add2a138b0b237801e77b09ee96a1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://908ee695e78145ef4375c4097d9f1e772502922d02a2fcb4566172a6ad9caaf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://908ee695e78145ef4375c4097d9f1e772502922d02a2fcb4566172a6ad9caaf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afb0a097f3556a8f9ed7cf8f762cb99c082fe0a36b8d8d3038bf188b08fe9a84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afb0a097f3556a8f9ed7cf8f762cb99c082fe0a36b8d8d3038bf188b08fe9a84\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce18f4f06db091e658684fa0bbc858e19b9f03010acf13b3ffccb1a6e3f88021\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce18f4f06db091e658684fa0bbc858e19b9f03010acf13b3ffccb1a6e3f88021\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-26nc6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:44Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:44 crc kubenswrapper[4860]: I1211 08:11:44.960931 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" Dec 11 08:11:44 crc kubenswrapper[4860]: I1211 08:11:44.979708 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:44 crc kubenswrapper[4860]: I1211 08:11:44.979773 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:44 crc kubenswrapper[4860]: I1211 08:11:44.979837 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:44 crc kubenswrapper[4860]: I1211 08:11:44.979865 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:44 crc kubenswrapper[4860]: I1211 08:11:44.979888 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:44Z","lastTransitionTime":"2025-12-11T08:11:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:44 crc kubenswrapper[4860]: I1211 08:11:44.979832 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0589a204-c98c-417d-8256-bfe3bf747660\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3333b4346f59b04017e32d61f457ba7ff457166731128f339d7987ae6b56b06d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52bb8ade314afb9494f4e026d0af2aea7b1970d64dc66102f55e257ae4678154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dfcc4370edaae36a5a32b0a9c29290c50f8e4f427e4e147dc3c779beaa8b780\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d856ae95e9077524c518d82d10fb6b93b7fbcaaf2b720ffafe6f7fae58bd7caa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1426084ac580895c42354630910dd8c8e3a48ddd25f22abce1d28dc3de332a63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9984d5d242bb258405efe91fc9caf9e7ea592834445df2e78add9553a921c76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://592dbdb2905039d89bdf741dbd0fd1e1fbd15acbabecdad6f07c758657fa610a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://526c2b9214d19b9ebe7a7078fd0459642a53e732d034602b2b8d2b8a33f4d18b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"message\\\":\\\"workPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1211 08:11:42.548303 6122 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1211 08:11:42.548667 6122 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1211 08:11:42.548830 6122 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1211 08:11:42.548871 6122 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1211 08:11:42.548960 6122 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1211 08:11:42.549108 6122 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1211 08:11:42.549161 6122 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1211 08:11:42.550530 6122 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:36Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://592dbdb2905039d89bdf741dbd0fd1e1fbd15acbabecdad6f07c758657fa610a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-11T08:11:44Z\\\",\\\"message\\\":\\\"es.lbConfig(nil)\\\\nI1211 08:11:44.095388 6339 services_controller.go:356] Processing sync for service openshift-ingress-canary/ingress-canary for network=default\\\\nI1211 08:11:44.095870 6339 services_controller.go:445] Built service openshift-kube-controller-manager/kube-controller-manager LB template configs for network=default: []services.lbConfig(nil)\\\\nF1211 08:11:44.095867 6339 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:44Z is after 2025-08-24T17:21:41Z]\\\\nI1211 08:11:44.095898 6339 services_controller.go:451] Built service openshift-kube-controller-manager/kube-controller-manager cluster-wide LB for network\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb80d9928e8349fc5963227d135355ad7ba820d4fd89877af601d005ca03334b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17cbcb49c06f3a9059cd5a0f05eedb8c785da30a3e502e31bae2f57b3486cb14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17cbcb49c06f3a9059cd5a0f05eedb8c785da30a3e502e31bae2f57b3486cb14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xvnqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:44Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:44 crc kubenswrapper[4860]: I1211 08:11:44.997864 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gsx9l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aba5b6b9-f88d-4763-9562-6b4dec606df9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://735ef012c523ddc952d46a431f27b2582a2c301d9e9da0c697a20ff6e2fb5d86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6zw9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gsx9l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:44Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:45 crc kubenswrapper[4860]: I1211 08:11:45.012505 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9aa0aed-62b3-4ea2-82ae-9a653c1101a5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ad4ba1b769c6cb2b34073798f8e3c25eb0177e6d3059484bf365e6981cab710\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://077aa1dd2fe5189ff60a0f2365e47a6a545cca0de7aa6379370908db4ea5838a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d7c919899b74a2b0fcd9221f3f769a9d013267354b200a282669d342d6413d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5483384f1d2ebc1fefa9706c9e67a3dc5a00ff643f803e3f2707c252e27ad94\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:45Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:45 crc kubenswrapper[4860]: I1211 08:11:45.024383 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-854v6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e6529ba7-2341-43ac-a187-95ab530fe775\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8986538a421f0184f52f288b0e171da695591caf2b10c67ad2dce1073a6e6016\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kh22b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0d4e3f761ad15ae1424204e40b9d3f60188ab4a9047bd4b5ac6a8a8ca1fb385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kh22b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-854v6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:45Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:45 crc kubenswrapper[4860]: I1211 08:11:45.048763 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9872be1e-b995-4a00-904b-875dff96ad56\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f74a32a8fefea704cc73ffd1114083dab68aa44dc33f8d9d432d2d24216546c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbe00b46063d3968d59c393b87178e135f982abb7d0f32fa798b212d4b83df6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eedc9988e1a4a3c8cae730dc4a352103a6d665535b825e157d0f87c331b137c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2137fee4b7abc9765f9517d7b3b6ea796601a31d03151bf167d17136ced8b25a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f683e48f3577a68862eb6302383538376e41f49b20b1e0ea3f1b3b143f24872a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5de794187456b3576b969069a3b839d75c7918418a97fa35dff8bb40bb76e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5de794187456b3576b969069a3b839d75c7918418a97fa35dff8bb40bb76e39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08ce42b974fba00c409c0af971f0de351e50f18b2f06f8e722150da3239bb908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08ce42b974fba00c409c0af971f0de351e50f18b2f06f8e722150da3239bb908\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cfac6be03e3cfa0b0f3450b48b67957f2d5172b9c65e27f14c2a260bda5d423f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cfac6be03e3cfa0b0f3450b48b67957f2d5172b9c65e27f14c2a260bda5d423f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:45Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:45 crc kubenswrapper[4860]: I1211 08:11:45.061902 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-pznwl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5310c1fc-66c6-40aa-b1bf-5a59a2410f9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2d6ff00ce322a28607bc3f6db05c5e4fb14acb6777ea315786ef53963ca504a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hkt48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-pznwl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:45Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:45 crc kubenswrapper[4860]: I1211 08:11:45.079822 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d8c2f28628dc2f1ffbdd98a55670e22cf6cdc8cf2c6442e9d1f8e5466b2f12e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://074ea63beac52bd196507ee7cfb1c3d7fc2d37d5ec6faa4d47dde62b922e1c43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:45Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:45 crc kubenswrapper[4860]: I1211 08:11:45.082792 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:45 crc kubenswrapper[4860]: I1211 08:11:45.082839 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:45 crc kubenswrapper[4860]: I1211 08:11:45.082876 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:45 crc kubenswrapper[4860]: I1211 08:11:45.082896 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:45 crc kubenswrapper[4860]: I1211 08:11:45.082907 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:45Z","lastTransitionTime":"2025-12-11T08:11:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:45 crc kubenswrapper[4860]: I1211 08:11:45.091803 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:45Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:45 crc kubenswrapper[4860]: I1211 08:11:45.102782 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25a6a154da2b74952951ae94efc914ba078ea4bbe5893360b3c84d46be6bf973\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:45Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:45 crc kubenswrapper[4860]: I1211 08:11:45.111680 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6xq7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a75ce3c-5e1e-4b9f-bbde-7a3b0cce5eae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe8d4d675567b41728e476351811a93c47b19868ccbadab8c613a4eb8eca8562\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5c6nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6xq7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:45Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:45 crc kubenswrapper[4860]: I1211 08:11:45.128306 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:45Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:45 crc kubenswrapper[4860]: I1211 08:11:45.147353 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a41f35db087418d1be16953f10377042abd3f4488e5ac1bc29ca5dcfc00a37e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:45Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:45 crc kubenswrapper[4860]: I1211 08:11:45.161621 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gjhlj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5cfd640a-f3d2-4b28-87e1-5f49f6341e21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knq7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knq7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:42Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gjhlj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:45Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:45 crc kubenswrapper[4860]: I1211 08:11:45.184235 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ec3758-85f1-4a22-a4ec-c805472607a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e831db1a15976d3967d56479689572c26f2599a766baa0402897a3008428c56f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a498aa90f914e1f0370df01dc5ef174df113a2a16d86b2e3f24b19a1b99b4c0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8caebd2f4dfcbb89c597029a07c335171889761d2d200206987537fb0a228db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34e91f615b76445ff3241a304bb2964fde59de7e8a8dfff6841b2e628a31e86d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2090cab02ad74ddfd6bb57ff50eba35d18f56b0cde1c230fe011c01994b693bc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 08:11:21.171011 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 08:11:21.177674 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-644352688/tls.crt::/tmp/serving-cert-644352688/tls.key\\\\\\\"\\\\nI1211 08:11:26.508606 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 08:11:26.512741 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 08:11:26.512778 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 08:11:26.512839 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 08:11:26.512858 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 08:11:26.521684 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1211 08:11:26.521700 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1211 08:11:26.521722 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:11:26.521732 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:11:26.521738 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 08:11:26.521746 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 08:11:26.521749 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 08:11:26.521753 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1211 08:11:26.524999 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2274fb298ca7d2b37b060b4dbf9b166d39ba77a108b08edeeb4d67196e72cb3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf0aff8a6ab816b6617b1feeca9146d41bb5e8808bf1db96f8d493500c135d85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf0aff8a6ab816b6617b1feeca9146d41bb5e8808bf1db96f8d493500c135d85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:45Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:45 crc kubenswrapper[4860]: I1211 08:11:45.188441 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:45 crc kubenswrapper[4860]: I1211 08:11:45.188476 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:45 crc kubenswrapper[4860]: I1211 08:11:45.188485 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:45 crc kubenswrapper[4860]: I1211 08:11:45.188500 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:45 crc kubenswrapper[4860]: I1211 08:11:45.188518 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:45Z","lastTransitionTime":"2025-12-11T08:11:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:45 crc kubenswrapper[4860]: I1211 08:11:45.198275 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:45Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:45 crc kubenswrapper[4860]: I1211 08:11:45.291561 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:45 crc kubenswrapper[4860]: I1211 08:11:45.291606 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:45 crc kubenswrapper[4860]: I1211 08:11:45.291615 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:45 crc kubenswrapper[4860]: I1211 08:11:45.291630 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:45 crc kubenswrapper[4860]: I1211 08:11:45.291657 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:45Z","lastTransitionTime":"2025-12-11T08:11:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:45 crc kubenswrapper[4860]: I1211 08:11:45.396558 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:45 crc kubenswrapper[4860]: I1211 08:11:45.396619 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:45 crc kubenswrapper[4860]: I1211 08:11:45.396685 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:45 crc kubenswrapper[4860]: I1211 08:11:45.396719 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:45 crc kubenswrapper[4860]: I1211 08:11:45.396745 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:45Z","lastTransitionTime":"2025-12-11T08:11:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:45 crc kubenswrapper[4860]: I1211 08:11:45.499800 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:45 crc kubenswrapper[4860]: I1211 08:11:45.499842 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:45 crc kubenswrapper[4860]: I1211 08:11:45.499858 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:45 crc kubenswrapper[4860]: I1211 08:11:45.499879 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:45 crc kubenswrapper[4860]: I1211 08:11:45.499900 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:45Z","lastTransitionTime":"2025-12-11T08:11:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:45 crc kubenswrapper[4860]: I1211 08:11:45.602581 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:45 crc kubenswrapper[4860]: I1211 08:11:45.602638 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:45 crc kubenswrapper[4860]: I1211 08:11:45.602670 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:45 crc kubenswrapper[4860]: I1211 08:11:45.602689 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:45 crc kubenswrapper[4860]: I1211 08:11:45.602704 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:45Z","lastTransitionTime":"2025-12-11T08:11:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:45 crc kubenswrapper[4860]: I1211 08:11:45.706734 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:45 crc kubenswrapper[4860]: I1211 08:11:45.706832 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:45 crc kubenswrapper[4860]: I1211 08:11:45.706874 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:45 crc kubenswrapper[4860]: I1211 08:11:45.706913 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:45 crc kubenswrapper[4860]: I1211 08:11:45.706937 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:45Z","lastTransitionTime":"2025-12-11T08:11:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:45 crc kubenswrapper[4860]: I1211 08:11:45.809556 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:45 crc kubenswrapper[4860]: I1211 08:11:45.809623 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:45 crc kubenswrapper[4860]: I1211 08:11:45.809669 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:45 crc kubenswrapper[4860]: I1211 08:11:45.809695 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:45 crc kubenswrapper[4860]: I1211 08:11:45.809712 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:45Z","lastTransitionTime":"2025-12-11T08:11:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:45 crc kubenswrapper[4860]: I1211 08:11:45.912434 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:45 crc kubenswrapper[4860]: I1211 08:11:45.912505 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:45 crc kubenswrapper[4860]: I1211 08:11:45.912530 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:45 crc kubenswrapper[4860]: I1211 08:11:45.912559 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:45 crc kubenswrapper[4860]: I1211 08:11:45.912579 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:45Z","lastTransitionTime":"2025-12-11T08:11:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:45 crc kubenswrapper[4860]: I1211 08:11:45.922045 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-xvnqp_0589a204-c98c-417d-8256-bfe3bf747660/ovnkube-controller/1.log" Dec 11 08:11:45 crc kubenswrapper[4860]: I1211 08:11:45.927511 4860 scope.go:117] "RemoveContainer" containerID="592dbdb2905039d89bdf741dbd0fd1e1fbd15acbabecdad6f07c758657fa610a" Dec 11 08:11:45 crc kubenswrapper[4860]: E1211 08:11:45.927916 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-xvnqp_openshift-ovn-kubernetes(0589a204-c98c-417d-8256-bfe3bf747660)\"" pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" podUID="0589a204-c98c-417d-8256-bfe3bf747660" Dec 11 08:11:45 crc kubenswrapper[4860]: I1211 08:11:45.949981 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ec3758-85f1-4a22-a4ec-c805472607a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e831db1a15976d3967d56479689572c26f2599a766baa0402897a3008428c56f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a498aa90f914e1f0370df01dc5ef174df113a2a16d86b2e3f24b19a1b99b4c0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8caebd2f4dfcbb89c597029a07c335171889761d2d200206987537fb0a228db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34e91f615b76445ff3241a304bb2964fde59de7e8a8dfff6841b2e628a31e86d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2090cab02ad74ddfd6bb57ff50eba35d18f56b0cde1c230fe011c01994b693bc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 08:11:21.171011 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 08:11:21.177674 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-644352688/tls.crt::/tmp/serving-cert-644352688/tls.key\\\\\\\"\\\\nI1211 08:11:26.508606 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 08:11:26.512741 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 08:11:26.512778 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 08:11:26.512839 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 08:11:26.512858 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 08:11:26.521684 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1211 08:11:26.521700 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1211 08:11:26.521722 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:11:26.521732 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:11:26.521738 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 08:11:26.521746 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 08:11:26.521749 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 08:11:26.521753 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1211 08:11:26.524999 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2274fb298ca7d2b37b060b4dbf9b166d39ba77a108b08edeeb4d67196e72cb3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf0aff8a6ab816b6617b1feeca9146d41bb5e8808bf1db96f8d493500c135d85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf0aff8a6ab816b6617b1feeca9146d41bb5e8808bf1db96f8d493500c135d85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:45Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:45 crc kubenswrapper[4860]: I1211 08:11:45.970719 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:45Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:45 crc kubenswrapper[4860]: I1211 08:11:45.993746 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a41f35db087418d1be16953f10377042abd3f4488e5ac1bc29ca5dcfc00a37e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:45Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:46 crc kubenswrapper[4860]: I1211 08:11:46.008102 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gjhlj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5cfd640a-f3d2-4b28-87e1-5f49f6341e21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knq7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knq7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:42Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gjhlj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:46Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:46 crc kubenswrapper[4860]: I1211 08:11:46.016165 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:46 crc kubenswrapper[4860]: I1211 08:11:46.016235 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:46 crc kubenswrapper[4860]: I1211 08:11:46.016255 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:46 crc kubenswrapper[4860]: I1211 08:11:46.016281 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:46 crc kubenswrapper[4860]: I1211 08:11:46.016299 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:46Z","lastTransitionTime":"2025-12-11T08:11:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:46 crc kubenswrapper[4860]: I1211 08:11:46.024557 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9aa0aed-62b3-4ea2-82ae-9a653c1101a5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ad4ba1b769c6cb2b34073798f8e3c25eb0177e6d3059484bf365e6981cab710\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://077aa1dd2fe5189ff60a0f2365e47a6a545cca0de7aa6379370908db4ea5838a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d7c919899b74a2b0fcd9221f3f769a9d013267354b200a282669d342d6413d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5483384f1d2ebc1fefa9706c9e67a3dc5a00ff643f803e3f2707c252e27ad94\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:46Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:46 crc kubenswrapper[4860]: I1211 08:11:46.040876 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31c30642-6e60-41b4-a477-0d802424e0aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79969402f5c33852a26fc839f37a45febf4f2377d77823284fbb01769e5585fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2d4d5de7106175087d91acb1f0c7552014f0c08bfcd65cbbf184d3f8d7826ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-99qgp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:46Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:46 crc kubenswrapper[4860]: I1211 08:11:46.056406 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-26nc6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1faa733-f408-41fd-b7aa-0cd86283976a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fce5b642b843ed30823c94e0b0c3d8bb1e5d2d93d691f7b8ef901ece03b0ac6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ac45db4f858a1c18a6b87472fc1cea2300f5845f1ad62c22260374995548a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ac45db4f858a1c18a6b87472fc1cea2300f5845f1ad62c22260374995548a0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b35296a5eb17dd2bb5e72c86c1cafdb30b7712cb330264f935d3b4fff05cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c6b35296a5eb17dd2bb5e72c86c1cafdb30b7712cb330264f935d3b4fff05cd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b9e82d6d3a247309ef5bdbf5b3a86ec083add2a138b0b237801e77b09ee96a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b9e82d6d3a247309ef5bdbf5b3a86ec083add2a138b0b237801e77b09ee96a1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://908ee695e78145ef4375c4097d9f1e772502922d02a2fcb4566172a6ad9caaf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://908ee695e78145ef4375c4097d9f1e772502922d02a2fcb4566172a6ad9caaf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afb0a097f3556a8f9ed7cf8f762cb99c082fe0a36b8d8d3038bf188b08fe9a84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afb0a097f3556a8f9ed7cf8f762cb99c082fe0a36b8d8d3038bf188b08fe9a84\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce18f4f06db091e658684fa0bbc858e19b9f03010acf13b3ffccb1a6e3f88021\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce18f4f06db091e658684fa0bbc858e19b9f03010acf13b3ffccb1a6e3f88021\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-26nc6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:46Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:46 crc kubenswrapper[4860]: I1211 08:11:46.076002 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0589a204-c98c-417d-8256-bfe3bf747660\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3333b4346f59b04017e32d61f457ba7ff457166731128f339d7987ae6b56b06d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52bb8ade314afb9494f4e026d0af2aea7b1970d64dc66102f55e257ae4678154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dfcc4370edaae36a5a32b0a9c29290c50f8e4f427e4e147dc3c779beaa8b780\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d856ae95e9077524c518d82d10fb6b93b7fbcaaf2b720ffafe6f7fae58bd7caa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1426084ac580895c42354630910dd8c8e3a48ddd25f22abce1d28dc3de332a63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9984d5d242bb258405efe91fc9caf9e7ea592834445df2e78add9553a921c76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://592dbdb2905039d89bdf741dbd0fd1e1fbd15acbabecdad6f07c758657fa610a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://592dbdb2905039d89bdf741dbd0fd1e1fbd15acbabecdad6f07c758657fa610a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-11T08:11:44Z\\\",\\\"message\\\":\\\"es.lbConfig(nil)\\\\nI1211 08:11:44.095388 6339 services_controller.go:356] Processing sync for service openshift-ingress-canary/ingress-canary for network=default\\\\nI1211 08:11:44.095870 6339 services_controller.go:445] Built service openshift-kube-controller-manager/kube-controller-manager LB template configs for network=default: []services.lbConfig(nil)\\\\nF1211 08:11:44.095867 6339 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:44Z is after 2025-08-24T17:21:41Z]\\\\nI1211 08:11:44.095898 6339 services_controller.go:451] Built service openshift-kube-controller-manager/kube-controller-manager cluster-wide LB for network\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:42Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-xvnqp_openshift-ovn-kubernetes(0589a204-c98c-417d-8256-bfe3bf747660)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb80d9928e8349fc5963227d135355ad7ba820d4fd89877af601d005ca03334b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17cbcb49c06f3a9059cd5a0f05eedb8c785da30a3e502e31bae2f57b3486cb14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17cbcb49c06f3a9059cd5a0f05eedb8c785da30a3e502e31bae2f57b3486cb14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xvnqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:46Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:46 crc kubenswrapper[4860]: I1211 08:11:46.088673 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gsx9l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aba5b6b9-f88d-4763-9562-6b4dec606df9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://735ef012c523ddc952d46a431f27b2582a2c301d9e9da0c697a20ff6e2fb5d86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6zw9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gsx9l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:46Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:46 crc kubenswrapper[4860]: I1211 08:11:46.111301 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9872be1e-b995-4a00-904b-875dff96ad56\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f74a32a8fefea704cc73ffd1114083dab68aa44dc33f8d9d432d2d24216546c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbe00b46063d3968d59c393b87178e135f982abb7d0f32fa798b212d4b83df6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eedc9988e1a4a3c8cae730dc4a352103a6d665535b825e157d0f87c331b137c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2137fee4b7abc9765f9517d7b3b6ea796601a31d03151bf167d17136ced8b25a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f683e48f3577a68862eb6302383538376e41f49b20b1e0ea3f1b3b143f24872a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5de794187456b3576b969069a3b839d75c7918418a97fa35dff8bb40bb76e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5de794187456b3576b969069a3b839d75c7918418a97fa35dff8bb40bb76e39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08ce42b974fba00c409c0af971f0de351e50f18b2f06f8e722150da3239bb908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08ce42b974fba00c409c0af971f0de351e50f18b2f06f8e722150da3239bb908\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cfac6be03e3cfa0b0f3450b48b67957f2d5172b9c65e27f14c2a260bda5d423f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cfac6be03e3cfa0b0f3450b48b67957f2d5172b9c65e27f14c2a260bda5d423f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:46Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:46 crc kubenswrapper[4860]: I1211 08:11:46.125418 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-pznwl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5310c1fc-66c6-40aa-b1bf-5a59a2410f9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2d6ff00ce322a28607bc3f6db05c5e4fb14acb6777ea315786ef53963ca504a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hkt48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-pznwl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:46Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:46 crc kubenswrapper[4860]: I1211 08:11:46.137404 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-854v6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e6529ba7-2341-43ac-a187-95ab530fe775\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8986538a421f0184f52f288b0e171da695591caf2b10c67ad2dce1073a6e6016\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kh22b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0d4e3f761ad15ae1424204e40b9d3f60188ab4a9047bd4b5ac6a8a8ca1fb385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kh22b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-854v6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:46Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:46 crc kubenswrapper[4860]: I1211 08:11:46.150941 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:46 crc kubenswrapper[4860]: I1211 08:11:46.150995 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:46 crc kubenswrapper[4860]: I1211 08:11:46.150949 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:46Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:46 crc kubenswrapper[4860]: I1211 08:11:46.151009 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:46 crc kubenswrapper[4860]: I1211 08:11:46.151213 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:46 crc kubenswrapper[4860]: I1211 08:11:46.151230 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:46Z","lastTransitionTime":"2025-12-11T08:11:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:46 crc kubenswrapper[4860]: I1211 08:11:46.167380 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d8c2f28628dc2f1ffbdd98a55670e22cf6cdc8cf2c6442e9d1f8e5466b2f12e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://074ea63beac52bd196507ee7cfb1c3d7fc2d37d5ec6faa4d47dde62b922e1c43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:46Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:46 crc kubenswrapper[4860]: I1211 08:11:46.183929 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:46Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:46 crc kubenswrapper[4860]: I1211 08:11:46.196821 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25a6a154da2b74952951ae94efc914ba078ea4bbe5893360b3c84d46be6bf973\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:46Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:46 crc kubenswrapper[4860]: I1211 08:11:46.206051 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6xq7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a75ce3c-5e1e-4b9f-bbde-7a3b0cce5eae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe8d4d675567b41728e476351811a93c47b19868ccbadab8c613a4eb8eca8562\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5c6nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6xq7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:46Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:46 crc kubenswrapper[4860]: I1211 08:11:46.212422 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5cfd640a-f3d2-4b28-87e1-5f49f6341e21-metrics-certs\") pod \"network-metrics-daemon-gjhlj\" (UID: \"5cfd640a-f3d2-4b28-87e1-5f49f6341e21\") " pod="openshift-multus/network-metrics-daemon-gjhlj" Dec 11 08:11:46 crc kubenswrapper[4860]: E1211 08:11:46.212542 4860 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 11 08:11:46 crc kubenswrapper[4860]: E1211 08:11:46.212592 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5cfd640a-f3d2-4b28-87e1-5f49f6341e21-metrics-certs podName:5cfd640a-f3d2-4b28-87e1-5f49f6341e21 nodeName:}" failed. No retries permitted until 2025-12-11 08:11:50.212577783 +0000 UTC m=+42.941096838 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5cfd640a-f3d2-4b28-87e1-5f49f6341e21-metrics-certs") pod "network-metrics-daemon-gjhlj" (UID: "5cfd640a-f3d2-4b28-87e1-5f49f6341e21") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 11 08:11:46 crc kubenswrapper[4860]: I1211 08:11:46.253389 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:46 crc kubenswrapper[4860]: I1211 08:11:46.253428 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:46 crc kubenswrapper[4860]: I1211 08:11:46.253439 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:46 crc kubenswrapper[4860]: I1211 08:11:46.253456 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:46 crc kubenswrapper[4860]: I1211 08:11:46.253467 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:46Z","lastTransitionTime":"2025-12-11T08:11:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:46 crc kubenswrapper[4860]: I1211 08:11:46.357170 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:46 crc kubenswrapper[4860]: I1211 08:11:46.357539 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:46 crc kubenswrapper[4860]: I1211 08:11:46.357815 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:46 crc kubenswrapper[4860]: I1211 08:11:46.358058 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:46 crc kubenswrapper[4860]: I1211 08:11:46.358209 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:46Z","lastTransitionTime":"2025-12-11T08:11:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:46 crc kubenswrapper[4860]: I1211 08:11:46.461957 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:46 crc kubenswrapper[4860]: I1211 08:11:46.462040 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:46 crc kubenswrapper[4860]: I1211 08:11:46.462060 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:46 crc kubenswrapper[4860]: I1211 08:11:46.462086 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:46 crc kubenswrapper[4860]: I1211 08:11:46.462105 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:46Z","lastTransitionTime":"2025-12-11T08:11:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:46 crc kubenswrapper[4860]: I1211 08:11:46.565541 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:46 crc kubenswrapper[4860]: I1211 08:11:46.565613 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:46 crc kubenswrapper[4860]: I1211 08:11:46.565700 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:46 crc kubenswrapper[4860]: I1211 08:11:46.565753 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:46 crc kubenswrapper[4860]: I1211 08:11:46.565772 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:46Z","lastTransitionTime":"2025-12-11T08:11:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:46 crc kubenswrapper[4860]: I1211 08:11:46.578808 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:11:46 crc kubenswrapper[4860]: I1211 08:11:46.578906 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:11:46 crc kubenswrapper[4860]: I1211 08:11:46.578932 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gjhlj" Dec 11 08:11:46 crc kubenswrapper[4860]: I1211 08:11:46.578833 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:11:46 crc kubenswrapper[4860]: E1211 08:11:46.579386 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:11:46 crc kubenswrapper[4860]: E1211 08:11:46.579526 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gjhlj" podUID="5cfd640a-f3d2-4b28-87e1-5f49f6341e21" Dec 11 08:11:46 crc kubenswrapper[4860]: E1211 08:11:46.579100 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:11:46 crc kubenswrapper[4860]: E1211 08:11:46.579735 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:11:46 crc kubenswrapper[4860]: I1211 08:11:46.668354 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:46 crc kubenswrapper[4860]: I1211 08:11:46.668480 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:46 crc kubenswrapper[4860]: I1211 08:11:46.668563 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:46 crc kubenswrapper[4860]: I1211 08:11:46.668597 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:46 crc kubenswrapper[4860]: I1211 08:11:46.668620 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:46Z","lastTransitionTime":"2025-12-11T08:11:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:46 crc kubenswrapper[4860]: I1211 08:11:46.771310 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:46 crc kubenswrapper[4860]: I1211 08:11:46.771628 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:46 crc kubenswrapper[4860]: I1211 08:11:46.771783 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:46 crc kubenswrapper[4860]: I1211 08:11:46.771942 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:46 crc kubenswrapper[4860]: I1211 08:11:46.772073 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:46Z","lastTransitionTime":"2025-12-11T08:11:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:46 crc kubenswrapper[4860]: I1211 08:11:46.874939 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:46 crc kubenswrapper[4860]: I1211 08:11:46.875003 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:46 crc kubenswrapper[4860]: I1211 08:11:46.875021 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:46 crc kubenswrapper[4860]: I1211 08:11:46.875046 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:46 crc kubenswrapper[4860]: I1211 08:11:46.875064 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:46Z","lastTransitionTime":"2025-12-11T08:11:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:46 crc kubenswrapper[4860]: I1211 08:11:46.978456 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:46 crc kubenswrapper[4860]: I1211 08:11:46.978524 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:46 crc kubenswrapper[4860]: I1211 08:11:46.978546 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:46 crc kubenswrapper[4860]: I1211 08:11:46.978574 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:46 crc kubenswrapper[4860]: I1211 08:11:46.978595 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:46Z","lastTransitionTime":"2025-12-11T08:11:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:47 crc kubenswrapper[4860]: I1211 08:11:47.081351 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:47 crc kubenswrapper[4860]: I1211 08:11:47.081436 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:47 crc kubenswrapper[4860]: I1211 08:11:47.081460 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:47 crc kubenswrapper[4860]: I1211 08:11:47.081489 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:47 crc kubenswrapper[4860]: I1211 08:11:47.081507 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:47Z","lastTransitionTime":"2025-12-11T08:11:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:47 crc kubenswrapper[4860]: I1211 08:11:47.184928 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:47 crc kubenswrapper[4860]: I1211 08:11:47.184993 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:47 crc kubenswrapper[4860]: I1211 08:11:47.185012 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:47 crc kubenswrapper[4860]: I1211 08:11:47.185047 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:47 crc kubenswrapper[4860]: I1211 08:11:47.185082 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:47Z","lastTransitionTime":"2025-12-11T08:11:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:47 crc kubenswrapper[4860]: I1211 08:11:47.288692 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:47 crc kubenswrapper[4860]: I1211 08:11:47.288814 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:47 crc kubenswrapper[4860]: I1211 08:11:47.288834 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:47 crc kubenswrapper[4860]: I1211 08:11:47.288859 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:47 crc kubenswrapper[4860]: I1211 08:11:47.288876 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:47Z","lastTransitionTime":"2025-12-11T08:11:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:47 crc kubenswrapper[4860]: I1211 08:11:47.391676 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:47 crc kubenswrapper[4860]: I1211 08:11:47.391736 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:47 crc kubenswrapper[4860]: I1211 08:11:47.391748 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:47 crc kubenswrapper[4860]: I1211 08:11:47.391767 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:47 crc kubenswrapper[4860]: I1211 08:11:47.391778 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:47Z","lastTransitionTime":"2025-12-11T08:11:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:47 crc kubenswrapper[4860]: I1211 08:11:47.495392 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:47 crc kubenswrapper[4860]: I1211 08:11:47.495466 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:47 crc kubenswrapper[4860]: I1211 08:11:47.495481 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:47 crc kubenswrapper[4860]: I1211 08:11:47.495510 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:47 crc kubenswrapper[4860]: I1211 08:11:47.495530 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:47Z","lastTransitionTime":"2025-12-11T08:11:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:47 crc kubenswrapper[4860]: I1211 08:11:47.597393 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31c30642-6e60-41b4-a477-0d802424e0aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79969402f5c33852a26fc839f37a45febf4f2377d77823284fbb01769e5585fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2d4d5de7106175087d91acb1f0c7552014f0c08bfcd65cbbf184d3f8d7826ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-99qgp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:47Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:47 crc kubenswrapper[4860]: I1211 08:11:47.599528 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:47 crc kubenswrapper[4860]: I1211 08:11:47.599578 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:47 crc kubenswrapper[4860]: I1211 08:11:47.599597 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:47 crc kubenswrapper[4860]: I1211 08:11:47.599623 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:47 crc kubenswrapper[4860]: I1211 08:11:47.599675 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:47Z","lastTransitionTime":"2025-12-11T08:11:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:47 crc kubenswrapper[4860]: I1211 08:11:47.621925 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-26nc6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1faa733-f408-41fd-b7aa-0cd86283976a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fce5b642b843ed30823c94e0b0c3d8bb1e5d2d93d691f7b8ef901ece03b0ac6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ac45db4f858a1c18a6b87472fc1cea2300f5845f1ad62c22260374995548a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ac45db4f858a1c18a6b87472fc1cea2300f5845f1ad62c22260374995548a0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b35296a5eb17dd2bb5e72c86c1cafdb30b7712cb330264f935d3b4fff05cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c6b35296a5eb17dd2bb5e72c86c1cafdb30b7712cb330264f935d3b4fff05cd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b9e82d6d3a247309ef5bdbf5b3a86ec083add2a138b0b237801e77b09ee96a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b9e82d6d3a247309ef5bdbf5b3a86ec083add2a138b0b237801e77b09ee96a1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://908ee695e78145ef4375c4097d9f1e772502922d02a2fcb4566172a6ad9caaf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://908ee695e78145ef4375c4097d9f1e772502922d02a2fcb4566172a6ad9caaf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afb0a097f3556a8f9ed7cf8f762cb99c082fe0a36b8d8d3038bf188b08fe9a84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afb0a097f3556a8f9ed7cf8f762cb99c082fe0a36b8d8d3038bf188b08fe9a84\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce18f4f06db091e658684fa0bbc858e19b9f03010acf13b3ffccb1a6e3f88021\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce18f4f06db091e658684fa0bbc858e19b9f03010acf13b3ffccb1a6e3f88021\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-26nc6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:47Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:47 crc kubenswrapper[4860]: I1211 08:11:47.660816 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0589a204-c98c-417d-8256-bfe3bf747660\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3333b4346f59b04017e32d61f457ba7ff457166731128f339d7987ae6b56b06d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52bb8ade314afb9494f4e026d0af2aea7b1970d64dc66102f55e257ae4678154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dfcc4370edaae36a5a32b0a9c29290c50f8e4f427e4e147dc3c779beaa8b780\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d856ae95e9077524c518d82d10fb6b93b7fbcaaf2b720ffafe6f7fae58bd7caa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1426084ac580895c42354630910dd8c8e3a48ddd25f22abce1d28dc3de332a63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9984d5d242bb258405efe91fc9caf9e7ea592834445df2e78add9553a921c76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://592dbdb2905039d89bdf741dbd0fd1e1fbd15acbabecdad6f07c758657fa610a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://592dbdb2905039d89bdf741dbd0fd1e1fbd15acbabecdad6f07c758657fa610a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-11T08:11:44Z\\\",\\\"message\\\":\\\"es.lbConfig(nil)\\\\nI1211 08:11:44.095388 6339 services_controller.go:356] Processing sync for service openshift-ingress-canary/ingress-canary for network=default\\\\nI1211 08:11:44.095870 6339 services_controller.go:445] Built service openshift-kube-controller-manager/kube-controller-manager LB template configs for network=default: []services.lbConfig(nil)\\\\nF1211 08:11:44.095867 6339 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:44Z is after 2025-08-24T17:21:41Z]\\\\nI1211 08:11:44.095898 6339 services_controller.go:451] Built service openshift-kube-controller-manager/kube-controller-manager cluster-wide LB for network\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:42Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-xvnqp_openshift-ovn-kubernetes(0589a204-c98c-417d-8256-bfe3bf747660)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb80d9928e8349fc5963227d135355ad7ba820d4fd89877af601d005ca03334b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17cbcb49c06f3a9059cd5a0f05eedb8c785da30a3e502e31bae2f57b3486cb14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17cbcb49c06f3a9059cd5a0f05eedb8c785da30a3e502e31bae2f57b3486cb14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xvnqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:47Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:47 crc kubenswrapper[4860]: I1211 08:11:47.678281 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gsx9l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aba5b6b9-f88d-4763-9562-6b4dec606df9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://735ef012c523ddc952d46a431f27b2582a2c301d9e9da0c697a20ff6e2fb5d86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6zw9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gsx9l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:47Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:47 crc kubenswrapper[4860]: I1211 08:11:47.699473 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9aa0aed-62b3-4ea2-82ae-9a653c1101a5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ad4ba1b769c6cb2b34073798f8e3c25eb0177e6d3059484bf365e6981cab710\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://077aa1dd2fe5189ff60a0f2365e47a6a545cca0de7aa6379370908db4ea5838a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d7c919899b74a2b0fcd9221f3f769a9d013267354b200a282669d342d6413d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5483384f1d2ebc1fefa9706c9e67a3dc5a00ff643f803e3f2707c252e27ad94\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:47Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:47 crc kubenswrapper[4860]: I1211 08:11:47.703202 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:47 crc kubenswrapper[4860]: I1211 08:11:47.703269 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:47 crc kubenswrapper[4860]: I1211 08:11:47.703283 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:47 crc kubenswrapper[4860]: I1211 08:11:47.703305 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:47 crc kubenswrapper[4860]: I1211 08:11:47.703319 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:47Z","lastTransitionTime":"2025-12-11T08:11:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:47 crc kubenswrapper[4860]: I1211 08:11:47.717522 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-854v6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e6529ba7-2341-43ac-a187-95ab530fe775\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8986538a421f0184f52f288b0e171da695591caf2b10c67ad2dce1073a6e6016\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kh22b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0d4e3f761ad15ae1424204e40b9d3f60188ab4a9047bd4b5ac6a8a8ca1fb385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kh22b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-854v6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:47Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:47 crc kubenswrapper[4860]: I1211 08:11:47.746094 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9872be1e-b995-4a00-904b-875dff96ad56\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f74a32a8fefea704cc73ffd1114083dab68aa44dc33f8d9d432d2d24216546c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbe00b46063d3968d59c393b87178e135f982abb7d0f32fa798b212d4b83df6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eedc9988e1a4a3c8cae730dc4a352103a6d665535b825e157d0f87c331b137c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2137fee4b7abc9765f9517d7b3b6ea796601a31d03151bf167d17136ced8b25a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f683e48f3577a68862eb6302383538376e41f49b20b1e0ea3f1b3b143f24872a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5de794187456b3576b969069a3b839d75c7918418a97fa35dff8bb40bb76e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5de794187456b3576b969069a3b839d75c7918418a97fa35dff8bb40bb76e39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08ce42b974fba00c409c0af971f0de351e50f18b2f06f8e722150da3239bb908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08ce42b974fba00c409c0af971f0de351e50f18b2f06f8e722150da3239bb908\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cfac6be03e3cfa0b0f3450b48b67957f2d5172b9c65e27f14c2a260bda5d423f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cfac6be03e3cfa0b0f3450b48b67957f2d5172b9c65e27f14c2a260bda5d423f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:47Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:47 crc kubenswrapper[4860]: I1211 08:11:47.763445 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-pznwl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5310c1fc-66c6-40aa-b1bf-5a59a2410f9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2d6ff00ce322a28607bc3f6db05c5e4fb14acb6777ea315786ef53963ca504a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hkt48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-pznwl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:47Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:47 crc kubenswrapper[4860]: I1211 08:11:47.777924 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d8c2f28628dc2f1ffbdd98a55670e22cf6cdc8cf2c6442e9d1f8e5466b2f12e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://074ea63beac52bd196507ee7cfb1c3d7fc2d37d5ec6faa4d47dde62b922e1c43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:47Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:47 crc kubenswrapper[4860]: I1211 08:11:47.790845 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:47Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:47 crc kubenswrapper[4860]: I1211 08:11:47.801547 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25a6a154da2b74952951ae94efc914ba078ea4bbe5893360b3c84d46be6bf973\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:47Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:47 crc kubenswrapper[4860]: I1211 08:11:47.804910 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:47 crc kubenswrapper[4860]: I1211 08:11:47.804940 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:47 crc kubenswrapper[4860]: I1211 08:11:47.804952 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:47 crc kubenswrapper[4860]: I1211 08:11:47.804970 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:47 crc kubenswrapper[4860]: I1211 08:11:47.804982 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:47Z","lastTransitionTime":"2025-12-11T08:11:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:47 crc kubenswrapper[4860]: I1211 08:11:47.811502 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6xq7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a75ce3c-5e1e-4b9f-bbde-7a3b0cce5eae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe8d4d675567b41728e476351811a93c47b19868ccbadab8c613a4eb8eca8562\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5c6nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6xq7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:47Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:47 crc kubenswrapper[4860]: I1211 08:11:47.821945 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:47Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:47 crc kubenswrapper[4860]: I1211 08:11:47.839846 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a41f35db087418d1be16953f10377042abd3f4488e5ac1bc29ca5dcfc00a37e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:47Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:47 crc kubenswrapper[4860]: I1211 08:11:47.850878 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gjhlj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5cfd640a-f3d2-4b28-87e1-5f49f6341e21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knq7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knq7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:42Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gjhlj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:47Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:47 crc kubenswrapper[4860]: I1211 08:11:47.863848 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ec3758-85f1-4a22-a4ec-c805472607a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e831db1a15976d3967d56479689572c26f2599a766baa0402897a3008428c56f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a498aa90f914e1f0370df01dc5ef174df113a2a16d86b2e3f24b19a1b99b4c0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8caebd2f4dfcbb89c597029a07c335171889761d2d200206987537fb0a228db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34e91f615b76445ff3241a304bb2964fde59de7e8a8dfff6841b2e628a31e86d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2090cab02ad74ddfd6bb57ff50eba35d18f56b0cde1c230fe011c01994b693bc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 08:11:21.171011 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 08:11:21.177674 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-644352688/tls.crt::/tmp/serving-cert-644352688/tls.key\\\\\\\"\\\\nI1211 08:11:26.508606 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 08:11:26.512741 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 08:11:26.512778 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 08:11:26.512839 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 08:11:26.512858 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 08:11:26.521684 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1211 08:11:26.521700 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1211 08:11:26.521722 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:11:26.521732 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:11:26.521738 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 08:11:26.521746 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 08:11:26.521749 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 08:11:26.521753 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1211 08:11:26.524999 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2274fb298ca7d2b37b060b4dbf9b166d39ba77a108b08edeeb4d67196e72cb3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf0aff8a6ab816b6617b1feeca9146d41bb5e8808bf1db96f8d493500c135d85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf0aff8a6ab816b6617b1feeca9146d41bb5e8808bf1db96f8d493500c135d85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:47Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:47 crc kubenswrapper[4860]: I1211 08:11:47.876210 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:47Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:47 crc kubenswrapper[4860]: I1211 08:11:47.908295 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:47 crc kubenswrapper[4860]: I1211 08:11:47.908558 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:47 crc kubenswrapper[4860]: I1211 08:11:47.908723 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:47 crc kubenswrapper[4860]: I1211 08:11:47.908866 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:47 crc kubenswrapper[4860]: I1211 08:11:47.908985 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:47Z","lastTransitionTime":"2025-12-11T08:11:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:48 crc kubenswrapper[4860]: I1211 08:11:48.012883 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:48 crc kubenswrapper[4860]: I1211 08:11:48.013084 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:48 crc kubenswrapper[4860]: I1211 08:11:48.013110 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:48 crc kubenswrapper[4860]: I1211 08:11:48.013197 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:48 crc kubenswrapper[4860]: I1211 08:11:48.013296 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:48Z","lastTransitionTime":"2025-12-11T08:11:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:48 crc kubenswrapper[4860]: I1211 08:11:48.116993 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:48 crc kubenswrapper[4860]: I1211 08:11:48.117063 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:48 crc kubenswrapper[4860]: I1211 08:11:48.117086 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:48 crc kubenswrapper[4860]: I1211 08:11:48.117113 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:48 crc kubenswrapper[4860]: I1211 08:11:48.117132 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:48Z","lastTransitionTime":"2025-12-11T08:11:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:48 crc kubenswrapper[4860]: I1211 08:11:48.220575 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:48 crc kubenswrapper[4860]: I1211 08:11:48.220708 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:48 crc kubenswrapper[4860]: I1211 08:11:48.220736 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:48 crc kubenswrapper[4860]: I1211 08:11:48.220770 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:48 crc kubenswrapper[4860]: I1211 08:11:48.220792 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:48Z","lastTransitionTime":"2025-12-11T08:11:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:48 crc kubenswrapper[4860]: I1211 08:11:48.324618 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:48 crc kubenswrapper[4860]: I1211 08:11:48.324713 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:48 crc kubenswrapper[4860]: I1211 08:11:48.324730 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:48 crc kubenswrapper[4860]: I1211 08:11:48.324755 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:48 crc kubenswrapper[4860]: I1211 08:11:48.324781 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:48Z","lastTransitionTime":"2025-12-11T08:11:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:48 crc kubenswrapper[4860]: I1211 08:11:48.428160 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:48 crc kubenswrapper[4860]: I1211 08:11:48.428238 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:48 crc kubenswrapper[4860]: I1211 08:11:48.428260 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:48 crc kubenswrapper[4860]: I1211 08:11:48.428290 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:48 crc kubenswrapper[4860]: I1211 08:11:48.428308 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:48Z","lastTransitionTime":"2025-12-11T08:11:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:48 crc kubenswrapper[4860]: I1211 08:11:48.530691 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:48 crc kubenswrapper[4860]: I1211 08:11:48.530765 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:48 crc kubenswrapper[4860]: I1211 08:11:48.530786 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:48 crc kubenswrapper[4860]: I1211 08:11:48.530812 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:48 crc kubenswrapper[4860]: I1211 08:11:48.530830 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:48Z","lastTransitionTime":"2025-12-11T08:11:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:48 crc kubenswrapper[4860]: I1211 08:11:48.578868 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gjhlj" Dec 11 08:11:48 crc kubenswrapper[4860]: I1211 08:11:48.578996 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:11:48 crc kubenswrapper[4860]: E1211 08:11:48.579105 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gjhlj" podUID="5cfd640a-f3d2-4b28-87e1-5f49f6341e21" Dec 11 08:11:48 crc kubenswrapper[4860]: I1211 08:11:48.579131 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:11:48 crc kubenswrapper[4860]: I1211 08:11:48.579194 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:11:48 crc kubenswrapper[4860]: E1211 08:11:48.579300 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:11:48 crc kubenswrapper[4860]: E1211 08:11:48.579603 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:11:48 crc kubenswrapper[4860]: E1211 08:11:48.579763 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:11:48 crc kubenswrapper[4860]: I1211 08:11:48.633433 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:48 crc kubenswrapper[4860]: I1211 08:11:48.633494 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:48 crc kubenswrapper[4860]: I1211 08:11:48.633515 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:48 crc kubenswrapper[4860]: I1211 08:11:48.633539 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:48 crc kubenswrapper[4860]: I1211 08:11:48.633559 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:48Z","lastTransitionTime":"2025-12-11T08:11:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:48 crc kubenswrapper[4860]: I1211 08:11:48.737037 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:48 crc kubenswrapper[4860]: I1211 08:11:48.737100 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:48 crc kubenswrapper[4860]: I1211 08:11:48.737118 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:48 crc kubenswrapper[4860]: I1211 08:11:48.737146 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:48 crc kubenswrapper[4860]: I1211 08:11:48.737164 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:48Z","lastTransitionTime":"2025-12-11T08:11:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:48 crc kubenswrapper[4860]: I1211 08:11:48.840697 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:48 crc kubenswrapper[4860]: I1211 08:11:48.840884 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:48 crc kubenswrapper[4860]: I1211 08:11:48.840917 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:48 crc kubenswrapper[4860]: I1211 08:11:48.840947 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:48 crc kubenswrapper[4860]: I1211 08:11:48.840969 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:48Z","lastTransitionTime":"2025-12-11T08:11:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:48 crc kubenswrapper[4860]: I1211 08:11:48.943113 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:48 crc kubenswrapper[4860]: I1211 08:11:48.943174 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:48 crc kubenswrapper[4860]: I1211 08:11:48.943196 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:48 crc kubenswrapper[4860]: I1211 08:11:48.943228 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:48 crc kubenswrapper[4860]: I1211 08:11:48.943253 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:48Z","lastTransitionTime":"2025-12-11T08:11:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:49 crc kubenswrapper[4860]: I1211 08:11:49.046394 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:49 crc kubenswrapper[4860]: I1211 08:11:49.046489 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:49 crc kubenswrapper[4860]: I1211 08:11:49.046506 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:49 crc kubenswrapper[4860]: I1211 08:11:49.046529 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:49 crc kubenswrapper[4860]: I1211 08:11:49.046546 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:49Z","lastTransitionTime":"2025-12-11T08:11:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:49 crc kubenswrapper[4860]: I1211 08:11:49.149816 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:49 crc kubenswrapper[4860]: I1211 08:11:49.149898 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:49 crc kubenswrapper[4860]: I1211 08:11:49.149912 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:49 crc kubenswrapper[4860]: I1211 08:11:49.149934 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:49 crc kubenswrapper[4860]: I1211 08:11:49.149949 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:49Z","lastTransitionTime":"2025-12-11T08:11:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:49 crc kubenswrapper[4860]: I1211 08:11:49.252811 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:49 crc kubenswrapper[4860]: I1211 08:11:49.252870 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:49 crc kubenswrapper[4860]: I1211 08:11:49.252889 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:49 crc kubenswrapper[4860]: I1211 08:11:49.252917 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:49 crc kubenswrapper[4860]: I1211 08:11:49.252937 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:49Z","lastTransitionTime":"2025-12-11T08:11:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:49 crc kubenswrapper[4860]: I1211 08:11:49.356343 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:49 crc kubenswrapper[4860]: I1211 08:11:49.356437 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:49 crc kubenswrapper[4860]: I1211 08:11:49.356463 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:49 crc kubenswrapper[4860]: I1211 08:11:49.356502 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:49 crc kubenswrapper[4860]: I1211 08:11:49.356531 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:49Z","lastTransitionTime":"2025-12-11T08:11:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:49 crc kubenswrapper[4860]: I1211 08:11:49.459492 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:49 crc kubenswrapper[4860]: I1211 08:11:49.459553 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:49 crc kubenswrapper[4860]: I1211 08:11:49.459565 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:49 crc kubenswrapper[4860]: I1211 08:11:49.459590 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:49 crc kubenswrapper[4860]: I1211 08:11:49.459604 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:49Z","lastTransitionTime":"2025-12-11T08:11:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:49 crc kubenswrapper[4860]: I1211 08:11:49.562849 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:49 crc kubenswrapper[4860]: I1211 08:11:49.562914 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:49 crc kubenswrapper[4860]: I1211 08:11:49.562931 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:49 crc kubenswrapper[4860]: I1211 08:11:49.562955 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:49 crc kubenswrapper[4860]: I1211 08:11:49.562979 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:49Z","lastTransitionTime":"2025-12-11T08:11:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:49 crc kubenswrapper[4860]: I1211 08:11:49.666686 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:49 crc kubenswrapper[4860]: I1211 08:11:49.666765 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:49 crc kubenswrapper[4860]: I1211 08:11:49.666783 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:49 crc kubenswrapper[4860]: I1211 08:11:49.666814 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:49 crc kubenswrapper[4860]: I1211 08:11:49.666836 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:49Z","lastTransitionTime":"2025-12-11T08:11:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:49 crc kubenswrapper[4860]: I1211 08:11:49.769993 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:49 crc kubenswrapper[4860]: I1211 08:11:49.770414 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:49 crc kubenswrapper[4860]: I1211 08:11:49.770444 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:49 crc kubenswrapper[4860]: I1211 08:11:49.770467 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:49 crc kubenswrapper[4860]: I1211 08:11:49.770484 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:49Z","lastTransitionTime":"2025-12-11T08:11:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:49 crc kubenswrapper[4860]: I1211 08:11:49.872971 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:49 crc kubenswrapper[4860]: I1211 08:11:49.873020 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:49 crc kubenswrapper[4860]: I1211 08:11:49.873038 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:49 crc kubenswrapper[4860]: I1211 08:11:49.873060 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:49 crc kubenswrapper[4860]: I1211 08:11:49.873077 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:49Z","lastTransitionTime":"2025-12-11T08:11:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:49 crc kubenswrapper[4860]: I1211 08:11:49.976977 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:49 crc kubenswrapper[4860]: I1211 08:11:49.977051 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:49 crc kubenswrapper[4860]: I1211 08:11:49.977070 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:49 crc kubenswrapper[4860]: I1211 08:11:49.977097 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:49 crc kubenswrapper[4860]: I1211 08:11:49.977118 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:49Z","lastTransitionTime":"2025-12-11T08:11:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:50 crc kubenswrapper[4860]: I1211 08:11:50.080862 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:50 crc kubenswrapper[4860]: I1211 08:11:50.080957 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:50 crc kubenswrapper[4860]: I1211 08:11:50.081028 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:50 crc kubenswrapper[4860]: I1211 08:11:50.081059 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:50 crc kubenswrapper[4860]: I1211 08:11:50.081108 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:50Z","lastTransitionTime":"2025-12-11T08:11:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:50 crc kubenswrapper[4860]: I1211 08:11:50.184737 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:50 crc kubenswrapper[4860]: I1211 08:11:50.184827 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:50 crc kubenswrapper[4860]: I1211 08:11:50.184852 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:50 crc kubenswrapper[4860]: I1211 08:11:50.184886 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:50 crc kubenswrapper[4860]: I1211 08:11:50.184912 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:50Z","lastTransitionTime":"2025-12-11T08:11:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:50 crc kubenswrapper[4860]: I1211 08:11:50.257021 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5cfd640a-f3d2-4b28-87e1-5f49f6341e21-metrics-certs\") pod \"network-metrics-daemon-gjhlj\" (UID: \"5cfd640a-f3d2-4b28-87e1-5f49f6341e21\") " pod="openshift-multus/network-metrics-daemon-gjhlj" Dec 11 08:11:50 crc kubenswrapper[4860]: E1211 08:11:50.257241 4860 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 11 08:11:50 crc kubenswrapper[4860]: E1211 08:11:50.257360 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5cfd640a-f3d2-4b28-87e1-5f49f6341e21-metrics-certs podName:5cfd640a-f3d2-4b28-87e1-5f49f6341e21 nodeName:}" failed. No retries permitted until 2025-12-11 08:11:58.257325364 +0000 UTC m=+50.985844449 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5cfd640a-f3d2-4b28-87e1-5f49f6341e21-metrics-certs") pod "network-metrics-daemon-gjhlj" (UID: "5cfd640a-f3d2-4b28-87e1-5f49f6341e21") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 11 08:11:50 crc kubenswrapper[4860]: I1211 08:11:50.288738 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:50 crc kubenswrapper[4860]: I1211 08:11:50.288810 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:50 crc kubenswrapper[4860]: I1211 08:11:50.288835 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:50 crc kubenswrapper[4860]: I1211 08:11:50.288862 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:50 crc kubenswrapper[4860]: I1211 08:11:50.288879 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:50Z","lastTransitionTime":"2025-12-11T08:11:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:50 crc kubenswrapper[4860]: I1211 08:11:50.392960 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:50 crc kubenswrapper[4860]: I1211 08:11:50.393008 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:50 crc kubenswrapper[4860]: I1211 08:11:50.393025 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:50 crc kubenswrapper[4860]: I1211 08:11:50.393046 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:50 crc kubenswrapper[4860]: I1211 08:11:50.393062 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:50Z","lastTransitionTime":"2025-12-11T08:11:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:50 crc kubenswrapper[4860]: I1211 08:11:50.497186 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:50 crc kubenswrapper[4860]: I1211 08:11:50.497258 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:50 crc kubenswrapper[4860]: I1211 08:11:50.497282 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:50 crc kubenswrapper[4860]: I1211 08:11:50.497306 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:50 crc kubenswrapper[4860]: I1211 08:11:50.497324 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:50Z","lastTransitionTime":"2025-12-11T08:11:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:50 crc kubenswrapper[4860]: I1211 08:11:50.578825 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gjhlj" Dec 11 08:11:50 crc kubenswrapper[4860]: I1211 08:11:50.578891 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:11:50 crc kubenswrapper[4860]: E1211 08:11:50.579053 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gjhlj" podUID="5cfd640a-f3d2-4b28-87e1-5f49f6341e21" Dec 11 08:11:50 crc kubenswrapper[4860]: I1211 08:11:50.578847 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:11:50 crc kubenswrapper[4860]: E1211 08:11:50.579320 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:11:50 crc kubenswrapper[4860]: I1211 08:11:50.579401 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:11:50 crc kubenswrapper[4860]: E1211 08:11:50.579491 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:11:50 crc kubenswrapper[4860]: E1211 08:11:50.579595 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:11:50 crc kubenswrapper[4860]: I1211 08:11:50.600678 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:50 crc kubenswrapper[4860]: I1211 08:11:50.600802 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:50 crc kubenswrapper[4860]: I1211 08:11:50.601039 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:50 crc kubenswrapper[4860]: I1211 08:11:50.601077 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:50 crc kubenswrapper[4860]: I1211 08:11:50.601440 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:50Z","lastTransitionTime":"2025-12-11T08:11:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:50 crc kubenswrapper[4860]: I1211 08:11:50.705224 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:50 crc kubenswrapper[4860]: I1211 08:11:50.705302 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:50 crc kubenswrapper[4860]: I1211 08:11:50.705322 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:50 crc kubenswrapper[4860]: I1211 08:11:50.705345 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:50 crc kubenswrapper[4860]: I1211 08:11:50.705363 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:50Z","lastTransitionTime":"2025-12-11T08:11:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:50 crc kubenswrapper[4860]: I1211 08:11:50.808869 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:50 crc kubenswrapper[4860]: I1211 08:11:50.808934 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:50 crc kubenswrapper[4860]: I1211 08:11:50.808952 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:50 crc kubenswrapper[4860]: I1211 08:11:50.808980 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:50 crc kubenswrapper[4860]: I1211 08:11:50.809002 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:50Z","lastTransitionTime":"2025-12-11T08:11:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:50 crc kubenswrapper[4860]: I1211 08:11:50.912678 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:50 crc kubenswrapper[4860]: I1211 08:11:50.912729 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:50 crc kubenswrapper[4860]: I1211 08:11:50.912740 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:50 crc kubenswrapper[4860]: I1211 08:11:50.912761 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:50 crc kubenswrapper[4860]: I1211 08:11:50.912774 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:50Z","lastTransitionTime":"2025-12-11T08:11:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:50 crc kubenswrapper[4860]: I1211 08:11:50.944315 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:50 crc kubenswrapper[4860]: I1211 08:11:50.944349 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:50 crc kubenswrapper[4860]: I1211 08:11:50.944360 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:50 crc kubenswrapper[4860]: I1211 08:11:50.944372 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:50 crc kubenswrapper[4860]: I1211 08:11:50.944381 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:50Z","lastTransitionTime":"2025-12-11T08:11:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:50 crc kubenswrapper[4860]: E1211 08:11:50.960072 4860 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:11:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:11:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:11:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:11:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"52631a1a-bc27-4dd5-8309-5dc39b5dcc41\\\",\\\"systemUUID\\\":\\\"299d3cd2-e1ad-4f5c-8af4-f12c0a5d76e7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:50Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:50 crc kubenswrapper[4860]: I1211 08:11:50.965637 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:50 crc kubenswrapper[4860]: I1211 08:11:50.965719 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:50 crc kubenswrapper[4860]: I1211 08:11:50.965737 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:50 crc kubenswrapper[4860]: I1211 08:11:50.965763 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:50 crc kubenswrapper[4860]: I1211 08:11:50.965783 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:50Z","lastTransitionTime":"2025-12-11T08:11:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:50 crc kubenswrapper[4860]: E1211 08:11:50.986189 4860 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:11:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:11:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:11:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:11:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"52631a1a-bc27-4dd5-8309-5dc39b5dcc41\\\",\\\"systemUUID\\\":\\\"299d3cd2-e1ad-4f5c-8af4-f12c0a5d76e7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:50Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:50 crc kubenswrapper[4860]: I1211 08:11:50.992780 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:50 crc kubenswrapper[4860]: I1211 08:11:50.992818 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:50 crc kubenswrapper[4860]: I1211 08:11:50.992828 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:50 crc kubenswrapper[4860]: I1211 08:11:50.992847 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:50 crc kubenswrapper[4860]: I1211 08:11:50.992860 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:50Z","lastTransitionTime":"2025-12-11T08:11:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:51 crc kubenswrapper[4860]: E1211 08:11:51.013343 4860 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:11:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:11:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:11:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:11:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"52631a1a-bc27-4dd5-8309-5dc39b5dcc41\\\",\\\"systemUUID\\\":\\\"299d3cd2-e1ad-4f5c-8af4-f12c0a5d76e7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:51Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:51 crc kubenswrapper[4860]: I1211 08:11:51.019352 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:51 crc kubenswrapper[4860]: I1211 08:11:51.019428 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:51 crc kubenswrapper[4860]: I1211 08:11:51.019447 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:51 crc kubenswrapper[4860]: I1211 08:11:51.019476 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:51 crc kubenswrapper[4860]: I1211 08:11:51.019495 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:51Z","lastTransitionTime":"2025-12-11T08:11:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:51 crc kubenswrapper[4860]: E1211 08:11:51.038685 4860 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:11:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:11:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:11:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:11:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"52631a1a-bc27-4dd5-8309-5dc39b5dcc41\\\",\\\"systemUUID\\\":\\\"299d3cd2-e1ad-4f5c-8af4-f12c0a5d76e7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:51Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:51 crc kubenswrapper[4860]: I1211 08:11:51.044629 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:51 crc kubenswrapper[4860]: I1211 08:11:51.044727 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:51 crc kubenswrapper[4860]: I1211 08:11:51.044743 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:51 crc kubenswrapper[4860]: I1211 08:11:51.044766 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:51 crc kubenswrapper[4860]: I1211 08:11:51.044779 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:51Z","lastTransitionTime":"2025-12-11T08:11:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:51 crc kubenswrapper[4860]: E1211 08:11:51.064681 4860 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:11:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:11:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:11:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:11:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"52631a1a-bc27-4dd5-8309-5dc39b5dcc41\\\",\\\"systemUUID\\\":\\\"299d3cd2-e1ad-4f5c-8af4-f12c0a5d76e7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:51Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:51 crc kubenswrapper[4860]: E1211 08:11:51.064845 4860 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 11 08:11:51 crc kubenswrapper[4860]: I1211 08:11:51.067624 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:51 crc kubenswrapper[4860]: I1211 08:11:51.067751 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:51 crc kubenswrapper[4860]: I1211 08:11:51.067778 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:51 crc kubenswrapper[4860]: I1211 08:11:51.067811 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:51 crc kubenswrapper[4860]: I1211 08:11:51.067835 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:51Z","lastTransitionTime":"2025-12-11T08:11:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:51 crc kubenswrapper[4860]: I1211 08:11:51.170810 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:51 crc kubenswrapper[4860]: I1211 08:11:51.170864 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:51 crc kubenswrapper[4860]: I1211 08:11:51.170873 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:51 crc kubenswrapper[4860]: I1211 08:11:51.170896 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:51 crc kubenswrapper[4860]: I1211 08:11:51.170907 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:51Z","lastTransitionTime":"2025-12-11T08:11:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:51 crc kubenswrapper[4860]: I1211 08:11:51.274527 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:51 crc kubenswrapper[4860]: I1211 08:11:51.274584 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:51 crc kubenswrapper[4860]: I1211 08:11:51.274601 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:51 crc kubenswrapper[4860]: I1211 08:11:51.274628 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:51 crc kubenswrapper[4860]: I1211 08:11:51.274673 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:51Z","lastTransitionTime":"2025-12-11T08:11:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:51 crc kubenswrapper[4860]: I1211 08:11:51.378184 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:51 crc kubenswrapper[4860]: I1211 08:11:51.378265 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:51 crc kubenswrapper[4860]: I1211 08:11:51.378291 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:51 crc kubenswrapper[4860]: I1211 08:11:51.378320 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:51 crc kubenswrapper[4860]: I1211 08:11:51.378340 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:51Z","lastTransitionTime":"2025-12-11T08:11:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:51 crc kubenswrapper[4860]: I1211 08:11:51.481546 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:51 crc kubenswrapper[4860]: I1211 08:11:51.481614 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:51 crc kubenswrapper[4860]: I1211 08:11:51.481632 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:51 crc kubenswrapper[4860]: I1211 08:11:51.481687 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:51 crc kubenswrapper[4860]: I1211 08:11:51.481705 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:51Z","lastTransitionTime":"2025-12-11T08:11:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:51 crc kubenswrapper[4860]: I1211 08:11:51.585711 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:51 crc kubenswrapper[4860]: I1211 08:11:51.585777 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:51 crc kubenswrapper[4860]: I1211 08:11:51.585794 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:51 crc kubenswrapper[4860]: I1211 08:11:51.585819 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:51 crc kubenswrapper[4860]: I1211 08:11:51.585838 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:51Z","lastTransitionTime":"2025-12-11T08:11:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:51 crc kubenswrapper[4860]: I1211 08:11:51.689161 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:51 crc kubenswrapper[4860]: I1211 08:11:51.689219 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:51 crc kubenswrapper[4860]: I1211 08:11:51.689235 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:51 crc kubenswrapper[4860]: I1211 08:11:51.689290 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:51 crc kubenswrapper[4860]: I1211 08:11:51.689311 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:51Z","lastTransitionTime":"2025-12-11T08:11:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:51 crc kubenswrapper[4860]: I1211 08:11:51.792916 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:51 crc kubenswrapper[4860]: I1211 08:11:51.793057 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:51 crc kubenswrapper[4860]: I1211 08:11:51.793081 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:51 crc kubenswrapper[4860]: I1211 08:11:51.793106 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:51 crc kubenswrapper[4860]: I1211 08:11:51.793127 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:51Z","lastTransitionTime":"2025-12-11T08:11:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:51 crc kubenswrapper[4860]: I1211 08:11:51.896478 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:51 crc kubenswrapper[4860]: I1211 08:11:51.896552 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:51 crc kubenswrapper[4860]: I1211 08:11:51.896574 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:51 crc kubenswrapper[4860]: I1211 08:11:51.896602 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:51 crc kubenswrapper[4860]: I1211 08:11:51.896623 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:51Z","lastTransitionTime":"2025-12-11T08:11:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:52 crc kubenswrapper[4860]: I1211 08:11:52.000333 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:52 crc kubenswrapper[4860]: I1211 08:11:52.000391 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:52 crc kubenswrapper[4860]: I1211 08:11:52.000409 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:52 crc kubenswrapper[4860]: I1211 08:11:52.000433 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:52 crc kubenswrapper[4860]: I1211 08:11:52.000450 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:52Z","lastTransitionTime":"2025-12-11T08:11:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:52 crc kubenswrapper[4860]: I1211 08:11:52.104117 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:52 crc kubenswrapper[4860]: I1211 08:11:52.104204 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:52 crc kubenswrapper[4860]: I1211 08:11:52.104224 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:52 crc kubenswrapper[4860]: I1211 08:11:52.104253 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:52 crc kubenswrapper[4860]: I1211 08:11:52.104270 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:52Z","lastTransitionTime":"2025-12-11T08:11:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:52 crc kubenswrapper[4860]: I1211 08:11:52.207524 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:52 crc kubenswrapper[4860]: I1211 08:11:52.207575 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:52 crc kubenswrapper[4860]: I1211 08:11:52.207583 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:52 crc kubenswrapper[4860]: I1211 08:11:52.207598 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:52 crc kubenswrapper[4860]: I1211 08:11:52.207608 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:52Z","lastTransitionTime":"2025-12-11T08:11:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:52 crc kubenswrapper[4860]: I1211 08:11:52.310251 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:52 crc kubenswrapper[4860]: I1211 08:11:52.310304 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:52 crc kubenswrapper[4860]: I1211 08:11:52.310322 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:52 crc kubenswrapper[4860]: I1211 08:11:52.310344 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:52 crc kubenswrapper[4860]: I1211 08:11:52.310364 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:52Z","lastTransitionTime":"2025-12-11T08:11:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:52 crc kubenswrapper[4860]: I1211 08:11:52.413063 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:52 crc kubenswrapper[4860]: I1211 08:11:52.413105 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:52 crc kubenswrapper[4860]: I1211 08:11:52.413117 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:52 crc kubenswrapper[4860]: I1211 08:11:52.413133 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:52 crc kubenswrapper[4860]: I1211 08:11:52.413145 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:52Z","lastTransitionTime":"2025-12-11T08:11:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:52 crc kubenswrapper[4860]: I1211 08:11:52.515252 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:52 crc kubenswrapper[4860]: I1211 08:11:52.515287 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:52 crc kubenswrapper[4860]: I1211 08:11:52.515316 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:52 crc kubenswrapper[4860]: I1211 08:11:52.515332 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:52 crc kubenswrapper[4860]: I1211 08:11:52.515340 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:52Z","lastTransitionTime":"2025-12-11T08:11:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:52 crc kubenswrapper[4860]: I1211 08:11:52.578921 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:11:52 crc kubenswrapper[4860]: I1211 08:11:52.579042 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:11:52 crc kubenswrapper[4860]: I1211 08:11:52.578953 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:11:52 crc kubenswrapper[4860]: I1211 08:11:52.578958 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gjhlj" Dec 11 08:11:52 crc kubenswrapper[4860]: E1211 08:11:52.579395 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:11:52 crc kubenswrapper[4860]: E1211 08:11:52.579532 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:11:52 crc kubenswrapper[4860]: E1211 08:11:52.579728 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:11:52 crc kubenswrapper[4860]: E1211 08:11:52.580074 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gjhlj" podUID="5cfd640a-f3d2-4b28-87e1-5f49f6341e21" Dec 11 08:11:52 crc kubenswrapper[4860]: I1211 08:11:52.619891 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:52 crc kubenswrapper[4860]: I1211 08:11:52.619976 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:52 crc kubenswrapper[4860]: I1211 08:11:52.620003 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:52 crc kubenswrapper[4860]: I1211 08:11:52.620087 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:52 crc kubenswrapper[4860]: I1211 08:11:52.620122 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:52Z","lastTransitionTime":"2025-12-11T08:11:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:52 crc kubenswrapper[4860]: I1211 08:11:52.722886 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:52 crc kubenswrapper[4860]: I1211 08:11:52.722947 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:52 crc kubenswrapper[4860]: I1211 08:11:52.722970 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:52 crc kubenswrapper[4860]: I1211 08:11:52.723006 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:52 crc kubenswrapper[4860]: I1211 08:11:52.723030 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:52Z","lastTransitionTime":"2025-12-11T08:11:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:52 crc kubenswrapper[4860]: I1211 08:11:52.825908 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:52 crc kubenswrapper[4860]: I1211 08:11:52.825963 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:52 crc kubenswrapper[4860]: I1211 08:11:52.825976 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:52 crc kubenswrapper[4860]: I1211 08:11:52.825995 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:52 crc kubenswrapper[4860]: I1211 08:11:52.826007 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:52Z","lastTransitionTime":"2025-12-11T08:11:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:52 crc kubenswrapper[4860]: I1211 08:11:52.928693 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:52 crc kubenswrapper[4860]: I1211 08:11:52.928749 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:52 crc kubenswrapper[4860]: I1211 08:11:52.928762 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:52 crc kubenswrapper[4860]: I1211 08:11:52.928785 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:52 crc kubenswrapper[4860]: I1211 08:11:52.928800 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:52Z","lastTransitionTime":"2025-12-11T08:11:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:53 crc kubenswrapper[4860]: I1211 08:11:53.032126 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:53 crc kubenswrapper[4860]: I1211 08:11:53.032195 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:53 crc kubenswrapper[4860]: I1211 08:11:53.032215 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:53 crc kubenswrapper[4860]: I1211 08:11:53.032240 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:53 crc kubenswrapper[4860]: I1211 08:11:53.032260 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:53Z","lastTransitionTime":"2025-12-11T08:11:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:53 crc kubenswrapper[4860]: I1211 08:11:53.136843 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:53 crc kubenswrapper[4860]: I1211 08:11:53.136946 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:53 crc kubenswrapper[4860]: I1211 08:11:53.136979 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:53 crc kubenswrapper[4860]: I1211 08:11:53.137017 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:53 crc kubenswrapper[4860]: I1211 08:11:53.137045 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:53Z","lastTransitionTime":"2025-12-11T08:11:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:53 crc kubenswrapper[4860]: I1211 08:11:53.241100 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:53 crc kubenswrapper[4860]: I1211 08:11:53.241148 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:53 crc kubenswrapper[4860]: I1211 08:11:53.241159 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:53 crc kubenswrapper[4860]: I1211 08:11:53.241175 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:53 crc kubenswrapper[4860]: I1211 08:11:53.241186 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:53Z","lastTransitionTime":"2025-12-11T08:11:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:53 crc kubenswrapper[4860]: I1211 08:11:53.344005 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:53 crc kubenswrapper[4860]: I1211 08:11:53.344065 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:53 crc kubenswrapper[4860]: I1211 08:11:53.344083 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:53 crc kubenswrapper[4860]: I1211 08:11:53.344107 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:53 crc kubenswrapper[4860]: I1211 08:11:53.344124 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:53Z","lastTransitionTime":"2025-12-11T08:11:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:53 crc kubenswrapper[4860]: I1211 08:11:53.447720 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:53 crc kubenswrapper[4860]: I1211 08:11:53.447766 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:53 crc kubenswrapper[4860]: I1211 08:11:53.447778 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:53 crc kubenswrapper[4860]: I1211 08:11:53.447798 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:53 crc kubenswrapper[4860]: I1211 08:11:53.447810 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:53Z","lastTransitionTime":"2025-12-11T08:11:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:53 crc kubenswrapper[4860]: I1211 08:11:53.550950 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:53 crc kubenswrapper[4860]: I1211 08:11:53.551006 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:53 crc kubenswrapper[4860]: I1211 08:11:53.551017 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:53 crc kubenswrapper[4860]: I1211 08:11:53.551034 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:53 crc kubenswrapper[4860]: I1211 08:11:53.551046 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:53Z","lastTransitionTime":"2025-12-11T08:11:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:53 crc kubenswrapper[4860]: I1211 08:11:53.654001 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:53 crc kubenswrapper[4860]: I1211 08:11:53.654081 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:53 crc kubenswrapper[4860]: I1211 08:11:53.654106 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:53 crc kubenswrapper[4860]: I1211 08:11:53.654130 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:53 crc kubenswrapper[4860]: I1211 08:11:53.654146 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:53Z","lastTransitionTime":"2025-12-11T08:11:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:53 crc kubenswrapper[4860]: I1211 08:11:53.757518 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:53 crc kubenswrapper[4860]: I1211 08:11:53.757964 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:53 crc kubenswrapper[4860]: I1211 08:11:53.758104 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:53 crc kubenswrapper[4860]: I1211 08:11:53.758264 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:53 crc kubenswrapper[4860]: I1211 08:11:53.758399 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:53Z","lastTransitionTime":"2025-12-11T08:11:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:53 crc kubenswrapper[4860]: I1211 08:11:53.862055 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:53 crc kubenswrapper[4860]: I1211 08:11:53.862357 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:53 crc kubenswrapper[4860]: I1211 08:11:53.862486 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:53 crc kubenswrapper[4860]: I1211 08:11:53.862735 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:53 crc kubenswrapper[4860]: I1211 08:11:53.862971 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:53Z","lastTransitionTime":"2025-12-11T08:11:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:53 crc kubenswrapper[4860]: I1211 08:11:53.965432 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:53 crc kubenswrapper[4860]: I1211 08:11:53.965477 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:53 crc kubenswrapper[4860]: I1211 08:11:53.965492 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:53 crc kubenswrapper[4860]: I1211 08:11:53.965514 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:53 crc kubenswrapper[4860]: I1211 08:11:53.965530 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:53Z","lastTransitionTime":"2025-12-11T08:11:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:54 crc kubenswrapper[4860]: I1211 08:11:54.066177 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 11 08:11:54 crc kubenswrapper[4860]: I1211 08:11:54.068442 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:54 crc kubenswrapper[4860]: I1211 08:11:54.068496 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:54 crc kubenswrapper[4860]: I1211 08:11:54.068520 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:54 crc kubenswrapper[4860]: I1211 08:11:54.068547 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:54 crc kubenswrapper[4860]: I1211 08:11:54.068569 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:54Z","lastTransitionTime":"2025-12-11T08:11:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:54 crc kubenswrapper[4860]: I1211 08:11:54.078459 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 11 08:11:54 crc kubenswrapper[4860]: I1211 08:11:54.085549 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-854v6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e6529ba7-2341-43ac-a187-95ab530fe775\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8986538a421f0184f52f288b0e171da695591caf2b10c67ad2dce1073a6e6016\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kh22b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0d4e3f761ad15ae1424204e40b9d3f60188ab4a9047bd4b5ac6a8a8ca1fb385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kh22b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-854v6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:54Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:54 crc kubenswrapper[4860]: I1211 08:11:54.118361 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9872be1e-b995-4a00-904b-875dff96ad56\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f74a32a8fefea704cc73ffd1114083dab68aa44dc33f8d9d432d2d24216546c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbe00b46063d3968d59c393b87178e135f982abb7d0f32fa798b212d4b83df6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eedc9988e1a4a3c8cae730dc4a352103a6d665535b825e157d0f87c331b137c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2137fee4b7abc9765f9517d7b3b6ea796601a31d03151bf167d17136ced8b25a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f683e48f3577a68862eb6302383538376e41f49b20b1e0ea3f1b3b143f24872a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5de794187456b3576b969069a3b839d75c7918418a97fa35dff8bb40bb76e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5de794187456b3576b969069a3b839d75c7918418a97fa35dff8bb40bb76e39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08ce42b974fba00c409c0af971f0de351e50f18b2f06f8e722150da3239bb908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08ce42b974fba00c409c0af971f0de351e50f18b2f06f8e722150da3239bb908\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cfac6be03e3cfa0b0f3450b48b67957f2d5172b9c65e27f14c2a260bda5d423f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cfac6be03e3cfa0b0f3450b48b67957f2d5172b9c65e27f14c2a260bda5d423f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:54Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:54 crc kubenswrapper[4860]: I1211 08:11:54.136852 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-pznwl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5310c1fc-66c6-40aa-b1bf-5a59a2410f9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2d6ff00ce322a28607bc3f6db05c5e4fb14acb6777ea315786ef53963ca504a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hkt48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-pznwl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:54Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:54 crc kubenswrapper[4860]: I1211 08:11:54.155988 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d8c2f28628dc2f1ffbdd98a55670e22cf6cdc8cf2c6442e9d1f8e5466b2f12e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://074ea63beac52bd196507ee7cfb1c3d7fc2d37d5ec6faa4d47dde62b922e1c43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:54Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:54 crc kubenswrapper[4860]: I1211 08:11:54.171270 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:54Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:54 crc kubenswrapper[4860]: I1211 08:11:54.172535 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:54 crc kubenswrapper[4860]: I1211 08:11:54.172570 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:54 crc kubenswrapper[4860]: I1211 08:11:54.172579 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:54 crc kubenswrapper[4860]: I1211 08:11:54.172593 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:54 crc kubenswrapper[4860]: I1211 08:11:54.172606 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:54Z","lastTransitionTime":"2025-12-11T08:11:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:54 crc kubenswrapper[4860]: I1211 08:11:54.188501 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25a6a154da2b74952951ae94efc914ba078ea4bbe5893360b3c84d46be6bf973\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:54Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:54 crc kubenswrapper[4860]: I1211 08:11:54.201039 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6xq7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a75ce3c-5e1e-4b9f-bbde-7a3b0cce5eae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe8d4d675567b41728e476351811a93c47b19868ccbadab8c613a4eb8eca8562\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5c6nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6xq7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:54Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:54 crc kubenswrapper[4860]: I1211 08:11:54.215153 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:54Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:54 crc kubenswrapper[4860]: I1211 08:11:54.233773 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a41f35db087418d1be16953f10377042abd3f4488e5ac1bc29ca5dcfc00a37e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:54Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:54 crc kubenswrapper[4860]: I1211 08:11:54.248562 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gjhlj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5cfd640a-f3d2-4b28-87e1-5f49f6341e21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knq7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knq7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:42Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gjhlj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:54Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:54 crc kubenswrapper[4860]: I1211 08:11:54.271748 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ec3758-85f1-4a22-a4ec-c805472607a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e831db1a15976d3967d56479689572c26f2599a766baa0402897a3008428c56f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a498aa90f914e1f0370df01dc5ef174df113a2a16d86b2e3f24b19a1b99b4c0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8caebd2f4dfcbb89c597029a07c335171889761d2d200206987537fb0a228db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34e91f615b76445ff3241a304bb2964fde59de7e8a8dfff6841b2e628a31e86d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2090cab02ad74ddfd6bb57ff50eba35d18f56b0cde1c230fe011c01994b693bc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 08:11:21.171011 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 08:11:21.177674 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-644352688/tls.crt::/tmp/serving-cert-644352688/tls.key\\\\\\\"\\\\nI1211 08:11:26.508606 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 08:11:26.512741 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 08:11:26.512778 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 08:11:26.512839 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 08:11:26.512858 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 08:11:26.521684 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1211 08:11:26.521700 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1211 08:11:26.521722 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:11:26.521732 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:11:26.521738 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 08:11:26.521746 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 08:11:26.521749 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 08:11:26.521753 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1211 08:11:26.524999 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2274fb298ca7d2b37b060b4dbf9b166d39ba77a108b08edeeb4d67196e72cb3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf0aff8a6ab816b6617b1feeca9146d41bb5e8808bf1db96f8d493500c135d85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf0aff8a6ab816b6617b1feeca9146d41bb5e8808bf1db96f8d493500c135d85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:54Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:54 crc kubenswrapper[4860]: I1211 08:11:54.276211 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:54 crc kubenswrapper[4860]: I1211 08:11:54.276275 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:54 crc kubenswrapper[4860]: I1211 08:11:54.276294 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:54 crc kubenswrapper[4860]: I1211 08:11:54.276322 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:54 crc kubenswrapper[4860]: I1211 08:11:54.276340 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:54Z","lastTransitionTime":"2025-12-11T08:11:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:54 crc kubenswrapper[4860]: I1211 08:11:54.291357 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:54Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:54 crc kubenswrapper[4860]: I1211 08:11:54.309947 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31c30642-6e60-41b4-a477-0d802424e0aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79969402f5c33852a26fc839f37a45febf4f2377d77823284fbb01769e5585fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2d4d5de7106175087d91acb1f0c7552014f0c08bfcd65cbbf184d3f8d7826ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-99qgp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:54Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:54 crc kubenswrapper[4860]: I1211 08:11:54.340021 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-26nc6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1faa733-f408-41fd-b7aa-0cd86283976a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fce5b642b843ed30823c94e0b0c3d8bb1e5d2d93d691f7b8ef901ece03b0ac6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ac45db4f858a1c18a6b87472fc1cea2300f5845f1ad62c22260374995548a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ac45db4f858a1c18a6b87472fc1cea2300f5845f1ad62c22260374995548a0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b35296a5eb17dd2bb5e72c86c1cafdb30b7712cb330264f935d3b4fff05cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c6b35296a5eb17dd2bb5e72c86c1cafdb30b7712cb330264f935d3b4fff05cd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b9e82d6d3a247309ef5bdbf5b3a86ec083add2a138b0b237801e77b09ee96a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b9e82d6d3a247309ef5bdbf5b3a86ec083add2a138b0b237801e77b09ee96a1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://908ee695e78145ef4375c4097d9f1e772502922d02a2fcb4566172a6ad9caaf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://908ee695e78145ef4375c4097d9f1e772502922d02a2fcb4566172a6ad9caaf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afb0a097f3556a8f9ed7cf8f762cb99c082fe0a36b8d8d3038bf188b08fe9a84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afb0a097f3556a8f9ed7cf8f762cb99c082fe0a36b8d8d3038bf188b08fe9a84\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce18f4f06db091e658684fa0bbc858e19b9f03010acf13b3ffccb1a6e3f88021\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce18f4f06db091e658684fa0bbc858e19b9f03010acf13b3ffccb1a6e3f88021\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-26nc6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:54Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:54 crc kubenswrapper[4860]: I1211 08:11:54.372367 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0589a204-c98c-417d-8256-bfe3bf747660\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3333b4346f59b04017e32d61f457ba7ff457166731128f339d7987ae6b56b06d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52bb8ade314afb9494f4e026d0af2aea7b1970d64dc66102f55e257ae4678154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dfcc4370edaae36a5a32b0a9c29290c50f8e4f427e4e147dc3c779beaa8b780\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d856ae95e9077524c518d82d10fb6b93b7fbcaaf2b720ffafe6f7fae58bd7caa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1426084ac580895c42354630910dd8c8e3a48ddd25f22abce1d28dc3de332a63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9984d5d242bb258405efe91fc9caf9e7ea592834445df2e78add9553a921c76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://592dbdb2905039d89bdf741dbd0fd1e1fbd15acbabecdad6f07c758657fa610a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://592dbdb2905039d89bdf741dbd0fd1e1fbd15acbabecdad6f07c758657fa610a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-11T08:11:44Z\\\",\\\"message\\\":\\\"es.lbConfig(nil)\\\\nI1211 08:11:44.095388 6339 services_controller.go:356] Processing sync for service openshift-ingress-canary/ingress-canary for network=default\\\\nI1211 08:11:44.095870 6339 services_controller.go:445] Built service openshift-kube-controller-manager/kube-controller-manager LB template configs for network=default: []services.lbConfig(nil)\\\\nF1211 08:11:44.095867 6339 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:44Z is after 2025-08-24T17:21:41Z]\\\\nI1211 08:11:44.095898 6339 services_controller.go:451] Built service openshift-kube-controller-manager/kube-controller-manager cluster-wide LB for network\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:42Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-xvnqp_openshift-ovn-kubernetes(0589a204-c98c-417d-8256-bfe3bf747660)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb80d9928e8349fc5963227d135355ad7ba820d4fd89877af601d005ca03334b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17cbcb49c06f3a9059cd5a0f05eedb8c785da30a3e502e31bae2f57b3486cb14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17cbcb49c06f3a9059cd5a0f05eedb8c785da30a3e502e31bae2f57b3486cb14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xvnqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:54Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:54 crc kubenswrapper[4860]: I1211 08:11:54.379920 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:54 crc kubenswrapper[4860]: I1211 08:11:54.380197 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:54 crc kubenswrapper[4860]: I1211 08:11:54.380222 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:54 crc kubenswrapper[4860]: I1211 08:11:54.380253 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:54 crc kubenswrapper[4860]: I1211 08:11:54.380277 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:54Z","lastTransitionTime":"2025-12-11T08:11:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:54 crc kubenswrapper[4860]: I1211 08:11:54.416847 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gsx9l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aba5b6b9-f88d-4763-9562-6b4dec606df9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://735ef012c523ddc952d46a431f27b2582a2c301d9e9da0c697a20ff6e2fb5d86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6zw9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gsx9l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:54Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:54 crc kubenswrapper[4860]: I1211 08:11:54.440986 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9aa0aed-62b3-4ea2-82ae-9a653c1101a5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ad4ba1b769c6cb2b34073798f8e3c25eb0177e6d3059484bf365e6981cab710\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://077aa1dd2fe5189ff60a0f2365e47a6a545cca0de7aa6379370908db4ea5838a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d7c919899b74a2b0fcd9221f3f769a9d013267354b200a282669d342d6413d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5483384f1d2ebc1fefa9706c9e67a3dc5a00ff643f803e3f2707c252e27ad94\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:54Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:54 crc kubenswrapper[4860]: I1211 08:11:54.483958 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:54 crc kubenswrapper[4860]: I1211 08:11:54.484022 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:54 crc kubenswrapper[4860]: I1211 08:11:54.484045 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:54 crc kubenswrapper[4860]: I1211 08:11:54.484074 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:54 crc kubenswrapper[4860]: I1211 08:11:54.484094 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:54Z","lastTransitionTime":"2025-12-11T08:11:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:54 crc kubenswrapper[4860]: I1211 08:11:54.578403 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:11:54 crc kubenswrapper[4860]: I1211 08:11:54.578436 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:11:54 crc kubenswrapper[4860]: I1211 08:11:54.578438 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gjhlj" Dec 11 08:11:54 crc kubenswrapper[4860]: E1211 08:11:54.578596 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:11:54 crc kubenswrapper[4860]: I1211 08:11:54.578743 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:11:54 crc kubenswrapper[4860]: E1211 08:11:54.578758 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:11:54 crc kubenswrapper[4860]: E1211 08:11:54.578924 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gjhlj" podUID="5cfd640a-f3d2-4b28-87e1-5f49f6341e21" Dec 11 08:11:54 crc kubenswrapper[4860]: E1211 08:11:54.579101 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:11:54 crc kubenswrapper[4860]: I1211 08:11:54.590053 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:54 crc kubenswrapper[4860]: I1211 08:11:54.590139 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:54 crc kubenswrapper[4860]: I1211 08:11:54.590167 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:54 crc kubenswrapper[4860]: I1211 08:11:54.590197 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:54 crc kubenswrapper[4860]: I1211 08:11:54.590219 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:54Z","lastTransitionTime":"2025-12-11T08:11:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:54 crc kubenswrapper[4860]: I1211 08:11:54.692878 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:54 crc kubenswrapper[4860]: I1211 08:11:54.692952 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:54 crc kubenswrapper[4860]: I1211 08:11:54.692977 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:54 crc kubenswrapper[4860]: I1211 08:11:54.693005 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:54 crc kubenswrapper[4860]: I1211 08:11:54.693028 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:54Z","lastTransitionTime":"2025-12-11T08:11:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:54 crc kubenswrapper[4860]: I1211 08:11:54.796094 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:54 crc kubenswrapper[4860]: I1211 08:11:54.796154 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:54 crc kubenswrapper[4860]: I1211 08:11:54.796179 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:54 crc kubenswrapper[4860]: I1211 08:11:54.796206 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:54 crc kubenswrapper[4860]: I1211 08:11:54.796227 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:54Z","lastTransitionTime":"2025-12-11T08:11:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:54 crc kubenswrapper[4860]: I1211 08:11:54.899494 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:54 crc kubenswrapper[4860]: I1211 08:11:54.899570 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:54 crc kubenswrapper[4860]: I1211 08:11:54.899592 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:54 crc kubenswrapper[4860]: I1211 08:11:54.899619 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:54 crc kubenswrapper[4860]: I1211 08:11:54.899671 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:54Z","lastTransitionTime":"2025-12-11T08:11:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:55 crc kubenswrapper[4860]: I1211 08:11:55.003084 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:55 crc kubenswrapper[4860]: I1211 08:11:55.003167 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:55 crc kubenswrapper[4860]: I1211 08:11:55.003191 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:55 crc kubenswrapper[4860]: I1211 08:11:55.003225 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:55 crc kubenswrapper[4860]: I1211 08:11:55.003251 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:55Z","lastTransitionTime":"2025-12-11T08:11:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:55 crc kubenswrapper[4860]: I1211 08:11:55.106100 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:55 crc kubenswrapper[4860]: I1211 08:11:55.106171 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:55 crc kubenswrapper[4860]: I1211 08:11:55.106187 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:55 crc kubenswrapper[4860]: I1211 08:11:55.106212 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:55 crc kubenswrapper[4860]: I1211 08:11:55.106229 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:55Z","lastTransitionTime":"2025-12-11T08:11:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:55 crc kubenswrapper[4860]: I1211 08:11:55.209918 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:55 crc kubenswrapper[4860]: I1211 08:11:55.210448 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:55 crc kubenswrapper[4860]: I1211 08:11:55.210477 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:55 crc kubenswrapper[4860]: I1211 08:11:55.210509 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:55 crc kubenswrapper[4860]: I1211 08:11:55.210533 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:55Z","lastTransitionTime":"2025-12-11T08:11:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:55 crc kubenswrapper[4860]: I1211 08:11:55.313507 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:55 crc kubenswrapper[4860]: I1211 08:11:55.313550 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:55 crc kubenswrapper[4860]: I1211 08:11:55.313559 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:55 crc kubenswrapper[4860]: I1211 08:11:55.313573 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:55 crc kubenswrapper[4860]: I1211 08:11:55.313583 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:55Z","lastTransitionTime":"2025-12-11T08:11:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:55 crc kubenswrapper[4860]: I1211 08:11:55.415903 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:55 crc kubenswrapper[4860]: I1211 08:11:55.415945 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:55 crc kubenswrapper[4860]: I1211 08:11:55.415956 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:55 crc kubenswrapper[4860]: I1211 08:11:55.415971 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:55 crc kubenswrapper[4860]: I1211 08:11:55.415983 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:55Z","lastTransitionTime":"2025-12-11T08:11:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:55 crc kubenswrapper[4860]: I1211 08:11:55.519278 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:55 crc kubenswrapper[4860]: I1211 08:11:55.519350 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:55 crc kubenswrapper[4860]: I1211 08:11:55.519369 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:55 crc kubenswrapper[4860]: I1211 08:11:55.519394 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:55 crc kubenswrapper[4860]: I1211 08:11:55.519412 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:55Z","lastTransitionTime":"2025-12-11T08:11:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:55 crc kubenswrapper[4860]: I1211 08:11:55.623087 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:55 crc kubenswrapper[4860]: I1211 08:11:55.623154 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:55 crc kubenswrapper[4860]: I1211 08:11:55.623172 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:55 crc kubenswrapper[4860]: I1211 08:11:55.623195 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:55 crc kubenswrapper[4860]: I1211 08:11:55.623211 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:55Z","lastTransitionTime":"2025-12-11T08:11:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:55 crc kubenswrapper[4860]: I1211 08:11:55.729791 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:55 crc kubenswrapper[4860]: I1211 08:11:55.729864 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:55 crc kubenswrapper[4860]: I1211 08:11:55.729873 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:55 crc kubenswrapper[4860]: I1211 08:11:55.730084 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:55 crc kubenswrapper[4860]: I1211 08:11:55.730101 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:55Z","lastTransitionTime":"2025-12-11T08:11:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:55 crc kubenswrapper[4860]: I1211 08:11:55.833853 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:55 crc kubenswrapper[4860]: I1211 08:11:55.833944 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:55 crc kubenswrapper[4860]: I1211 08:11:55.833967 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:55 crc kubenswrapper[4860]: I1211 08:11:55.833997 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:55 crc kubenswrapper[4860]: I1211 08:11:55.834019 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:55Z","lastTransitionTime":"2025-12-11T08:11:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:55 crc kubenswrapper[4860]: I1211 08:11:55.936781 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:55 crc kubenswrapper[4860]: I1211 08:11:55.936867 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:55 crc kubenswrapper[4860]: I1211 08:11:55.936885 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:55 crc kubenswrapper[4860]: I1211 08:11:55.936910 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:55 crc kubenswrapper[4860]: I1211 08:11:55.936926 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:55Z","lastTransitionTime":"2025-12-11T08:11:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:56 crc kubenswrapper[4860]: I1211 08:11:56.040359 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:56 crc kubenswrapper[4860]: I1211 08:11:56.040429 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:56 crc kubenswrapper[4860]: I1211 08:11:56.040455 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:56 crc kubenswrapper[4860]: I1211 08:11:56.040485 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:56 crc kubenswrapper[4860]: I1211 08:11:56.040505 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:56Z","lastTransitionTime":"2025-12-11T08:11:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:56 crc kubenswrapper[4860]: I1211 08:11:56.143536 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:56 crc kubenswrapper[4860]: I1211 08:11:56.143596 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:56 crc kubenswrapper[4860]: I1211 08:11:56.143613 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:56 crc kubenswrapper[4860]: I1211 08:11:56.143637 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:56 crc kubenswrapper[4860]: I1211 08:11:56.143695 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:56Z","lastTransitionTime":"2025-12-11T08:11:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:56 crc kubenswrapper[4860]: I1211 08:11:56.247043 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:56 crc kubenswrapper[4860]: I1211 08:11:56.247177 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:56 crc kubenswrapper[4860]: I1211 08:11:56.247194 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:56 crc kubenswrapper[4860]: I1211 08:11:56.247220 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:56 crc kubenswrapper[4860]: I1211 08:11:56.247237 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:56Z","lastTransitionTime":"2025-12-11T08:11:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:56 crc kubenswrapper[4860]: I1211 08:11:56.350603 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:56 crc kubenswrapper[4860]: I1211 08:11:56.350711 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:56 crc kubenswrapper[4860]: I1211 08:11:56.350734 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:56 crc kubenswrapper[4860]: I1211 08:11:56.350762 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:56 crc kubenswrapper[4860]: I1211 08:11:56.350784 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:56Z","lastTransitionTime":"2025-12-11T08:11:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:56 crc kubenswrapper[4860]: I1211 08:11:56.456618 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:56 crc kubenswrapper[4860]: I1211 08:11:56.456715 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:56 crc kubenswrapper[4860]: I1211 08:11:56.456735 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:56 crc kubenswrapper[4860]: I1211 08:11:56.456767 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:56 crc kubenswrapper[4860]: I1211 08:11:56.456794 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:56Z","lastTransitionTime":"2025-12-11T08:11:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:56 crc kubenswrapper[4860]: I1211 08:11:56.561014 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:56 crc kubenswrapper[4860]: I1211 08:11:56.561074 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:56 crc kubenswrapper[4860]: I1211 08:11:56.561092 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:56 crc kubenswrapper[4860]: I1211 08:11:56.561120 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:56 crc kubenswrapper[4860]: I1211 08:11:56.561140 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:56Z","lastTransitionTime":"2025-12-11T08:11:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:56 crc kubenswrapper[4860]: I1211 08:11:56.578879 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gjhlj" Dec 11 08:11:56 crc kubenswrapper[4860]: I1211 08:11:56.578896 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:11:56 crc kubenswrapper[4860]: I1211 08:11:56.578959 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:11:56 crc kubenswrapper[4860]: I1211 08:11:56.579036 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:11:56 crc kubenswrapper[4860]: E1211 08:11:56.579226 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gjhlj" podUID="5cfd640a-f3d2-4b28-87e1-5f49f6341e21" Dec 11 08:11:56 crc kubenswrapper[4860]: E1211 08:11:56.579398 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:11:56 crc kubenswrapper[4860]: E1211 08:11:56.579519 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:11:56 crc kubenswrapper[4860]: E1211 08:11:56.579671 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:11:56 crc kubenswrapper[4860]: I1211 08:11:56.664079 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:56 crc kubenswrapper[4860]: I1211 08:11:56.664164 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:56 crc kubenswrapper[4860]: I1211 08:11:56.664198 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:56 crc kubenswrapper[4860]: I1211 08:11:56.664228 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:56 crc kubenswrapper[4860]: I1211 08:11:56.664249 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:56Z","lastTransitionTime":"2025-12-11T08:11:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:56 crc kubenswrapper[4860]: I1211 08:11:56.768030 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:56 crc kubenswrapper[4860]: I1211 08:11:56.768097 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:56 crc kubenswrapper[4860]: I1211 08:11:56.768119 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:56 crc kubenswrapper[4860]: I1211 08:11:56.768147 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:56 crc kubenswrapper[4860]: I1211 08:11:56.768172 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:56Z","lastTransitionTime":"2025-12-11T08:11:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:56 crc kubenswrapper[4860]: I1211 08:11:56.871198 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:56 crc kubenswrapper[4860]: I1211 08:11:56.871271 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:56 crc kubenswrapper[4860]: I1211 08:11:56.871300 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:56 crc kubenswrapper[4860]: I1211 08:11:56.871346 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:56 crc kubenswrapper[4860]: I1211 08:11:56.871368 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:56Z","lastTransitionTime":"2025-12-11T08:11:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:56 crc kubenswrapper[4860]: I1211 08:11:56.974355 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:56 crc kubenswrapper[4860]: I1211 08:11:56.974416 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:56 crc kubenswrapper[4860]: I1211 08:11:56.974438 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:56 crc kubenswrapper[4860]: I1211 08:11:56.974465 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:56 crc kubenswrapper[4860]: I1211 08:11:56.974489 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:56Z","lastTransitionTime":"2025-12-11T08:11:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:57 crc kubenswrapper[4860]: I1211 08:11:57.076753 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:57 crc kubenswrapper[4860]: I1211 08:11:57.076789 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:57 crc kubenswrapper[4860]: I1211 08:11:57.076801 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:57 crc kubenswrapper[4860]: I1211 08:11:57.076817 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:57 crc kubenswrapper[4860]: I1211 08:11:57.076828 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:57Z","lastTransitionTime":"2025-12-11T08:11:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:57 crc kubenswrapper[4860]: I1211 08:11:57.179493 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:57 crc kubenswrapper[4860]: I1211 08:11:57.179547 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:57 crc kubenswrapper[4860]: I1211 08:11:57.179564 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:57 crc kubenswrapper[4860]: I1211 08:11:57.179585 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:57 crc kubenswrapper[4860]: I1211 08:11:57.179600 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:57Z","lastTransitionTime":"2025-12-11T08:11:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:57 crc kubenswrapper[4860]: I1211 08:11:57.282531 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:57 crc kubenswrapper[4860]: I1211 08:11:57.282584 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:57 crc kubenswrapper[4860]: I1211 08:11:57.282599 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:57 crc kubenswrapper[4860]: I1211 08:11:57.282619 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:57 crc kubenswrapper[4860]: I1211 08:11:57.282634 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:57Z","lastTransitionTime":"2025-12-11T08:11:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:57 crc kubenswrapper[4860]: I1211 08:11:57.386080 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:57 crc kubenswrapper[4860]: I1211 08:11:57.386255 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:57 crc kubenswrapper[4860]: I1211 08:11:57.386281 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:57 crc kubenswrapper[4860]: I1211 08:11:57.386312 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:57 crc kubenswrapper[4860]: I1211 08:11:57.386334 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:57Z","lastTransitionTime":"2025-12-11T08:11:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:57 crc kubenswrapper[4860]: I1211 08:11:57.489630 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:57 crc kubenswrapper[4860]: I1211 08:11:57.489727 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:57 crc kubenswrapper[4860]: I1211 08:11:57.489745 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:57 crc kubenswrapper[4860]: I1211 08:11:57.489769 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:57 crc kubenswrapper[4860]: I1211 08:11:57.489786 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:57Z","lastTransitionTime":"2025-12-11T08:11:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:57 crc kubenswrapper[4860]: I1211 08:11:57.592497 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:57 crc kubenswrapper[4860]: I1211 08:11:57.592601 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:57 crc kubenswrapper[4860]: I1211 08:11:57.592627 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:57 crc kubenswrapper[4860]: I1211 08:11:57.592687 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:57 crc kubenswrapper[4860]: I1211 08:11:57.592707 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:57Z","lastTransitionTime":"2025-12-11T08:11:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:57 crc kubenswrapper[4860]: I1211 08:11:57.600851 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-26nc6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1faa733-f408-41fd-b7aa-0cd86283976a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fce5b642b843ed30823c94e0b0c3d8bb1e5d2d93d691f7b8ef901ece03b0ac6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ac45db4f858a1c18a6b87472fc1cea2300f5845f1ad62c22260374995548a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ac45db4f858a1c18a6b87472fc1cea2300f5845f1ad62c22260374995548a0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b35296a5eb17dd2bb5e72c86c1cafdb30b7712cb330264f935d3b4fff05cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c6b35296a5eb17dd2bb5e72c86c1cafdb30b7712cb330264f935d3b4fff05cd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b9e82d6d3a247309ef5bdbf5b3a86ec083add2a138b0b237801e77b09ee96a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b9e82d6d3a247309ef5bdbf5b3a86ec083add2a138b0b237801e77b09ee96a1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://908ee695e78145ef4375c4097d9f1e772502922d02a2fcb4566172a6ad9caaf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://908ee695e78145ef4375c4097d9f1e772502922d02a2fcb4566172a6ad9caaf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afb0a097f3556a8f9ed7cf8f762cb99c082fe0a36b8d8d3038bf188b08fe9a84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afb0a097f3556a8f9ed7cf8f762cb99c082fe0a36b8d8d3038bf188b08fe9a84\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce18f4f06db091e658684fa0bbc858e19b9f03010acf13b3ffccb1a6e3f88021\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce18f4f06db091e658684fa0bbc858e19b9f03010acf13b3ffccb1a6e3f88021\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-26nc6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:57Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:57 crc kubenswrapper[4860]: I1211 08:11:57.635335 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0589a204-c98c-417d-8256-bfe3bf747660\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3333b4346f59b04017e32d61f457ba7ff457166731128f339d7987ae6b56b06d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52bb8ade314afb9494f4e026d0af2aea7b1970d64dc66102f55e257ae4678154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dfcc4370edaae36a5a32b0a9c29290c50f8e4f427e4e147dc3c779beaa8b780\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d856ae95e9077524c518d82d10fb6b93b7fbcaaf2b720ffafe6f7fae58bd7caa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1426084ac580895c42354630910dd8c8e3a48ddd25f22abce1d28dc3de332a63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9984d5d242bb258405efe91fc9caf9e7ea592834445df2e78add9553a921c76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://592dbdb2905039d89bdf741dbd0fd1e1fbd15acbabecdad6f07c758657fa610a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://592dbdb2905039d89bdf741dbd0fd1e1fbd15acbabecdad6f07c758657fa610a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-11T08:11:44Z\\\",\\\"message\\\":\\\"es.lbConfig(nil)\\\\nI1211 08:11:44.095388 6339 services_controller.go:356] Processing sync for service openshift-ingress-canary/ingress-canary for network=default\\\\nI1211 08:11:44.095870 6339 services_controller.go:445] Built service openshift-kube-controller-manager/kube-controller-manager LB template configs for network=default: []services.lbConfig(nil)\\\\nF1211 08:11:44.095867 6339 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:44Z is after 2025-08-24T17:21:41Z]\\\\nI1211 08:11:44.095898 6339 services_controller.go:451] Built service openshift-kube-controller-manager/kube-controller-manager cluster-wide LB for network\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:42Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-xvnqp_openshift-ovn-kubernetes(0589a204-c98c-417d-8256-bfe3bf747660)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb80d9928e8349fc5963227d135355ad7ba820d4fd89877af601d005ca03334b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17cbcb49c06f3a9059cd5a0f05eedb8c785da30a3e502e31bae2f57b3486cb14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17cbcb49c06f3a9059cd5a0f05eedb8c785da30a3e502e31bae2f57b3486cb14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xvnqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:57Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:57 crc kubenswrapper[4860]: I1211 08:11:57.652633 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gsx9l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aba5b6b9-f88d-4763-9562-6b4dec606df9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://735ef012c523ddc952d46a431f27b2582a2c301d9e9da0c697a20ff6e2fb5d86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6zw9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gsx9l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:57Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:57 crc kubenswrapper[4860]: I1211 08:11:57.673689 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9aa0aed-62b3-4ea2-82ae-9a653c1101a5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ad4ba1b769c6cb2b34073798f8e3c25eb0177e6d3059484bf365e6981cab710\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://077aa1dd2fe5189ff60a0f2365e47a6a545cca0de7aa6379370908db4ea5838a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d7c919899b74a2b0fcd9221f3f769a9d013267354b200a282669d342d6413d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5483384f1d2ebc1fefa9706c9e67a3dc5a00ff643f803e3f2707c252e27ad94\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:57Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:57 crc kubenswrapper[4860]: I1211 08:11:57.691397 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31c30642-6e60-41b4-a477-0d802424e0aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79969402f5c33852a26fc839f37a45febf4f2377d77823284fbb01769e5585fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2d4d5de7106175087d91acb1f0c7552014f0c08bfcd65cbbf184d3f8d7826ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-99qgp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:57Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:57 crc kubenswrapper[4860]: I1211 08:11:57.696107 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:57 crc kubenswrapper[4860]: I1211 08:11:57.696174 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:57 crc kubenswrapper[4860]: I1211 08:11:57.696211 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:57 crc kubenswrapper[4860]: I1211 08:11:57.696242 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:57 crc kubenswrapper[4860]: I1211 08:11:57.696267 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:57Z","lastTransitionTime":"2025-12-11T08:11:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:57 crc kubenswrapper[4860]: I1211 08:11:57.725724 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9872be1e-b995-4a00-904b-875dff96ad56\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f74a32a8fefea704cc73ffd1114083dab68aa44dc33f8d9d432d2d24216546c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbe00b46063d3968d59c393b87178e135f982abb7d0f32fa798b212d4b83df6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eedc9988e1a4a3c8cae730dc4a352103a6d665535b825e157d0f87c331b137c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2137fee4b7abc9765f9517d7b3b6ea796601a31d03151bf167d17136ced8b25a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f683e48f3577a68862eb6302383538376e41f49b20b1e0ea3f1b3b143f24872a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5de794187456b3576b969069a3b839d75c7918418a97fa35dff8bb40bb76e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5de794187456b3576b969069a3b839d75c7918418a97fa35dff8bb40bb76e39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08ce42b974fba00c409c0af971f0de351e50f18b2f06f8e722150da3239bb908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08ce42b974fba00c409c0af971f0de351e50f18b2f06f8e722150da3239bb908\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cfac6be03e3cfa0b0f3450b48b67957f2d5172b9c65e27f14c2a260bda5d423f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cfac6be03e3cfa0b0f3450b48b67957f2d5172b9c65e27f14c2a260bda5d423f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:57Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:57 crc kubenswrapper[4860]: I1211 08:11:57.748042 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-pznwl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5310c1fc-66c6-40aa-b1bf-5a59a2410f9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2d6ff00ce322a28607bc3f6db05c5e4fb14acb6777ea315786ef53963ca504a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hkt48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-pznwl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:57Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:57 crc kubenswrapper[4860]: I1211 08:11:57.763923 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-854v6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e6529ba7-2341-43ac-a187-95ab530fe775\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8986538a421f0184f52f288b0e171da695591caf2b10c67ad2dce1073a6e6016\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kh22b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0d4e3f761ad15ae1424204e40b9d3f60188ab4a9047bd4b5ac6a8a8ca1fb385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kh22b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-854v6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:57Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:57 crc kubenswrapper[4860]: I1211 08:11:57.782073 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:57Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:57 crc kubenswrapper[4860]: I1211 08:11:57.800714 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25a6a154da2b74952951ae94efc914ba078ea4bbe5893360b3c84d46be6bf973\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:57Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:57 crc kubenswrapper[4860]: I1211 08:11:57.801468 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:57 crc kubenswrapper[4860]: I1211 08:11:57.801537 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:57 crc kubenswrapper[4860]: I1211 08:11:57.801561 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:57 crc kubenswrapper[4860]: I1211 08:11:57.801589 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:57 crc kubenswrapper[4860]: I1211 08:11:57.801606 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:57Z","lastTransitionTime":"2025-12-11T08:11:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:57 crc kubenswrapper[4860]: I1211 08:11:57.814551 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6xq7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a75ce3c-5e1e-4b9f-bbde-7a3b0cce5eae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe8d4d675567b41728e476351811a93c47b19868ccbadab8c613a4eb8eca8562\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5c6nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6xq7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:57Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:57 crc kubenswrapper[4860]: I1211 08:11:57.832381 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"910ec852-e5d2-4370-806d-77f16a36e630\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://122eb7e14a042027de3aafaa8db89562d80565e7df20f302f872e48d412a26b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1348b83ed7b1e8c25f9dfb1b2d2d035b5adc96dc23854ee90236f06fb155f155\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://867cdac386b0576ee2958a520afd522a2558a8878ea6b5af977b136dc7bba4a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4602fc6635e6188bf4cb75c5c7cf1e71a64fc0c5420626bfafd87beae150722d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4602fc6635e6188bf4cb75c5c7cf1e71a64fc0c5420626bfafd87beae150722d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:57Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:57 crc kubenswrapper[4860]: I1211 08:11:57.850065 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:57Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:57 crc kubenswrapper[4860]: I1211 08:11:57.872661 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d8c2f28628dc2f1ffbdd98a55670e22cf6cdc8cf2c6442e9d1f8e5466b2f12e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://074ea63beac52bd196507ee7cfb1c3d7fc2d37d5ec6faa4d47dde62b922e1c43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:57Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:57 crc kubenswrapper[4860]: I1211 08:11:57.885023 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gjhlj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5cfd640a-f3d2-4b28-87e1-5f49f6341e21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knq7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knq7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:42Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gjhlj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:57Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:57 crc kubenswrapper[4860]: I1211 08:11:57.901521 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ec3758-85f1-4a22-a4ec-c805472607a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e831db1a15976d3967d56479689572c26f2599a766baa0402897a3008428c56f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a498aa90f914e1f0370df01dc5ef174df113a2a16d86b2e3f24b19a1b99b4c0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8caebd2f4dfcbb89c597029a07c335171889761d2d200206987537fb0a228db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34e91f615b76445ff3241a304bb2964fde59de7e8a8dfff6841b2e628a31e86d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2090cab02ad74ddfd6bb57ff50eba35d18f56b0cde1c230fe011c01994b693bc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 08:11:21.171011 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 08:11:21.177674 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-644352688/tls.crt::/tmp/serving-cert-644352688/tls.key\\\\\\\"\\\\nI1211 08:11:26.508606 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 08:11:26.512741 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 08:11:26.512778 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 08:11:26.512839 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 08:11:26.512858 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 08:11:26.521684 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1211 08:11:26.521700 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1211 08:11:26.521722 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:11:26.521732 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:11:26.521738 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 08:11:26.521746 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 08:11:26.521749 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 08:11:26.521753 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1211 08:11:26.524999 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2274fb298ca7d2b37b060b4dbf9b166d39ba77a108b08edeeb4d67196e72cb3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf0aff8a6ab816b6617b1feeca9146d41bb5e8808bf1db96f8d493500c135d85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf0aff8a6ab816b6617b1feeca9146d41bb5e8808bf1db96f8d493500c135d85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:57Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:57 crc kubenswrapper[4860]: I1211 08:11:57.904262 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:57 crc kubenswrapper[4860]: I1211 08:11:57.904307 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:57 crc kubenswrapper[4860]: I1211 08:11:57.904324 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:57 crc kubenswrapper[4860]: I1211 08:11:57.904343 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:57 crc kubenswrapper[4860]: I1211 08:11:57.904357 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:57Z","lastTransitionTime":"2025-12-11T08:11:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:57 crc kubenswrapper[4860]: I1211 08:11:57.920188 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:57Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:57 crc kubenswrapper[4860]: I1211 08:11:57.936314 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a41f35db087418d1be16953f10377042abd3f4488e5ac1bc29ca5dcfc00a37e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:57Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:58 crc kubenswrapper[4860]: I1211 08:11:58.006610 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:58 crc kubenswrapper[4860]: I1211 08:11:58.006733 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:58 crc kubenswrapper[4860]: I1211 08:11:58.006762 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:58 crc kubenswrapper[4860]: I1211 08:11:58.006792 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:58 crc kubenswrapper[4860]: I1211 08:11:58.006814 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:58Z","lastTransitionTime":"2025-12-11T08:11:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:58 crc kubenswrapper[4860]: I1211 08:11:58.109684 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:58 crc kubenswrapper[4860]: I1211 08:11:58.109790 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:58 crc kubenswrapper[4860]: I1211 08:11:58.109808 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:58 crc kubenswrapper[4860]: I1211 08:11:58.109831 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:58 crc kubenswrapper[4860]: I1211 08:11:58.109850 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:58Z","lastTransitionTime":"2025-12-11T08:11:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:58 crc kubenswrapper[4860]: I1211 08:11:58.213171 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:58 crc kubenswrapper[4860]: I1211 08:11:58.213241 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:58 crc kubenswrapper[4860]: I1211 08:11:58.213252 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:58 crc kubenswrapper[4860]: I1211 08:11:58.213269 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:58 crc kubenswrapper[4860]: I1211 08:11:58.213298 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:58Z","lastTransitionTime":"2025-12-11T08:11:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:58 crc kubenswrapper[4860]: I1211 08:11:58.317272 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:58 crc kubenswrapper[4860]: I1211 08:11:58.317328 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:58 crc kubenswrapper[4860]: I1211 08:11:58.317372 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:58 crc kubenswrapper[4860]: I1211 08:11:58.317392 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:58 crc kubenswrapper[4860]: I1211 08:11:58.317405 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:58Z","lastTransitionTime":"2025-12-11T08:11:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:58 crc kubenswrapper[4860]: I1211 08:11:58.352248 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5cfd640a-f3d2-4b28-87e1-5f49f6341e21-metrics-certs\") pod \"network-metrics-daemon-gjhlj\" (UID: \"5cfd640a-f3d2-4b28-87e1-5f49f6341e21\") " pod="openshift-multus/network-metrics-daemon-gjhlj" Dec 11 08:11:58 crc kubenswrapper[4860]: E1211 08:11:58.352712 4860 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 11 08:11:58 crc kubenswrapper[4860]: E1211 08:11:58.352846 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5cfd640a-f3d2-4b28-87e1-5f49f6341e21-metrics-certs podName:5cfd640a-f3d2-4b28-87e1-5f49f6341e21 nodeName:}" failed. No retries permitted until 2025-12-11 08:12:14.352814769 +0000 UTC m=+67.081333864 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5cfd640a-f3d2-4b28-87e1-5f49f6341e21-metrics-certs") pod "network-metrics-daemon-gjhlj" (UID: "5cfd640a-f3d2-4b28-87e1-5f49f6341e21") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 11 08:11:58 crc kubenswrapper[4860]: I1211 08:11:58.420544 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:58 crc kubenswrapper[4860]: I1211 08:11:58.420699 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:58 crc kubenswrapper[4860]: I1211 08:11:58.420728 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:58 crc kubenswrapper[4860]: I1211 08:11:58.420764 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:58 crc kubenswrapper[4860]: I1211 08:11:58.420788 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:58Z","lastTransitionTime":"2025-12-11T08:11:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:58 crc kubenswrapper[4860]: I1211 08:11:58.452979 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:11:58 crc kubenswrapper[4860]: E1211 08:11:58.453276 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:12:30.453238086 +0000 UTC m=+83.181757181 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:11:58 crc kubenswrapper[4860]: I1211 08:11:58.453474 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:11:58 crc kubenswrapper[4860]: E1211 08:11:58.453723 4860 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 11 08:11:58 crc kubenswrapper[4860]: E1211 08:11:58.453804 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-11 08:12:30.453785984 +0000 UTC m=+83.182305079 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 11 08:11:58 crc kubenswrapper[4860]: I1211 08:11:58.524369 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:58 crc kubenswrapper[4860]: I1211 08:11:58.524439 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:58 crc kubenswrapper[4860]: I1211 08:11:58.524449 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:58 crc kubenswrapper[4860]: I1211 08:11:58.524470 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:58 crc kubenswrapper[4860]: I1211 08:11:58.524481 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:58Z","lastTransitionTime":"2025-12-11T08:11:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:58 crc kubenswrapper[4860]: I1211 08:11:58.554399 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:11:58 crc kubenswrapper[4860]: I1211 08:11:58.554480 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:11:58 crc kubenswrapper[4860]: I1211 08:11:58.554535 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:11:58 crc kubenswrapper[4860]: E1211 08:11:58.554726 4860 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 11 08:11:58 crc kubenswrapper[4860]: E1211 08:11:58.554760 4860 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 11 08:11:58 crc kubenswrapper[4860]: E1211 08:11:58.554794 4860 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 11 08:11:58 crc kubenswrapper[4860]: E1211 08:11:58.554858 4860 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 11 08:11:58 crc kubenswrapper[4860]: E1211 08:11:58.554886 4860 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 08:11:58 crc kubenswrapper[4860]: E1211 08:11:58.554896 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-11 08:12:30.554862122 +0000 UTC m=+83.283381227 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 11 08:11:58 crc kubenswrapper[4860]: E1211 08:11:58.554793 4860 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 11 08:11:58 crc kubenswrapper[4860]: E1211 08:11:58.554937 4860 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 08:11:58 crc kubenswrapper[4860]: E1211 08:11:58.554951 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-11 08:12:30.554932524 +0000 UTC m=+83.283451619 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 08:11:58 crc kubenswrapper[4860]: E1211 08:11:58.555047 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-11 08:12:30.555017277 +0000 UTC m=+83.283536542 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 08:11:58 crc kubenswrapper[4860]: I1211 08:11:58.578088 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gjhlj" Dec 11 08:11:58 crc kubenswrapper[4860]: I1211 08:11:58.578136 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:11:58 crc kubenswrapper[4860]: I1211 08:11:58.578107 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:11:58 crc kubenswrapper[4860]: I1211 08:11:58.578107 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:11:58 crc kubenswrapper[4860]: E1211 08:11:58.578292 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gjhlj" podUID="5cfd640a-f3d2-4b28-87e1-5f49f6341e21" Dec 11 08:11:58 crc kubenswrapper[4860]: E1211 08:11:58.578418 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:11:58 crc kubenswrapper[4860]: E1211 08:11:58.578472 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:11:58 crc kubenswrapper[4860]: E1211 08:11:58.579139 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:11:58 crc kubenswrapper[4860]: I1211 08:11:58.580595 4860 scope.go:117] "RemoveContainer" containerID="592dbdb2905039d89bdf741dbd0fd1e1fbd15acbabecdad6f07c758657fa610a" Dec 11 08:11:58 crc kubenswrapper[4860]: I1211 08:11:58.629943 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:58 crc kubenswrapper[4860]: I1211 08:11:58.629992 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:58 crc kubenswrapper[4860]: I1211 08:11:58.630005 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:58 crc kubenswrapper[4860]: I1211 08:11:58.630026 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:58 crc kubenswrapper[4860]: I1211 08:11:58.630049 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:58Z","lastTransitionTime":"2025-12-11T08:11:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:58 crc kubenswrapper[4860]: I1211 08:11:58.732671 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:58 crc kubenswrapper[4860]: I1211 08:11:58.732719 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:58 crc kubenswrapper[4860]: I1211 08:11:58.732736 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:58 crc kubenswrapper[4860]: I1211 08:11:58.732758 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:58 crc kubenswrapper[4860]: I1211 08:11:58.732776 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:58Z","lastTransitionTime":"2025-12-11T08:11:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:58 crc kubenswrapper[4860]: I1211 08:11:58.835294 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:58 crc kubenswrapper[4860]: I1211 08:11:58.835325 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:58 crc kubenswrapper[4860]: I1211 08:11:58.835333 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:58 crc kubenswrapper[4860]: I1211 08:11:58.835348 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:58 crc kubenswrapper[4860]: I1211 08:11:58.835358 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:58Z","lastTransitionTime":"2025-12-11T08:11:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:58 crc kubenswrapper[4860]: I1211 08:11:58.938222 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:58 crc kubenswrapper[4860]: I1211 08:11:58.938267 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:58 crc kubenswrapper[4860]: I1211 08:11:58.938276 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:58 crc kubenswrapper[4860]: I1211 08:11:58.938292 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:58 crc kubenswrapper[4860]: I1211 08:11:58.938303 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:58Z","lastTransitionTime":"2025-12-11T08:11:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:58 crc kubenswrapper[4860]: I1211 08:11:58.975441 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-xvnqp_0589a204-c98c-417d-8256-bfe3bf747660/ovnkube-controller/1.log" Dec 11 08:11:58 crc kubenswrapper[4860]: I1211 08:11:58.979019 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" event={"ID":"0589a204-c98c-417d-8256-bfe3bf747660","Type":"ContainerStarted","Data":"769a74d7b125e7e981e5b0541324a777d021651182d900a57685c9db458395f1"} Dec 11 08:11:58 crc kubenswrapper[4860]: I1211 08:11:58.979851 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" Dec 11 08:11:59 crc kubenswrapper[4860]: I1211 08:11:59.011142 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9872be1e-b995-4a00-904b-875dff96ad56\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f74a32a8fefea704cc73ffd1114083dab68aa44dc33f8d9d432d2d24216546c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbe00b46063d3968d59c393b87178e135f982abb7d0f32fa798b212d4b83df6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eedc9988e1a4a3c8cae730dc4a352103a6d665535b825e157d0f87c331b137c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2137fee4b7abc9765f9517d7b3b6ea796601a31d03151bf167d17136ced8b25a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f683e48f3577a68862eb6302383538376e41f49b20b1e0ea3f1b3b143f24872a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5de794187456b3576b969069a3b839d75c7918418a97fa35dff8bb40bb76e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5de794187456b3576b969069a3b839d75c7918418a97fa35dff8bb40bb76e39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08ce42b974fba00c409c0af971f0de351e50f18b2f06f8e722150da3239bb908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08ce42b974fba00c409c0af971f0de351e50f18b2f06f8e722150da3239bb908\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cfac6be03e3cfa0b0f3450b48b67957f2d5172b9c65e27f14c2a260bda5d423f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cfac6be03e3cfa0b0f3450b48b67957f2d5172b9c65e27f14c2a260bda5d423f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:59Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:59 crc kubenswrapper[4860]: I1211 08:11:59.033127 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-pznwl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5310c1fc-66c6-40aa-b1bf-5a59a2410f9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2d6ff00ce322a28607bc3f6db05c5e4fb14acb6777ea315786ef53963ca504a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hkt48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-pznwl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:59Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:59 crc kubenswrapper[4860]: I1211 08:11:59.041295 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:59 crc kubenswrapper[4860]: I1211 08:11:59.041349 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:59 crc kubenswrapper[4860]: I1211 08:11:59.041366 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:59 crc kubenswrapper[4860]: I1211 08:11:59.041388 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:59 crc kubenswrapper[4860]: I1211 08:11:59.041403 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:59Z","lastTransitionTime":"2025-12-11T08:11:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:59 crc kubenswrapper[4860]: I1211 08:11:59.048926 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-854v6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e6529ba7-2341-43ac-a187-95ab530fe775\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8986538a421f0184f52f288b0e171da695591caf2b10c67ad2dce1073a6e6016\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kh22b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0d4e3f761ad15ae1424204e40b9d3f60188ab4a9047bd4b5ac6a8a8ca1fb385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kh22b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-854v6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:59Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:59 crc kubenswrapper[4860]: I1211 08:11:59.073702 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:59Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:59 crc kubenswrapper[4860]: I1211 08:11:59.093330 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25a6a154da2b74952951ae94efc914ba078ea4bbe5893360b3c84d46be6bf973\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:59Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:59 crc kubenswrapper[4860]: I1211 08:11:59.105752 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6xq7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a75ce3c-5e1e-4b9f-bbde-7a3b0cce5eae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe8d4d675567b41728e476351811a93c47b19868ccbadab8c613a4eb8eca8562\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5c6nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6xq7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:59Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:59 crc kubenswrapper[4860]: I1211 08:11:59.119222 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"910ec852-e5d2-4370-806d-77f16a36e630\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://122eb7e14a042027de3aafaa8db89562d80565e7df20f302f872e48d412a26b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1348b83ed7b1e8c25f9dfb1b2d2d035b5adc96dc23854ee90236f06fb155f155\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://867cdac386b0576ee2958a520afd522a2558a8878ea6b5af977b136dc7bba4a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4602fc6635e6188bf4cb75c5c7cf1e71a64fc0c5420626bfafd87beae150722d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4602fc6635e6188bf4cb75c5c7cf1e71a64fc0c5420626bfafd87beae150722d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:59Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:59 crc kubenswrapper[4860]: I1211 08:11:59.132824 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:59Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:59 crc kubenswrapper[4860]: I1211 08:11:59.144145 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:59 crc kubenswrapper[4860]: I1211 08:11:59.144189 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:59 crc kubenswrapper[4860]: I1211 08:11:59.144202 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:59 crc kubenswrapper[4860]: I1211 08:11:59.144221 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:59 crc kubenswrapper[4860]: I1211 08:11:59.144236 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:59Z","lastTransitionTime":"2025-12-11T08:11:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:59 crc kubenswrapper[4860]: I1211 08:11:59.147111 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d8c2f28628dc2f1ffbdd98a55670e22cf6cdc8cf2c6442e9d1f8e5466b2f12e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://074ea63beac52bd196507ee7cfb1c3d7fc2d37d5ec6faa4d47dde62b922e1c43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:59Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:59 crc kubenswrapper[4860]: I1211 08:11:59.162833 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gjhlj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5cfd640a-f3d2-4b28-87e1-5f49f6341e21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knq7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knq7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:42Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gjhlj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:59Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:59 crc kubenswrapper[4860]: I1211 08:11:59.183628 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ec3758-85f1-4a22-a4ec-c805472607a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e831db1a15976d3967d56479689572c26f2599a766baa0402897a3008428c56f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a498aa90f914e1f0370df01dc5ef174df113a2a16d86b2e3f24b19a1b99b4c0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8caebd2f4dfcbb89c597029a07c335171889761d2d200206987537fb0a228db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34e91f615b76445ff3241a304bb2964fde59de7e8a8dfff6841b2e628a31e86d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2090cab02ad74ddfd6bb57ff50eba35d18f56b0cde1c230fe011c01994b693bc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 08:11:21.171011 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 08:11:21.177674 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-644352688/tls.crt::/tmp/serving-cert-644352688/tls.key\\\\\\\"\\\\nI1211 08:11:26.508606 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 08:11:26.512741 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 08:11:26.512778 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 08:11:26.512839 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 08:11:26.512858 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 08:11:26.521684 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1211 08:11:26.521700 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1211 08:11:26.521722 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:11:26.521732 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:11:26.521738 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 08:11:26.521746 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 08:11:26.521749 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 08:11:26.521753 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1211 08:11:26.524999 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2274fb298ca7d2b37b060b4dbf9b166d39ba77a108b08edeeb4d67196e72cb3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf0aff8a6ab816b6617b1feeca9146d41bb5e8808bf1db96f8d493500c135d85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf0aff8a6ab816b6617b1feeca9146d41bb5e8808bf1db96f8d493500c135d85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:59Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:59 crc kubenswrapper[4860]: I1211 08:11:59.199147 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:59Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:59 crc kubenswrapper[4860]: I1211 08:11:59.215333 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a41f35db087418d1be16953f10377042abd3f4488e5ac1bc29ca5dcfc00a37e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:59Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:59 crc kubenswrapper[4860]: I1211 08:11:59.235028 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-26nc6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1faa733-f408-41fd-b7aa-0cd86283976a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fce5b642b843ed30823c94e0b0c3d8bb1e5d2d93d691f7b8ef901ece03b0ac6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ac45db4f858a1c18a6b87472fc1cea2300f5845f1ad62c22260374995548a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ac45db4f858a1c18a6b87472fc1cea2300f5845f1ad62c22260374995548a0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b35296a5eb17dd2bb5e72c86c1cafdb30b7712cb330264f935d3b4fff05cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c6b35296a5eb17dd2bb5e72c86c1cafdb30b7712cb330264f935d3b4fff05cd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b9e82d6d3a247309ef5bdbf5b3a86ec083add2a138b0b237801e77b09ee96a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b9e82d6d3a247309ef5bdbf5b3a86ec083add2a138b0b237801e77b09ee96a1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://908ee695e78145ef4375c4097d9f1e772502922d02a2fcb4566172a6ad9caaf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://908ee695e78145ef4375c4097d9f1e772502922d02a2fcb4566172a6ad9caaf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afb0a097f3556a8f9ed7cf8f762cb99c082fe0a36b8d8d3038bf188b08fe9a84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afb0a097f3556a8f9ed7cf8f762cb99c082fe0a36b8d8d3038bf188b08fe9a84\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce18f4f06db091e658684fa0bbc858e19b9f03010acf13b3ffccb1a6e3f88021\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce18f4f06db091e658684fa0bbc858e19b9f03010acf13b3ffccb1a6e3f88021\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-26nc6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:59Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:59 crc kubenswrapper[4860]: I1211 08:11:59.247029 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:59 crc kubenswrapper[4860]: I1211 08:11:59.247082 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:59 crc kubenswrapper[4860]: I1211 08:11:59.247095 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:59 crc kubenswrapper[4860]: I1211 08:11:59.247134 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:59 crc kubenswrapper[4860]: I1211 08:11:59.247144 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:59Z","lastTransitionTime":"2025-12-11T08:11:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:59 crc kubenswrapper[4860]: I1211 08:11:59.264226 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0589a204-c98c-417d-8256-bfe3bf747660\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3333b4346f59b04017e32d61f457ba7ff457166731128f339d7987ae6b56b06d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52bb8ade314afb9494f4e026d0af2aea7b1970d64dc66102f55e257ae4678154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dfcc4370edaae36a5a32b0a9c29290c50f8e4f427e4e147dc3c779beaa8b780\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d856ae95e9077524c518d82d10fb6b93b7fbcaaf2b720ffafe6f7fae58bd7caa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1426084ac580895c42354630910dd8c8e3a48ddd25f22abce1d28dc3de332a63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9984d5d242bb258405efe91fc9caf9e7ea592834445df2e78add9553a921c76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://769a74d7b125e7e981e5b0541324a777d021651182d900a57685c9db458395f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://592dbdb2905039d89bdf741dbd0fd1e1fbd15acbabecdad6f07c758657fa610a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-11T08:11:44Z\\\",\\\"message\\\":\\\"es.lbConfig(nil)\\\\nI1211 08:11:44.095388 6339 services_controller.go:356] Processing sync for service openshift-ingress-canary/ingress-canary for network=default\\\\nI1211 08:11:44.095870 6339 services_controller.go:445] Built service openshift-kube-controller-manager/kube-controller-manager LB template configs for network=default: []services.lbConfig(nil)\\\\nF1211 08:11:44.095867 6339 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:44Z is after 2025-08-24T17:21:41Z]\\\\nI1211 08:11:44.095898 6339 services_controller.go:451] Built service openshift-kube-controller-manager/kube-controller-manager cluster-wide LB for network\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:42Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb80d9928e8349fc5963227d135355ad7ba820d4fd89877af601d005ca03334b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17cbcb49c06f3a9059cd5a0f05eedb8c785da30a3e502e31bae2f57b3486cb14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17cbcb49c06f3a9059cd5a0f05eedb8c785da30a3e502e31bae2f57b3486cb14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xvnqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:59Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:59 crc kubenswrapper[4860]: I1211 08:11:59.277227 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gsx9l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aba5b6b9-f88d-4763-9562-6b4dec606df9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://735ef012c523ddc952d46a431f27b2582a2c301d9e9da0c697a20ff6e2fb5d86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6zw9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gsx9l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:59Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:59 crc kubenswrapper[4860]: I1211 08:11:59.297678 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9aa0aed-62b3-4ea2-82ae-9a653c1101a5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ad4ba1b769c6cb2b34073798f8e3c25eb0177e6d3059484bf365e6981cab710\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://077aa1dd2fe5189ff60a0f2365e47a6a545cca0de7aa6379370908db4ea5838a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d7c919899b74a2b0fcd9221f3f769a9d013267354b200a282669d342d6413d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5483384f1d2ebc1fefa9706c9e67a3dc5a00ff643f803e3f2707c252e27ad94\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:59Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:59 crc kubenswrapper[4860]: I1211 08:11:59.312373 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31c30642-6e60-41b4-a477-0d802424e0aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79969402f5c33852a26fc839f37a45febf4f2377d77823284fbb01769e5585fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2d4d5de7106175087d91acb1f0c7552014f0c08bfcd65cbbf184d3f8d7826ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-99qgp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:59Z is after 2025-08-24T17:21:41Z" Dec 11 08:11:59 crc kubenswrapper[4860]: I1211 08:11:59.349899 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:59 crc kubenswrapper[4860]: I1211 08:11:59.349956 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:59 crc kubenswrapper[4860]: I1211 08:11:59.349967 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:59 crc kubenswrapper[4860]: I1211 08:11:59.349993 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:59 crc kubenswrapper[4860]: I1211 08:11:59.350007 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:59Z","lastTransitionTime":"2025-12-11T08:11:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:59 crc kubenswrapper[4860]: I1211 08:11:59.452568 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:59 crc kubenswrapper[4860]: I1211 08:11:59.452607 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:59 crc kubenswrapper[4860]: I1211 08:11:59.452616 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:59 crc kubenswrapper[4860]: I1211 08:11:59.452654 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:59 crc kubenswrapper[4860]: I1211 08:11:59.452667 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:59Z","lastTransitionTime":"2025-12-11T08:11:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:59 crc kubenswrapper[4860]: I1211 08:11:59.554994 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:59 crc kubenswrapper[4860]: I1211 08:11:59.555026 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:59 crc kubenswrapper[4860]: I1211 08:11:59.555034 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:59 crc kubenswrapper[4860]: I1211 08:11:59.555064 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:59 crc kubenswrapper[4860]: I1211 08:11:59.555073 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:59Z","lastTransitionTime":"2025-12-11T08:11:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:59 crc kubenswrapper[4860]: I1211 08:11:59.659120 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:59 crc kubenswrapper[4860]: I1211 08:11:59.659221 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:59 crc kubenswrapper[4860]: I1211 08:11:59.659278 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:59 crc kubenswrapper[4860]: I1211 08:11:59.659310 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:59 crc kubenswrapper[4860]: I1211 08:11:59.659363 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:59Z","lastTransitionTime":"2025-12-11T08:11:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:59 crc kubenswrapper[4860]: I1211 08:11:59.762219 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:59 crc kubenswrapper[4860]: I1211 08:11:59.762264 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:59 crc kubenswrapper[4860]: I1211 08:11:59.762276 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:59 crc kubenswrapper[4860]: I1211 08:11:59.762295 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:59 crc kubenswrapper[4860]: I1211 08:11:59.762309 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:59Z","lastTransitionTime":"2025-12-11T08:11:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:59 crc kubenswrapper[4860]: I1211 08:11:59.866275 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:59 crc kubenswrapper[4860]: I1211 08:11:59.866341 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:59 crc kubenswrapper[4860]: I1211 08:11:59.866357 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:59 crc kubenswrapper[4860]: I1211 08:11:59.866387 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:59 crc kubenswrapper[4860]: I1211 08:11:59.866406 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:59Z","lastTransitionTime":"2025-12-11T08:11:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:59 crc kubenswrapper[4860]: I1211 08:11:59.969974 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:11:59 crc kubenswrapper[4860]: I1211 08:11:59.970062 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:11:59 crc kubenswrapper[4860]: I1211 08:11:59.970086 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:11:59 crc kubenswrapper[4860]: I1211 08:11:59.970120 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:11:59 crc kubenswrapper[4860]: I1211 08:11:59.970145 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:11:59Z","lastTransitionTime":"2025-12-11T08:11:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:11:59 crc kubenswrapper[4860]: I1211 08:11:59.986984 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-xvnqp_0589a204-c98c-417d-8256-bfe3bf747660/ovnkube-controller/2.log" Dec 11 08:11:59 crc kubenswrapper[4860]: I1211 08:11:59.988181 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-xvnqp_0589a204-c98c-417d-8256-bfe3bf747660/ovnkube-controller/1.log" Dec 11 08:11:59 crc kubenswrapper[4860]: I1211 08:11:59.993756 4860 generic.go:334] "Generic (PLEG): container finished" podID="0589a204-c98c-417d-8256-bfe3bf747660" containerID="769a74d7b125e7e981e5b0541324a777d021651182d900a57685c9db458395f1" exitCode=1 Dec 11 08:11:59 crc kubenswrapper[4860]: I1211 08:11:59.993888 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" event={"ID":"0589a204-c98c-417d-8256-bfe3bf747660","Type":"ContainerDied","Data":"769a74d7b125e7e981e5b0541324a777d021651182d900a57685c9db458395f1"} Dec 11 08:11:59 crc kubenswrapper[4860]: I1211 08:11:59.993972 4860 scope.go:117] "RemoveContainer" containerID="592dbdb2905039d89bdf741dbd0fd1e1fbd15acbabecdad6f07c758657fa610a" Dec 11 08:11:59 crc kubenswrapper[4860]: I1211 08:11:59.995012 4860 scope.go:117] "RemoveContainer" containerID="769a74d7b125e7e981e5b0541324a777d021651182d900a57685c9db458395f1" Dec 11 08:11:59 crc kubenswrapper[4860]: E1211 08:11:59.995303 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-xvnqp_openshift-ovn-kubernetes(0589a204-c98c-417d-8256-bfe3bf747660)\"" pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" podUID="0589a204-c98c-417d-8256-bfe3bf747660" Dec 11 08:12:00 crc kubenswrapper[4860]: I1211 08:12:00.013956 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9aa0aed-62b3-4ea2-82ae-9a653c1101a5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ad4ba1b769c6cb2b34073798f8e3c25eb0177e6d3059484bf365e6981cab710\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://077aa1dd2fe5189ff60a0f2365e47a6a545cca0de7aa6379370908db4ea5838a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d7c919899b74a2b0fcd9221f3f769a9d013267354b200a282669d342d6413d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5483384f1d2ebc1fefa9706c9e67a3dc5a00ff643f803e3f2707c252e27ad94\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:00Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:00 crc kubenswrapper[4860]: I1211 08:12:00.030201 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31c30642-6e60-41b4-a477-0d802424e0aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79969402f5c33852a26fc839f37a45febf4f2377d77823284fbb01769e5585fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2d4d5de7106175087d91acb1f0c7552014f0c08bfcd65cbbf184d3f8d7826ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-99qgp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:00Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:00 crc kubenswrapper[4860]: I1211 08:12:00.051356 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-26nc6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1faa733-f408-41fd-b7aa-0cd86283976a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fce5b642b843ed30823c94e0b0c3d8bb1e5d2d93d691f7b8ef901ece03b0ac6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ac45db4f858a1c18a6b87472fc1cea2300f5845f1ad62c22260374995548a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ac45db4f858a1c18a6b87472fc1cea2300f5845f1ad62c22260374995548a0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b35296a5eb17dd2bb5e72c86c1cafdb30b7712cb330264f935d3b4fff05cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c6b35296a5eb17dd2bb5e72c86c1cafdb30b7712cb330264f935d3b4fff05cd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b9e82d6d3a247309ef5bdbf5b3a86ec083add2a138b0b237801e77b09ee96a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b9e82d6d3a247309ef5bdbf5b3a86ec083add2a138b0b237801e77b09ee96a1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://908ee695e78145ef4375c4097d9f1e772502922d02a2fcb4566172a6ad9caaf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://908ee695e78145ef4375c4097d9f1e772502922d02a2fcb4566172a6ad9caaf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afb0a097f3556a8f9ed7cf8f762cb99c082fe0a36b8d8d3038bf188b08fe9a84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afb0a097f3556a8f9ed7cf8f762cb99c082fe0a36b8d8d3038bf188b08fe9a84\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce18f4f06db091e658684fa0bbc858e19b9f03010acf13b3ffccb1a6e3f88021\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce18f4f06db091e658684fa0bbc858e19b9f03010acf13b3ffccb1a6e3f88021\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-26nc6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:00Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:00 crc kubenswrapper[4860]: I1211 08:12:00.080463 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:00 crc kubenswrapper[4860]: I1211 08:12:00.080566 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:00 crc kubenswrapper[4860]: I1211 08:12:00.080590 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:00 crc kubenswrapper[4860]: I1211 08:12:00.080620 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:00 crc kubenswrapper[4860]: I1211 08:12:00.080678 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:00Z","lastTransitionTime":"2025-12-11T08:12:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:00 crc kubenswrapper[4860]: I1211 08:12:00.086913 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0589a204-c98c-417d-8256-bfe3bf747660\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3333b4346f59b04017e32d61f457ba7ff457166731128f339d7987ae6b56b06d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52bb8ade314afb9494f4e026d0af2aea7b1970d64dc66102f55e257ae4678154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dfcc4370edaae36a5a32b0a9c29290c50f8e4f427e4e147dc3c779beaa8b780\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d856ae95e9077524c518d82d10fb6b93b7fbcaaf2b720ffafe6f7fae58bd7caa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1426084ac580895c42354630910dd8c8e3a48ddd25f22abce1d28dc3de332a63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9984d5d242bb258405efe91fc9caf9e7ea592834445df2e78add9553a921c76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://769a74d7b125e7e981e5b0541324a777d021651182d900a57685c9db458395f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://592dbdb2905039d89bdf741dbd0fd1e1fbd15acbabecdad6f07c758657fa610a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-11T08:11:44Z\\\",\\\"message\\\":\\\"es.lbConfig(nil)\\\\nI1211 08:11:44.095388 6339 services_controller.go:356] Processing sync for service openshift-ingress-canary/ingress-canary for network=default\\\\nI1211 08:11:44.095870 6339 services_controller.go:445] Built service openshift-kube-controller-manager/kube-controller-manager LB template configs for network=default: []services.lbConfig(nil)\\\\nF1211 08:11:44.095867 6339 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:11:44Z is after 2025-08-24T17:21:41Z]\\\\nI1211 08:11:44.095898 6339 services_controller.go:451] Built service openshift-kube-controller-manager/kube-controller-manager cluster-wide LB for network\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:42Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://769a74d7b125e7e981e5b0541324a777d021651182d900a57685c9db458395f1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-11T08:11:59Z\\\",\\\"message\\\":\\\"multus-pznwl\\\\nI1211 08:11:59.565441 6486 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1211 08:11:59.565459 6486 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-pznwl in node crc\\\\nI1211 08:11:59.565418 6486 ovn.go:134] Ensuring zone local for Pod openshift-kube-controller-manager/kube-controller-manager-crc in node crc\\\\nI1211 08:11:59.565467 6486 obj_retry.go:386] Retry successful for *v1.Pod openshift-multus/multus-pznwl after 0 failed attempt(s)\\\\nI1211 08:11:59.565474 6486 obj_retry.go:386] Retry successful for *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc after 0 failed attempt(s)\\\\nI1211 08:11:59.565476 6486 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-pznwl\\\\nI1211 08:11:59.565481 6486 default_network_controller.go:776] Recording success event on pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1211 08:11:59.565201 6486 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/network-metrics-daemon-gjhlj\\\\nI1211 08:11:59.565512 6486 ovn.go:134] Ensuring zone local for Pod openshift-multus/network-metrics-daemon-gjhlj in node crc\\\\nF1211 08:11:59.565540 6486 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb80d9928e8349fc5963227d135355ad7ba820d4fd89877af601d005ca03334b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17cbcb49c06f3a9059cd5a0f05eedb8c785da30a3e502e31bae2f57b3486cb14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17cbcb49c06f3a9059cd5a0f05eedb8c785da30a3e502e31bae2f57b3486cb14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xvnqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:00Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:00 crc kubenswrapper[4860]: I1211 08:12:00.103717 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gsx9l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aba5b6b9-f88d-4763-9562-6b4dec606df9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://735ef012c523ddc952d46a431f27b2582a2c301d9e9da0c697a20ff6e2fb5d86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6zw9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gsx9l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:00Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:00 crc kubenswrapper[4860]: I1211 08:12:00.138627 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9872be1e-b995-4a00-904b-875dff96ad56\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f74a32a8fefea704cc73ffd1114083dab68aa44dc33f8d9d432d2d24216546c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbe00b46063d3968d59c393b87178e135f982abb7d0f32fa798b212d4b83df6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eedc9988e1a4a3c8cae730dc4a352103a6d665535b825e157d0f87c331b137c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2137fee4b7abc9765f9517d7b3b6ea796601a31d03151bf167d17136ced8b25a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f683e48f3577a68862eb6302383538376e41f49b20b1e0ea3f1b3b143f24872a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5de794187456b3576b969069a3b839d75c7918418a97fa35dff8bb40bb76e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5de794187456b3576b969069a3b839d75c7918418a97fa35dff8bb40bb76e39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08ce42b974fba00c409c0af971f0de351e50f18b2f06f8e722150da3239bb908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08ce42b974fba00c409c0af971f0de351e50f18b2f06f8e722150da3239bb908\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cfac6be03e3cfa0b0f3450b48b67957f2d5172b9c65e27f14c2a260bda5d423f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cfac6be03e3cfa0b0f3450b48b67957f2d5172b9c65e27f14c2a260bda5d423f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:00Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:00 crc kubenswrapper[4860]: I1211 08:12:00.160953 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-pznwl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5310c1fc-66c6-40aa-b1bf-5a59a2410f9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2d6ff00ce322a28607bc3f6db05c5e4fb14acb6777ea315786ef53963ca504a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hkt48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-pznwl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:00Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:00 crc kubenswrapper[4860]: I1211 08:12:00.176749 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-854v6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e6529ba7-2341-43ac-a187-95ab530fe775\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8986538a421f0184f52f288b0e171da695591caf2b10c67ad2dce1073a6e6016\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kh22b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0d4e3f761ad15ae1424204e40b9d3f60188ab4a9047bd4b5ac6a8a8ca1fb385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kh22b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-854v6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:00Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:00 crc kubenswrapper[4860]: I1211 08:12:00.183068 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:00 crc kubenswrapper[4860]: I1211 08:12:00.183140 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:00 crc kubenswrapper[4860]: I1211 08:12:00.183160 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:00 crc kubenswrapper[4860]: I1211 08:12:00.183187 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:00 crc kubenswrapper[4860]: I1211 08:12:00.183208 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:00Z","lastTransitionTime":"2025-12-11T08:12:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:00 crc kubenswrapper[4860]: I1211 08:12:00.195941 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"910ec852-e5d2-4370-806d-77f16a36e630\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://122eb7e14a042027de3aafaa8db89562d80565e7df20f302f872e48d412a26b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1348b83ed7b1e8c25f9dfb1b2d2d035b5adc96dc23854ee90236f06fb155f155\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://867cdac386b0576ee2958a520afd522a2558a8878ea6b5af977b136dc7bba4a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4602fc6635e6188bf4cb75c5c7cf1e71a64fc0c5420626bfafd87beae150722d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4602fc6635e6188bf4cb75c5c7cf1e71a64fc0c5420626bfafd87beae150722d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:00Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:00 crc kubenswrapper[4860]: I1211 08:12:00.215364 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:00Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:00 crc kubenswrapper[4860]: I1211 08:12:00.233564 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d8c2f28628dc2f1ffbdd98a55670e22cf6cdc8cf2c6442e9d1f8e5466b2f12e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://074ea63beac52bd196507ee7cfb1c3d7fc2d37d5ec6faa4d47dde62b922e1c43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:00Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:00 crc kubenswrapper[4860]: I1211 08:12:00.254612 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:00Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:00 crc kubenswrapper[4860]: I1211 08:12:00.273074 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25a6a154da2b74952951ae94efc914ba078ea4bbe5893360b3c84d46be6bf973\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:00Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:00 crc kubenswrapper[4860]: I1211 08:12:00.286447 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:00 crc kubenswrapper[4860]: I1211 08:12:00.286549 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:00 crc kubenswrapper[4860]: I1211 08:12:00.286578 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:00 crc kubenswrapper[4860]: I1211 08:12:00.286611 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:00 crc kubenswrapper[4860]: I1211 08:12:00.286630 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:00Z","lastTransitionTime":"2025-12-11T08:12:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:00 crc kubenswrapper[4860]: I1211 08:12:00.289263 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6xq7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a75ce3c-5e1e-4b9f-bbde-7a3b0cce5eae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe8d4d675567b41728e476351811a93c47b19868ccbadab8c613a4eb8eca8562\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5c6nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6xq7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:00Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:00 crc kubenswrapper[4860]: I1211 08:12:00.311798 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ec3758-85f1-4a22-a4ec-c805472607a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e831db1a15976d3967d56479689572c26f2599a766baa0402897a3008428c56f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a498aa90f914e1f0370df01dc5ef174df113a2a16d86b2e3f24b19a1b99b4c0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8caebd2f4dfcbb89c597029a07c335171889761d2d200206987537fb0a228db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34e91f615b76445ff3241a304bb2964fde59de7e8a8dfff6841b2e628a31e86d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2090cab02ad74ddfd6bb57ff50eba35d18f56b0cde1c230fe011c01994b693bc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 08:11:21.171011 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 08:11:21.177674 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-644352688/tls.crt::/tmp/serving-cert-644352688/tls.key\\\\\\\"\\\\nI1211 08:11:26.508606 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 08:11:26.512741 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 08:11:26.512778 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 08:11:26.512839 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 08:11:26.512858 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 08:11:26.521684 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1211 08:11:26.521700 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1211 08:11:26.521722 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:11:26.521732 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:11:26.521738 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 08:11:26.521746 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 08:11:26.521749 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 08:11:26.521753 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1211 08:11:26.524999 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2274fb298ca7d2b37b060b4dbf9b166d39ba77a108b08edeeb4d67196e72cb3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf0aff8a6ab816b6617b1feeca9146d41bb5e8808bf1db96f8d493500c135d85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf0aff8a6ab816b6617b1feeca9146d41bb5e8808bf1db96f8d493500c135d85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:00Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:00 crc kubenswrapper[4860]: I1211 08:12:00.331837 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:00Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:00 crc kubenswrapper[4860]: I1211 08:12:00.355876 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a41f35db087418d1be16953f10377042abd3f4488e5ac1bc29ca5dcfc00a37e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:00Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:00 crc kubenswrapper[4860]: I1211 08:12:00.370518 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gjhlj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5cfd640a-f3d2-4b28-87e1-5f49f6341e21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knq7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knq7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:42Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gjhlj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:00Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:00 crc kubenswrapper[4860]: I1211 08:12:00.389842 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:00 crc kubenswrapper[4860]: I1211 08:12:00.389896 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:00 crc kubenswrapper[4860]: I1211 08:12:00.389908 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:00 crc kubenswrapper[4860]: I1211 08:12:00.389926 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:00 crc kubenswrapper[4860]: I1211 08:12:00.389938 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:00Z","lastTransitionTime":"2025-12-11T08:12:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:00 crc kubenswrapper[4860]: I1211 08:12:00.493279 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:00 crc kubenswrapper[4860]: I1211 08:12:00.493347 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:00 crc kubenswrapper[4860]: I1211 08:12:00.493367 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:00 crc kubenswrapper[4860]: I1211 08:12:00.493394 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:00 crc kubenswrapper[4860]: I1211 08:12:00.493411 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:00Z","lastTransitionTime":"2025-12-11T08:12:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:00 crc kubenswrapper[4860]: I1211 08:12:00.578566 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gjhlj" Dec 11 08:12:00 crc kubenswrapper[4860]: I1211 08:12:00.578631 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:12:00 crc kubenswrapper[4860]: I1211 08:12:00.578694 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:12:00 crc kubenswrapper[4860]: I1211 08:12:00.578575 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:12:00 crc kubenswrapper[4860]: E1211 08:12:00.578807 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gjhlj" podUID="5cfd640a-f3d2-4b28-87e1-5f49f6341e21" Dec 11 08:12:00 crc kubenswrapper[4860]: E1211 08:12:00.578958 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:12:00 crc kubenswrapper[4860]: E1211 08:12:00.579061 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:12:00 crc kubenswrapper[4860]: E1211 08:12:00.579497 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:12:00 crc kubenswrapper[4860]: I1211 08:12:00.596382 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:00 crc kubenswrapper[4860]: I1211 08:12:00.596429 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:00 crc kubenswrapper[4860]: I1211 08:12:00.596445 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:00 crc kubenswrapper[4860]: I1211 08:12:00.596469 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:00 crc kubenswrapper[4860]: I1211 08:12:00.596489 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:00Z","lastTransitionTime":"2025-12-11T08:12:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:00 crc kubenswrapper[4860]: I1211 08:12:00.698903 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:00 crc kubenswrapper[4860]: I1211 08:12:00.698960 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:00 crc kubenswrapper[4860]: I1211 08:12:00.698978 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:00 crc kubenswrapper[4860]: I1211 08:12:00.699001 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:00 crc kubenswrapper[4860]: I1211 08:12:00.699019 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:00Z","lastTransitionTime":"2025-12-11T08:12:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:00 crc kubenswrapper[4860]: I1211 08:12:00.802087 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:00 crc kubenswrapper[4860]: I1211 08:12:00.802158 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:00 crc kubenswrapper[4860]: I1211 08:12:00.802176 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:00 crc kubenswrapper[4860]: I1211 08:12:00.802198 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:00 crc kubenswrapper[4860]: I1211 08:12:00.802217 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:00Z","lastTransitionTime":"2025-12-11T08:12:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:00 crc kubenswrapper[4860]: I1211 08:12:00.905614 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:00 crc kubenswrapper[4860]: I1211 08:12:00.905709 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:00 crc kubenswrapper[4860]: I1211 08:12:00.905726 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:00 crc kubenswrapper[4860]: I1211 08:12:00.905748 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:00 crc kubenswrapper[4860]: I1211 08:12:00.905771 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:00Z","lastTransitionTime":"2025-12-11T08:12:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:01 crc kubenswrapper[4860]: I1211 08:12:01.001076 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-xvnqp_0589a204-c98c-417d-8256-bfe3bf747660/ovnkube-controller/2.log" Dec 11 08:12:01 crc kubenswrapper[4860]: I1211 08:12:01.007814 4860 scope.go:117] "RemoveContainer" containerID="769a74d7b125e7e981e5b0541324a777d021651182d900a57685c9db458395f1" Dec 11 08:12:01 crc kubenswrapper[4860]: I1211 08:12:01.007899 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:01 crc kubenswrapper[4860]: I1211 08:12:01.007944 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:01 crc kubenswrapper[4860]: I1211 08:12:01.007967 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:01 crc kubenswrapper[4860]: I1211 08:12:01.007993 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:01 crc kubenswrapper[4860]: I1211 08:12:01.008014 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:01Z","lastTransitionTime":"2025-12-11T08:12:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:01 crc kubenswrapper[4860]: E1211 08:12:01.008468 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-xvnqp_openshift-ovn-kubernetes(0589a204-c98c-417d-8256-bfe3bf747660)\"" pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" podUID="0589a204-c98c-417d-8256-bfe3bf747660" Dec 11 08:12:01 crc kubenswrapper[4860]: I1211 08:12:01.042454 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9872be1e-b995-4a00-904b-875dff96ad56\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f74a32a8fefea704cc73ffd1114083dab68aa44dc33f8d9d432d2d24216546c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbe00b46063d3968d59c393b87178e135f982abb7d0f32fa798b212d4b83df6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eedc9988e1a4a3c8cae730dc4a352103a6d665535b825e157d0f87c331b137c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2137fee4b7abc9765f9517d7b3b6ea796601a31d03151bf167d17136ced8b25a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f683e48f3577a68862eb6302383538376e41f49b20b1e0ea3f1b3b143f24872a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5de794187456b3576b969069a3b839d75c7918418a97fa35dff8bb40bb76e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5de794187456b3576b969069a3b839d75c7918418a97fa35dff8bb40bb76e39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08ce42b974fba00c409c0af971f0de351e50f18b2f06f8e722150da3239bb908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08ce42b974fba00c409c0af971f0de351e50f18b2f06f8e722150da3239bb908\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cfac6be03e3cfa0b0f3450b48b67957f2d5172b9c65e27f14c2a260bda5d423f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cfac6be03e3cfa0b0f3450b48b67957f2d5172b9c65e27f14c2a260bda5d423f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:01Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:01 crc kubenswrapper[4860]: I1211 08:12:01.064610 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-pznwl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5310c1fc-66c6-40aa-b1bf-5a59a2410f9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2d6ff00ce322a28607bc3f6db05c5e4fb14acb6777ea315786ef53963ca504a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hkt48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-pznwl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:01Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:01 crc kubenswrapper[4860]: I1211 08:12:01.079404 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-854v6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e6529ba7-2341-43ac-a187-95ab530fe775\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8986538a421f0184f52f288b0e171da695591caf2b10c67ad2dce1073a6e6016\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kh22b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0d4e3f761ad15ae1424204e40b9d3f60188ab4a9047bd4b5ac6a8a8ca1fb385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kh22b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-854v6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:01Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:01 crc kubenswrapper[4860]: I1211 08:12:01.093992 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:01Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:01 crc kubenswrapper[4860]: I1211 08:12:01.111251 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25a6a154da2b74952951ae94efc914ba078ea4bbe5893360b3c84d46be6bf973\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:01Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:01 crc kubenswrapper[4860]: I1211 08:12:01.111778 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:01 crc kubenswrapper[4860]: I1211 08:12:01.111844 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:01 crc kubenswrapper[4860]: I1211 08:12:01.111857 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:01 crc kubenswrapper[4860]: I1211 08:12:01.111897 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:01 crc kubenswrapper[4860]: I1211 08:12:01.111912 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:01Z","lastTransitionTime":"2025-12-11T08:12:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:01 crc kubenswrapper[4860]: I1211 08:12:01.127729 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6xq7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a75ce3c-5e1e-4b9f-bbde-7a3b0cce5eae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe8d4d675567b41728e476351811a93c47b19868ccbadab8c613a4eb8eca8562\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5c6nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6xq7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:01Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:01 crc kubenswrapper[4860]: I1211 08:12:01.143312 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"910ec852-e5d2-4370-806d-77f16a36e630\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://122eb7e14a042027de3aafaa8db89562d80565e7df20f302f872e48d412a26b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1348b83ed7b1e8c25f9dfb1b2d2d035b5adc96dc23854ee90236f06fb155f155\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://867cdac386b0576ee2958a520afd522a2558a8878ea6b5af977b136dc7bba4a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4602fc6635e6188bf4cb75c5c7cf1e71a64fc0c5420626bfafd87beae150722d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4602fc6635e6188bf4cb75c5c7cf1e71a64fc0c5420626bfafd87beae150722d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:01Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:01 crc kubenswrapper[4860]: I1211 08:12:01.159844 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:01Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:01 crc kubenswrapper[4860]: I1211 08:12:01.175062 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:01 crc kubenswrapper[4860]: I1211 08:12:01.175102 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:01 crc kubenswrapper[4860]: I1211 08:12:01.175113 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:01 crc kubenswrapper[4860]: I1211 08:12:01.175128 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:01 crc kubenswrapper[4860]: I1211 08:12:01.175137 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:01Z","lastTransitionTime":"2025-12-11T08:12:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:01 crc kubenswrapper[4860]: I1211 08:12:01.179744 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d8c2f28628dc2f1ffbdd98a55670e22cf6cdc8cf2c6442e9d1f8e5466b2f12e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://074ea63beac52bd196507ee7cfb1c3d7fc2d37d5ec6faa4d47dde62b922e1c43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:01Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:01 crc kubenswrapper[4860]: E1211 08:12:01.189550 4860 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:12:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:12:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:12:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:12:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"52631a1a-bc27-4dd5-8309-5dc39b5dcc41\\\",\\\"systemUUID\\\":\\\"299d3cd2-e1ad-4f5c-8af4-f12c0a5d76e7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:01Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:01 crc kubenswrapper[4860]: I1211 08:12:01.193667 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:01 crc kubenswrapper[4860]: I1211 08:12:01.193730 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:01 crc kubenswrapper[4860]: I1211 08:12:01.193749 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:01 crc kubenswrapper[4860]: I1211 08:12:01.193777 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:01 crc kubenswrapper[4860]: I1211 08:12:01.193796 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:01Z","lastTransitionTime":"2025-12-11T08:12:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:01 crc kubenswrapper[4860]: I1211 08:12:01.196439 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gjhlj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5cfd640a-f3d2-4b28-87e1-5f49f6341e21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knq7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knq7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:42Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gjhlj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:01Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:01 crc kubenswrapper[4860]: E1211 08:12:01.207689 4860 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:12:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:12:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:12:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:12:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"52631a1a-bc27-4dd5-8309-5dc39b5dcc41\\\",\\\"systemUUID\\\":\\\"299d3cd2-e1ad-4f5c-8af4-f12c0a5d76e7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:01Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:01 crc kubenswrapper[4860]: I1211 08:12:01.210801 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:01 crc kubenswrapper[4860]: I1211 08:12:01.210827 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:01 crc kubenswrapper[4860]: I1211 08:12:01.210836 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:01 crc kubenswrapper[4860]: I1211 08:12:01.210851 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:01 crc kubenswrapper[4860]: I1211 08:12:01.210859 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:01Z","lastTransitionTime":"2025-12-11T08:12:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:01 crc kubenswrapper[4860]: I1211 08:12:01.213313 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ec3758-85f1-4a22-a4ec-c805472607a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e831db1a15976d3967d56479689572c26f2599a766baa0402897a3008428c56f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a498aa90f914e1f0370df01dc5ef174df113a2a16d86b2e3f24b19a1b99b4c0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8caebd2f4dfcbb89c597029a07c335171889761d2d200206987537fb0a228db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34e91f615b76445ff3241a304bb2964fde59de7e8a8dfff6841b2e628a31e86d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2090cab02ad74ddfd6bb57ff50eba35d18f56b0cde1c230fe011c01994b693bc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 08:11:21.171011 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 08:11:21.177674 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-644352688/tls.crt::/tmp/serving-cert-644352688/tls.key\\\\\\\"\\\\nI1211 08:11:26.508606 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 08:11:26.512741 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 08:11:26.512778 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 08:11:26.512839 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 08:11:26.512858 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 08:11:26.521684 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1211 08:11:26.521700 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1211 08:11:26.521722 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:11:26.521732 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:11:26.521738 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 08:11:26.521746 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 08:11:26.521749 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 08:11:26.521753 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1211 08:11:26.524999 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2274fb298ca7d2b37b060b4dbf9b166d39ba77a108b08edeeb4d67196e72cb3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf0aff8a6ab816b6617b1feeca9146d41bb5e8808bf1db96f8d493500c135d85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf0aff8a6ab816b6617b1feeca9146d41bb5e8808bf1db96f8d493500c135d85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:01Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:01 crc kubenswrapper[4860]: E1211 08:12:01.220975 4860 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:12:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:12:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:12:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:12:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"52631a1a-bc27-4dd5-8309-5dc39b5dcc41\\\",\\\"systemUUID\\\":\\\"299d3cd2-e1ad-4f5c-8af4-f12c0a5d76e7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:01Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:01 crc kubenswrapper[4860]: I1211 08:12:01.224090 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:01Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:01 crc kubenswrapper[4860]: I1211 08:12:01.224134 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:01 crc kubenswrapper[4860]: I1211 08:12:01.224159 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:01 crc kubenswrapper[4860]: I1211 08:12:01.224172 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:01 crc kubenswrapper[4860]: I1211 08:12:01.224191 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:01 crc kubenswrapper[4860]: I1211 08:12:01.224206 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:01Z","lastTransitionTime":"2025-12-11T08:12:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:01 crc kubenswrapper[4860]: I1211 08:12:01.235568 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a41f35db087418d1be16953f10377042abd3f4488e5ac1bc29ca5dcfc00a37e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:01Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:01 crc kubenswrapper[4860]: E1211 08:12:01.237997 4860 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:12:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:12:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:12:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:12:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"52631a1a-bc27-4dd5-8309-5dc39b5dcc41\\\",\\\"systemUUID\\\":\\\"299d3cd2-e1ad-4f5c-8af4-f12c0a5d76e7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:01Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:01 crc kubenswrapper[4860]: I1211 08:12:01.244569 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:01 crc kubenswrapper[4860]: I1211 08:12:01.244599 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:01 crc kubenswrapper[4860]: I1211 08:12:01.244609 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:01 crc kubenswrapper[4860]: I1211 08:12:01.244623 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:01 crc kubenswrapper[4860]: I1211 08:12:01.244633 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:01Z","lastTransitionTime":"2025-12-11T08:12:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:01 crc kubenswrapper[4860]: I1211 08:12:01.247537 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-26nc6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1faa733-f408-41fd-b7aa-0cd86283976a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fce5b642b843ed30823c94e0b0c3d8bb1e5d2d93d691f7b8ef901ece03b0ac6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ac45db4f858a1c18a6b87472fc1cea2300f5845f1ad62c22260374995548a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ac45db4f858a1c18a6b87472fc1cea2300f5845f1ad62c22260374995548a0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b35296a5eb17dd2bb5e72c86c1cafdb30b7712cb330264f935d3b4fff05cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c6b35296a5eb17dd2bb5e72c86c1cafdb30b7712cb330264f935d3b4fff05cd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b9e82d6d3a247309ef5bdbf5b3a86ec083add2a138b0b237801e77b09ee96a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b9e82d6d3a247309ef5bdbf5b3a86ec083add2a138b0b237801e77b09ee96a1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://908ee695e78145ef4375c4097d9f1e772502922d02a2fcb4566172a6ad9caaf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://908ee695e78145ef4375c4097d9f1e772502922d02a2fcb4566172a6ad9caaf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afb0a097f3556a8f9ed7cf8f762cb99c082fe0a36b8d8d3038bf188b08fe9a84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afb0a097f3556a8f9ed7cf8f762cb99c082fe0a36b8d8d3038bf188b08fe9a84\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce18f4f06db091e658684fa0bbc858e19b9f03010acf13b3ffccb1a6e3f88021\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce18f4f06db091e658684fa0bbc858e19b9f03010acf13b3ffccb1a6e3f88021\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-26nc6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:01Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:01 crc kubenswrapper[4860]: E1211 08:12:01.254963 4860 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:12:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:12:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:12:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:12:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"52631a1a-bc27-4dd5-8309-5dc39b5dcc41\\\",\\\"systemUUID\\\":\\\"299d3cd2-e1ad-4f5c-8af4-f12c0a5d76e7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:01Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:01 crc kubenswrapper[4860]: E1211 08:12:01.255077 4860 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 11 08:12:01 crc kubenswrapper[4860]: I1211 08:12:01.256426 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:01 crc kubenswrapper[4860]: I1211 08:12:01.256455 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:01 crc kubenswrapper[4860]: I1211 08:12:01.256466 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:01 crc kubenswrapper[4860]: I1211 08:12:01.256483 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:01 crc kubenswrapper[4860]: I1211 08:12:01.256494 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:01Z","lastTransitionTime":"2025-12-11T08:12:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:01 crc kubenswrapper[4860]: I1211 08:12:01.264430 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0589a204-c98c-417d-8256-bfe3bf747660\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3333b4346f59b04017e32d61f457ba7ff457166731128f339d7987ae6b56b06d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52bb8ade314afb9494f4e026d0af2aea7b1970d64dc66102f55e257ae4678154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dfcc4370edaae36a5a32b0a9c29290c50f8e4f427e4e147dc3c779beaa8b780\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d856ae95e9077524c518d82d10fb6b93b7fbcaaf2b720ffafe6f7fae58bd7caa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1426084ac580895c42354630910dd8c8e3a48ddd25f22abce1d28dc3de332a63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9984d5d242bb258405efe91fc9caf9e7ea592834445df2e78add9553a921c76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://769a74d7b125e7e981e5b0541324a777d021651182d900a57685c9db458395f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://769a74d7b125e7e981e5b0541324a777d021651182d900a57685c9db458395f1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-11T08:11:59Z\\\",\\\"message\\\":\\\"multus-pznwl\\\\nI1211 08:11:59.565441 6486 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1211 08:11:59.565459 6486 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-pznwl in node crc\\\\nI1211 08:11:59.565418 6486 ovn.go:134] Ensuring zone local for Pod openshift-kube-controller-manager/kube-controller-manager-crc in node crc\\\\nI1211 08:11:59.565467 6486 obj_retry.go:386] Retry successful for *v1.Pod openshift-multus/multus-pznwl after 0 failed attempt(s)\\\\nI1211 08:11:59.565474 6486 obj_retry.go:386] Retry successful for *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc after 0 failed attempt(s)\\\\nI1211 08:11:59.565476 6486 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-pznwl\\\\nI1211 08:11:59.565481 6486 default_network_controller.go:776] Recording success event on pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1211 08:11:59.565201 6486 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/network-metrics-daemon-gjhlj\\\\nI1211 08:11:59.565512 6486 ovn.go:134] Ensuring zone local for Pod openshift-multus/network-metrics-daemon-gjhlj in node crc\\\\nF1211 08:11:59.565540 6486 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:58Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-xvnqp_openshift-ovn-kubernetes(0589a204-c98c-417d-8256-bfe3bf747660)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb80d9928e8349fc5963227d135355ad7ba820d4fd89877af601d005ca03334b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17cbcb49c06f3a9059cd5a0f05eedb8c785da30a3e502e31bae2f57b3486cb14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17cbcb49c06f3a9059cd5a0f05eedb8c785da30a3e502e31bae2f57b3486cb14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xvnqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:01Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:01 crc kubenswrapper[4860]: I1211 08:12:01.274840 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gsx9l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aba5b6b9-f88d-4763-9562-6b4dec606df9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://735ef012c523ddc952d46a431f27b2582a2c301d9e9da0c697a20ff6e2fb5d86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6zw9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gsx9l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:01Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:01 crc kubenswrapper[4860]: I1211 08:12:01.287657 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9aa0aed-62b3-4ea2-82ae-9a653c1101a5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ad4ba1b769c6cb2b34073798f8e3c25eb0177e6d3059484bf365e6981cab710\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://077aa1dd2fe5189ff60a0f2365e47a6a545cca0de7aa6379370908db4ea5838a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d7c919899b74a2b0fcd9221f3f769a9d013267354b200a282669d342d6413d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5483384f1d2ebc1fefa9706c9e67a3dc5a00ff643f803e3f2707c252e27ad94\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:01Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:01 crc kubenswrapper[4860]: I1211 08:12:01.297104 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31c30642-6e60-41b4-a477-0d802424e0aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79969402f5c33852a26fc839f37a45febf4f2377d77823284fbb01769e5585fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2d4d5de7106175087d91acb1f0c7552014f0c08bfcd65cbbf184d3f8d7826ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-99qgp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:01Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:01 crc kubenswrapper[4860]: I1211 08:12:01.359379 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:01 crc kubenswrapper[4860]: I1211 08:12:01.359440 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:01 crc kubenswrapper[4860]: I1211 08:12:01.359460 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:01 crc kubenswrapper[4860]: I1211 08:12:01.359486 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:01 crc kubenswrapper[4860]: I1211 08:12:01.359504 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:01Z","lastTransitionTime":"2025-12-11T08:12:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:01 crc kubenswrapper[4860]: I1211 08:12:01.462301 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:01 crc kubenswrapper[4860]: I1211 08:12:01.462395 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:01 crc kubenswrapper[4860]: I1211 08:12:01.462414 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:01 crc kubenswrapper[4860]: I1211 08:12:01.462437 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:01 crc kubenswrapper[4860]: I1211 08:12:01.462454 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:01Z","lastTransitionTime":"2025-12-11T08:12:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:01 crc kubenswrapper[4860]: I1211 08:12:01.565161 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:01 crc kubenswrapper[4860]: I1211 08:12:01.565215 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:01 crc kubenswrapper[4860]: I1211 08:12:01.565226 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:01 crc kubenswrapper[4860]: I1211 08:12:01.565243 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:01 crc kubenswrapper[4860]: I1211 08:12:01.565255 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:01Z","lastTransitionTime":"2025-12-11T08:12:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:01 crc kubenswrapper[4860]: I1211 08:12:01.668022 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:01 crc kubenswrapper[4860]: I1211 08:12:01.668442 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:01 crc kubenswrapper[4860]: I1211 08:12:01.668460 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:01 crc kubenswrapper[4860]: I1211 08:12:01.668481 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:01 crc kubenswrapper[4860]: I1211 08:12:01.668497 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:01Z","lastTransitionTime":"2025-12-11T08:12:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:01 crc kubenswrapper[4860]: I1211 08:12:01.771689 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:01 crc kubenswrapper[4860]: I1211 08:12:01.771766 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:01 crc kubenswrapper[4860]: I1211 08:12:01.771787 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:01 crc kubenswrapper[4860]: I1211 08:12:01.771841 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:01 crc kubenswrapper[4860]: I1211 08:12:01.771860 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:01Z","lastTransitionTime":"2025-12-11T08:12:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:01 crc kubenswrapper[4860]: I1211 08:12:01.875069 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:01 crc kubenswrapper[4860]: I1211 08:12:01.875128 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:01 crc kubenswrapper[4860]: I1211 08:12:01.875151 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:01 crc kubenswrapper[4860]: I1211 08:12:01.875179 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:01 crc kubenswrapper[4860]: I1211 08:12:01.875203 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:01Z","lastTransitionTime":"2025-12-11T08:12:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:01 crc kubenswrapper[4860]: I1211 08:12:01.978449 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:01 crc kubenswrapper[4860]: I1211 08:12:01.978518 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:01 crc kubenswrapper[4860]: I1211 08:12:01.978540 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:01 crc kubenswrapper[4860]: I1211 08:12:01.978571 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:01 crc kubenswrapper[4860]: I1211 08:12:01.978598 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:01Z","lastTransitionTime":"2025-12-11T08:12:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:02 crc kubenswrapper[4860]: I1211 08:12:02.081399 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:02 crc kubenswrapper[4860]: I1211 08:12:02.081451 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:02 crc kubenswrapper[4860]: I1211 08:12:02.081467 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:02 crc kubenswrapper[4860]: I1211 08:12:02.081490 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:02 crc kubenswrapper[4860]: I1211 08:12:02.081507 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:02Z","lastTransitionTime":"2025-12-11T08:12:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:02 crc kubenswrapper[4860]: I1211 08:12:02.184314 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:02 crc kubenswrapper[4860]: I1211 08:12:02.184374 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:02 crc kubenswrapper[4860]: I1211 08:12:02.184396 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:02 crc kubenswrapper[4860]: I1211 08:12:02.184423 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:02 crc kubenswrapper[4860]: I1211 08:12:02.184445 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:02Z","lastTransitionTime":"2025-12-11T08:12:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:02 crc kubenswrapper[4860]: I1211 08:12:02.287761 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:02 crc kubenswrapper[4860]: I1211 08:12:02.287840 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:02 crc kubenswrapper[4860]: I1211 08:12:02.287863 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:02 crc kubenswrapper[4860]: I1211 08:12:02.287898 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:02 crc kubenswrapper[4860]: I1211 08:12:02.287920 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:02Z","lastTransitionTime":"2025-12-11T08:12:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:02 crc kubenswrapper[4860]: I1211 08:12:02.391167 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:02 crc kubenswrapper[4860]: I1211 08:12:02.391230 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:02 crc kubenswrapper[4860]: I1211 08:12:02.391247 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:02 crc kubenswrapper[4860]: I1211 08:12:02.391273 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:02 crc kubenswrapper[4860]: I1211 08:12:02.391292 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:02Z","lastTransitionTime":"2025-12-11T08:12:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:02 crc kubenswrapper[4860]: I1211 08:12:02.493875 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:02 crc kubenswrapper[4860]: I1211 08:12:02.493949 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:02 crc kubenswrapper[4860]: I1211 08:12:02.493965 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:02 crc kubenswrapper[4860]: I1211 08:12:02.493987 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:02 crc kubenswrapper[4860]: I1211 08:12:02.494006 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:02Z","lastTransitionTime":"2025-12-11T08:12:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:02 crc kubenswrapper[4860]: I1211 08:12:02.578896 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:12:02 crc kubenswrapper[4860]: I1211 08:12:02.578945 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:12:02 crc kubenswrapper[4860]: I1211 08:12:02.578945 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:12:02 crc kubenswrapper[4860]: I1211 08:12:02.578896 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gjhlj" Dec 11 08:12:02 crc kubenswrapper[4860]: E1211 08:12:02.579148 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:12:02 crc kubenswrapper[4860]: E1211 08:12:02.579509 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:12:02 crc kubenswrapper[4860]: E1211 08:12:02.579705 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:12:02 crc kubenswrapper[4860]: E1211 08:12:02.579806 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gjhlj" podUID="5cfd640a-f3d2-4b28-87e1-5f49f6341e21" Dec 11 08:12:02 crc kubenswrapper[4860]: I1211 08:12:02.597051 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:02 crc kubenswrapper[4860]: I1211 08:12:02.597118 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:02 crc kubenswrapper[4860]: I1211 08:12:02.597142 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:02 crc kubenswrapper[4860]: I1211 08:12:02.597168 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:02 crc kubenswrapper[4860]: I1211 08:12:02.597186 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:02Z","lastTransitionTime":"2025-12-11T08:12:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:02 crc kubenswrapper[4860]: I1211 08:12:02.700829 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:02 crc kubenswrapper[4860]: I1211 08:12:02.700882 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:02 crc kubenswrapper[4860]: I1211 08:12:02.700945 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:02 crc kubenswrapper[4860]: I1211 08:12:02.701045 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:02 crc kubenswrapper[4860]: I1211 08:12:02.701080 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:02Z","lastTransitionTime":"2025-12-11T08:12:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:02 crc kubenswrapper[4860]: I1211 08:12:02.805160 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:02 crc kubenswrapper[4860]: I1211 08:12:02.805230 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:02 crc kubenswrapper[4860]: I1211 08:12:02.805243 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:02 crc kubenswrapper[4860]: I1211 08:12:02.805264 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:02 crc kubenswrapper[4860]: I1211 08:12:02.805297 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:02Z","lastTransitionTime":"2025-12-11T08:12:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:02 crc kubenswrapper[4860]: I1211 08:12:02.908418 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:02 crc kubenswrapper[4860]: I1211 08:12:02.908487 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:02 crc kubenswrapper[4860]: I1211 08:12:02.908509 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:02 crc kubenswrapper[4860]: I1211 08:12:02.908533 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:02 crc kubenswrapper[4860]: I1211 08:12:02.908553 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:02Z","lastTransitionTime":"2025-12-11T08:12:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:03 crc kubenswrapper[4860]: I1211 08:12:03.012477 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:03 crc kubenswrapper[4860]: I1211 08:12:03.012533 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:03 crc kubenswrapper[4860]: I1211 08:12:03.012556 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:03 crc kubenswrapper[4860]: I1211 08:12:03.012585 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:03 crc kubenswrapper[4860]: I1211 08:12:03.012607 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:03Z","lastTransitionTime":"2025-12-11T08:12:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:03 crc kubenswrapper[4860]: I1211 08:12:03.115783 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:03 crc kubenswrapper[4860]: I1211 08:12:03.115849 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:03 crc kubenswrapper[4860]: I1211 08:12:03.115860 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:03 crc kubenswrapper[4860]: I1211 08:12:03.115876 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:03 crc kubenswrapper[4860]: I1211 08:12:03.115906 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:03Z","lastTransitionTime":"2025-12-11T08:12:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:03 crc kubenswrapper[4860]: I1211 08:12:03.219831 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:03 crc kubenswrapper[4860]: I1211 08:12:03.219906 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:03 crc kubenswrapper[4860]: I1211 08:12:03.219927 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:03 crc kubenswrapper[4860]: I1211 08:12:03.219956 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:03 crc kubenswrapper[4860]: I1211 08:12:03.219974 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:03Z","lastTransitionTime":"2025-12-11T08:12:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:03 crc kubenswrapper[4860]: I1211 08:12:03.322969 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:03 crc kubenswrapper[4860]: I1211 08:12:03.323036 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:03 crc kubenswrapper[4860]: I1211 08:12:03.323054 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:03 crc kubenswrapper[4860]: I1211 08:12:03.323078 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:03 crc kubenswrapper[4860]: I1211 08:12:03.323099 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:03Z","lastTransitionTime":"2025-12-11T08:12:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:03 crc kubenswrapper[4860]: I1211 08:12:03.425901 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:03 crc kubenswrapper[4860]: I1211 08:12:03.425984 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:03 crc kubenswrapper[4860]: I1211 08:12:03.426009 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:03 crc kubenswrapper[4860]: I1211 08:12:03.426041 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:03 crc kubenswrapper[4860]: I1211 08:12:03.426066 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:03Z","lastTransitionTime":"2025-12-11T08:12:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:03 crc kubenswrapper[4860]: I1211 08:12:03.528864 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:03 crc kubenswrapper[4860]: I1211 08:12:03.528924 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:03 crc kubenswrapper[4860]: I1211 08:12:03.528942 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:03 crc kubenswrapper[4860]: I1211 08:12:03.528966 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:03 crc kubenswrapper[4860]: I1211 08:12:03.528983 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:03Z","lastTransitionTime":"2025-12-11T08:12:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:03 crc kubenswrapper[4860]: I1211 08:12:03.631936 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:03 crc kubenswrapper[4860]: I1211 08:12:03.632008 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:03 crc kubenswrapper[4860]: I1211 08:12:03.632028 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:03 crc kubenswrapper[4860]: I1211 08:12:03.632052 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:03 crc kubenswrapper[4860]: I1211 08:12:03.632070 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:03Z","lastTransitionTime":"2025-12-11T08:12:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:03 crc kubenswrapper[4860]: I1211 08:12:03.735684 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:03 crc kubenswrapper[4860]: I1211 08:12:03.735752 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:03 crc kubenswrapper[4860]: I1211 08:12:03.735771 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:03 crc kubenswrapper[4860]: I1211 08:12:03.735798 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:03 crc kubenswrapper[4860]: I1211 08:12:03.735816 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:03Z","lastTransitionTime":"2025-12-11T08:12:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:03 crc kubenswrapper[4860]: I1211 08:12:03.838746 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:03 crc kubenswrapper[4860]: I1211 08:12:03.838808 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:03 crc kubenswrapper[4860]: I1211 08:12:03.838821 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:03 crc kubenswrapper[4860]: I1211 08:12:03.838840 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:03 crc kubenswrapper[4860]: I1211 08:12:03.838858 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:03Z","lastTransitionTime":"2025-12-11T08:12:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:03 crc kubenswrapper[4860]: I1211 08:12:03.942560 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:03 crc kubenswrapper[4860]: I1211 08:12:03.942718 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:03 crc kubenswrapper[4860]: I1211 08:12:03.942737 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:03 crc kubenswrapper[4860]: I1211 08:12:03.942762 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:03 crc kubenswrapper[4860]: I1211 08:12:03.942780 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:03Z","lastTransitionTime":"2025-12-11T08:12:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:04 crc kubenswrapper[4860]: I1211 08:12:04.045325 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:04 crc kubenswrapper[4860]: I1211 08:12:04.045397 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:04 crc kubenswrapper[4860]: I1211 08:12:04.045413 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:04 crc kubenswrapper[4860]: I1211 08:12:04.045436 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:04 crc kubenswrapper[4860]: I1211 08:12:04.045449 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:04Z","lastTransitionTime":"2025-12-11T08:12:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:04 crc kubenswrapper[4860]: I1211 08:12:04.147791 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:04 crc kubenswrapper[4860]: I1211 08:12:04.147858 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:04 crc kubenswrapper[4860]: I1211 08:12:04.147876 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:04 crc kubenswrapper[4860]: I1211 08:12:04.147902 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:04 crc kubenswrapper[4860]: I1211 08:12:04.147920 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:04Z","lastTransitionTime":"2025-12-11T08:12:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:04 crc kubenswrapper[4860]: I1211 08:12:04.250834 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:04 crc kubenswrapper[4860]: I1211 08:12:04.250900 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:04 crc kubenswrapper[4860]: I1211 08:12:04.250924 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:04 crc kubenswrapper[4860]: I1211 08:12:04.250955 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:04 crc kubenswrapper[4860]: I1211 08:12:04.250980 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:04Z","lastTransitionTime":"2025-12-11T08:12:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:04 crc kubenswrapper[4860]: I1211 08:12:04.354181 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:04 crc kubenswrapper[4860]: I1211 08:12:04.354235 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:04 crc kubenswrapper[4860]: I1211 08:12:04.354252 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:04 crc kubenswrapper[4860]: I1211 08:12:04.354278 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:04 crc kubenswrapper[4860]: I1211 08:12:04.354295 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:04Z","lastTransitionTime":"2025-12-11T08:12:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:04 crc kubenswrapper[4860]: I1211 08:12:04.457317 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:04 crc kubenswrapper[4860]: I1211 08:12:04.457364 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:04 crc kubenswrapper[4860]: I1211 08:12:04.457372 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:04 crc kubenswrapper[4860]: I1211 08:12:04.457387 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:04 crc kubenswrapper[4860]: I1211 08:12:04.457396 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:04Z","lastTransitionTime":"2025-12-11T08:12:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:04 crc kubenswrapper[4860]: I1211 08:12:04.560474 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:04 crc kubenswrapper[4860]: I1211 08:12:04.560541 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:04 crc kubenswrapper[4860]: I1211 08:12:04.560558 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:04 crc kubenswrapper[4860]: I1211 08:12:04.560585 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:04 crc kubenswrapper[4860]: I1211 08:12:04.560606 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:04Z","lastTransitionTime":"2025-12-11T08:12:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:04 crc kubenswrapper[4860]: I1211 08:12:04.578765 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:12:04 crc kubenswrapper[4860]: I1211 08:12:04.578802 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:12:04 crc kubenswrapper[4860]: I1211 08:12:04.578870 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gjhlj" Dec 11 08:12:04 crc kubenswrapper[4860]: E1211 08:12:04.579035 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:12:04 crc kubenswrapper[4860]: I1211 08:12:04.579251 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:12:04 crc kubenswrapper[4860]: E1211 08:12:04.579358 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:12:04 crc kubenswrapper[4860]: E1211 08:12:04.579605 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:12:04 crc kubenswrapper[4860]: E1211 08:12:04.580344 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gjhlj" podUID="5cfd640a-f3d2-4b28-87e1-5f49f6341e21" Dec 11 08:12:04 crc kubenswrapper[4860]: I1211 08:12:04.663368 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:04 crc kubenswrapper[4860]: I1211 08:12:04.663453 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:04 crc kubenswrapper[4860]: I1211 08:12:04.663513 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:04 crc kubenswrapper[4860]: I1211 08:12:04.663541 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:04 crc kubenswrapper[4860]: I1211 08:12:04.663561 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:04Z","lastTransitionTime":"2025-12-11T08:12:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:04 crc kubenswrapper[4860]: I1211 08:12:04.767049 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:04 crc kubenswrapper[4860]: I1211 08:12:04.767143 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:04 crc kubenswrapper[4860]: I1211 08:12:04.767170 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:04 crc kubenswrapper[4860]: I1211 08:12:04.767203 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:04 crc kubenswrapper[4860]: I1211 08:12:04.767225 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:04Z","lastTransitionTime":"2025-12-11T08:12:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:04 crc kubenswrapper[4860]: I1211 08:12:04.870242 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:04 crc kubenswrapper[4860]: I1211 08:12:04.870285 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:04 crc kubenswrapper[4860]: I1211 08:12:04.870299 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:04 crc kubenswrapper[4860]: I1211 08:12:04.870319 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:04 crc kubenswrapper[4860]: I1211 08:12:04.870329 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:04Z","lastTransitionTime":"2025-12-11T08:12:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:04 crc kubenswrapper[4860]: I1211 08:12:04.974163 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:04 crc kubenswrapper[4860]: I1211 08:12:04.974244 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:04 crc kubenswrapper[4860]: I1211 08:12:04.974270 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:04 crc kubenswrapper[4860]: I1211 08:12:04.974301 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:04 crc kubenswrapper[4860]: I1211 08:12:04.974323 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:04Z","lastTransitionTime":"2025-12-11T08:12:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:05 crc kubenswrapper[4860]: I1211 08:12:05.077808 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:05 crc kubenswrapper[4860]: I1211 08:12:05.077870 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:05 crc kubenswrapper[4860]: I1211 08:12:05.077892 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:05 crc kubenswrapper[4860]: I1211 08:12:05.077921 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:05 crc kubenswrapper[4860]: I1211 08:12:05.077945 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:05Z","lastTransitionTime":"2025-12-11T08:12:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:05 crc kubenswrapper[4860]: I1211 08:12:05.181073 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:05 crc kubenswrapper[4860]: I1211 08:12:05.181118 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:05 crc kubenswrapper[4860]: I1211 08:12:05.181134 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:05 crc kubenswrapper[4860]: I1211 08:12:05.181151 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:05 crc kubenswrapper[4860]: I1211 08:12:05.181165 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:05Z","lastTransitionTime":"2025-12-11T08:12:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:05 crc kubenswrapper[4860]: I1211 08:12:05.284750 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:05 crc kubenswrapper[4860]: I1211 08:12:05.284815 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:05 crc kubenswrapper[4860]: I1211 08:12:05.284831 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:05 crc kubenswrapper[4860]: I1211 08:12:05.284856 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:05 crc kubenswrapper[4860]: I1211 08:12:05.284877 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:05Z","lastTransitionTime":"2025-12-11T08:12:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:05 crc kubenswrapper[4860]: I1211 08:12:05.387966 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:05 crc kubenswrapper[4860]: I1211 08:12:05.388012 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:05 crc kubenswrapper[4860]: I1211 08:12:05.388023 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:05 crc kubenswrapper[4860]: I1211 08:12:05.388039 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:05 crc kubenswrapper[4860]: I1211 08:12:05.388052 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:05Z","lastTransitionTime":"2025-12-11T08:12:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:05 crc kubenswrapper[4860]: I1211 08:12:05.490956 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:05 crc kubenswrapper[4860]: I1211 08:12:05.491029 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:05 crc kubenswrapper[4860]: I1211 08:12:05.491049 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:05 crc kubenswrapper[4860]: I1211 08:12:05.491074 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:05 crc kubenswrapper[4860]: I1211 08:12:05.491095 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:05Z","lastTransitionTime":"2025-12-11T08:12:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:05 crc kubenswrapper[4860]: I1211 08:12:05.594288 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:05 crc kubenswrapper[4860]: I1211 08:12:05.594348 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:05 crc kubenswrapper[4860]: I1211 08:12:05.594365 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:05 crc kubenswrapper[4860]: I1211 08:12:05.594389 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:05 crc kubenswrapper[4860]: I1211 08:12:05.594406 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:05Z","lastTransitionTime":"2025-12-11T08:12:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:05 crc kubenswrapper[4860]: I1211 08:12:05.696859 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:05 crc kubenswrapper[4860]: I1211 08:12:05.696933 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:05 crc kubenswrapper[4860]: I1211 08:12:05.696955 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:05 crc kubenswrapper[4860]: I1211 08:12:05.696988 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:05 crc kubenswrapper[4860]: I1211 08:12:05.697010 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:05Z","lastTransitionTime":"2025-12-11T08:12:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:05 crc kubenswrapper[4860]: I1211 08:12:05.800260 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:05 crc kubenswrapper[4860]: I1211 08:12:05.800299 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:05 crc kubenswrapper[4860]: I1211 08:12:05.800348 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:05 crc kubenswrapper[4860]: I1211 08:12:05.800368 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:05 crc kubenswrapper[4860]: I1211 08:12:05.800382 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:05Z","lastTransitionTime":"2025-12-11T08:12:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:05 crc kubenswrapper[4860]: I1211 08:12:05.904066 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:05 crc kubenswrapper[4860]: I1211 08:12:05.904131 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:05 crc kubenswrapper[4860]: I1211 08:12:05.904154 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:05 crc kubenswrapper[4860]: I1211 08:12:05.904181 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:05 crc kubenswrapper[4860]: I1211 08:12:05.904202 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:05Z","lastTransitionTime":"2025-12-11T08:12:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:06 crc kubenswrapper[4860]: I1211 08:12:06.006935 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:06 crc kubenswrapper[4860]: I1211 08:12:06.006971 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:06 crc kubenswrapper[4860]: I1211 08:12:06.006983 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:06 crc kubenswrapper[4860]: I1211 08:12:06.006999 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:06 crc kubenswrapper[4860]: I1211 08:12:06.007011 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:06Z","lastTransitionTime":"2025-12-11T08:12:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:06 crc kubenswrapper[4860]: I1211 08:12:06.109864 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:06 crc kubenswrapper[4860]: I1211 08:12:06.109974 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:06 crc kubenswrapper[4860]: I1211 08:12:06.109995 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:06 crc kubenswrapper[4860]: I1211 08:12:06.110019 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:06 crc kubenswrapper[4860]: I1211 08:12:06.110036 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:06Z","lastTransitionTime":"2025-12-11T08:12:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:06 crc kubenswrapper[4860]: I1211 08:12:06.213436 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:06 crc kubenswrapper[4860]: I1211 08:12:06.213507 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:06 crc kubenswrapper[4860]: I1211 08:12:06.213526 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:06 crc kubenswrapper[4860]: I1211 08:12:06.213551 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:06 crc kubenswrapper[4860]: I1211 08:12:06.213568 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:06Z","lastTransitionTime":"2025-12-11T08:12:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:06 crc kubenswrapper[4860]: I1211 08:12:06.316734 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:06 crc kubenswrapper[4860]: I1211 08:12:06.316789 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:06 crc kubenswrapper[4860]: I1211 08:12:06.316802 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:06 crc kubenswrapper[4860]: I1211 08:12:06.316822 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:06 crc kubenswrapper[4860]: I1211 08:12:06.316833 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:06Z","lastTransitionTime":"2025-12-11T08:12:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:06 crc kubenswrapper[4860]: I1211 08:12:06.419867 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:06 crc kubenswrapper[4860]: I1211 08:12:06.419935 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:06 crc kubenswrapper[4860]: I1211 08:12:06.419952 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:06 crc kubenswrapper[4860]: I1211 08:12:06.419977 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:06 crc kubenswrapper[4860]: I1211 08:12:06.419997 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:06Z","lastTransitionTime":"2025-12-11T08:12:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:06 crc kubenswrapper[4860]: I1211 08:12:06.522792 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:06 crc kubenswrapper[4860]: I1211 08:12:06.522873 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:06 crc kubenswrapper[4860]: I1211 08:12:06.522898 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:06 crc kubenswrapper[4860]: I1211 08:12:06.522930 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:06 crc kubenswrapper[4860]: I1211 08:12:06.522954 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:06Z","lastTransitionTime":"2025-12-11T08:12:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:06 crc kubenswrapper[4860]: I1211 08:12:06.578948 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gjhlj" Dec 11 08:12:06 crc kubenswrapper[4860]: I1211 08:12:06.579015 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:12:06 crc kubenswrapper[4860]: I1211 08:12:06.578984 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:12:06 crc kubenswrapper[4860]: I1211 08:12:06.578981 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:12:06 crc kubenswrapper[4860]: E1211 08:12:06.579963 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gjhlj" podUID="5cfd640a-f3d2-4b28-87e1-5f49f6341e21" Dec 11 08:12:06 crc kubenswrapper[4860]: E1211 08:12:06.580200 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:12:06 crc kubenswrapper[4860]: E1211 08:12:06.580357 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:12:06 crc kubenswrapper[4860]: E1211 08:12:06.580525 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:12:06 crc kubenswrapper[4860]: I1211 08:12:06.625626 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:06 crc kubenswrapper[4860]: I1211 08:12:06.625717 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:06 crc kubenswrapper[4860]: I1211 08:12:06.625735 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:06 crc kubenswrapper[4860]: I1211 08:12:06.625759 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:06 crc kubenswrapper[4860]: I1211 08:12:06.625780 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:06Z","lastTransitionTime":"2025-12-11T08:12:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:06 crc kubenswrapper[4860]: I1211 08:12:06.728571 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:06 crc kubenswrapper[4860]: I1211 08:12:06.728631 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:06 crc kubenswrapper[4860]: I1211 08:12:06.728662 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:06 crc kubenswrapper[4860]: I1211 08:12:06.728692 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:06 crc kubenswrapper[4860]: I1211 08:12:06.728710 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:06Z","lastTransitionTime":"2025-12-11T08:12:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:06 crc kubenswrapper[4860]: I1211 08:12:06.832625 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:06 crc kubenswrapper[4860]: I1211 08:12:06.832739 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:06 crc kubenswrapper[4860]: I1211 08:12:06.832763 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:06 crc kubenswrapper[4860]: I1211 08:12:06.832798 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:06 crc kubenswrapper[4860]: I1211 08:12:06.832826 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:06Z","lastTransitionTime":"2025-12-11T08:12:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:06 crc kubenswrapper[4860]: I1211 08:12:06.935438 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:06 crc kubenswrapper[4860]: I1211 08:12:06.935490 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:06 crc kubenswrapper[4860]: I1211 08:12:06.935502 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:06 crc kubenswrapper[4860]: I1211 08:12:06.935521 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:06 crc kubenswrapper[4860]: I1211 08:12:06.935534 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:06Z","lastTransitionTime":"2025-12-11T08:12:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:07 crc kubenswrapper[4860]: I1211 08:12:07.038019 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:07 crc kubenswrapper[4860]: I1211 08:12:07.038082 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:07 crc kubenswrapper[4860]: I1211 08:12:07.038161 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:07 crc kubenswrapper[4860]: I1211 08:12:07.038256 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:07 crc kubenswrapper[4860]: I1211 08:12:07.038277 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:07Z","lastTransitionTime":"2025-12-11T08:12:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:07 crc kubenswrapper[4860]: I1211 08:12:07.141911 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:07 crc kubenswrapper[4860]: I1211 08:12:07.141976 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:07 crc kubenswrapper[4860]: I1211 08:12:07.141993 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:07 crc kubenswrapper[4860]: I1211 08:12:07.142020 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:07 crc kubenswrapper[4860]: I1211 08:12:07.142038 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:07Z","lastTransitionTime":"2025-12-11T08:12:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:07 crc kubenswrapper[4860]: I1211 08:12:07.245134 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:07 crc kubenswrapper[4860]: I1211 08:12:07.245323 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:07 crc kubenswrapper[4860]: I1211 08:12:07.245349 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:07 crc kubenswrapper[4860]: I1211 08:12:07.245389 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:07 crc kubenswrapper[4860]: I1211 08:12:07.245413 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:07Z","lastTransitionTime":"2025-12-11T08:12:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:07 crc kubenswrapper[4860]: I1211 08:12:07.348822 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:07 crc kubenswrapper[4860]: I1211 08:12:07.348884 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:07 crc kubenswrapper[4860]: I1211 08:12:07.348904 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:07 crc kubenswrapper[4860]: I1211 08:12:07.348929 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:07 crc kubenswrapper[4860]: I1211 08:12:07.348946 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:07Z","lastTransitionTime":"2025-12-11T08:12:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:07 crc kubenswrapper[4860]: I1211 08:12:07.451822 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:07 crc kubenswrapper[4860]: I1211 08:12:07.451888 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:07 crc kubenswrapper[4860]: I1211 08:12:07.451975 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:07 crc kubenswrapper[4860]: I1211 08:12:07.451999 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:07 crc kubenswrapper[4860]: I1211 08:12:07.452017 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:07Z","lastTransitionTime":"2025-12-11T08:12:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:07 crc kubenswrapper[4860]: I1211 08:12:07.558485 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:07 crc kubenswrapper[4860]: I1211 08:12:07.558612 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:07 crc kubenswrapper[4860]: I1211 08:12:07.558754 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:07 crc kubenswrapper[4860]: I1211 08:12:07.559844 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:07 crc kubenswrapper[4860]: I1211 08:12:07.560054 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:07Z","lastTransitionTime":"2025-12-11T08:12:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:07 crc kubenswrapper[4860]: I1211 08:12:07.594908 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"910ec852-e5d2-4370-806d-77f16a36e630\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://122eb7e14a042027de3aafaa8db89562d80565e7df20f302f872e48d412a26b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1348b83ed7b1e8c25f9dfb1b2d2d035b5adc96dc23854ee90236f06fb155f155\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://867cdac386b0576ee2958a520afd522a2558a8878ea6b5af977b136dc7bba4a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4602fc6635e6188bf4cb75c5c7cf1e71a64fc0c5420626bfafd87beae150722d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4602fc6635e6188bf4cb75c5c7cf1e71a64fc0c5420626bfafd87beae150722d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:07Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:07 crc kubenswrapper[4860]: I1211 08:12:07.615331 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:07Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:07 crc kubenswrapper[4860]: I1211 08:12:07.637846 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d8c2f28628dc2f1ffbdd98a55670e22cf6cdc8cf2c6442e9d1f8e5466b2f12e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://074ea63beac52bd196507ee7cfb1c3d7fc2d37d5ec6faa4d47dde62b922e1c43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:07Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:07 crc kubenswrapper[4860]: I1211 08:12:07.655093 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:07Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:07 crc kubenswrapper[4860]: I1211 08:12:07.663001 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:07 crc kubenswrapper[4860]: I1211 08:12:07.663032 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:07 crc kubenswrapper[4860]: I1211 08:12:07.663044 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:07 crc kubenswrapper[4860]: I1211 08:12:07.663063 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:07 crc kubenswrapper[4860]: I1211 08:12:07.663077 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:07Z","lastTransitionTime":"2025-12-11T08:12:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:07 crc kubenswrapper[4860]: I1211 08:12:07.672177 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25a6a154da2b74952951ae94efc914ba078ea4bbe5893360b3c84d46be6bf973\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:07Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:07 crc kubenswrapper[4860]: I1211 08:12:07.684589 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6xq7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a75ce3c-5e1e-4b9f-bbde-7a3b0cce5eae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe8d4d675567b41728e476351811a93c47b19868ccbadab8c613a4eb8eca8562\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5c6nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6xq7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:07Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:07 crc kubenswrapper[4860]: I1211 08:12:07.702104 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ec3758-85f1-4a22-a4ec-c805472607a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e831db1a15976d3967d56479689572c26f2599a766baa0402897a3008428c56f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a498aa90f914e1f0370df01dc5ef174df113a2a16d86b2e3f24b19a1b99b4c0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8caebd2f4dfcbb89c597029a07c335171889761d2d200206987537fb0a228db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34e91f615b76445ff3241a304bb2964fde59de7e8a8dfff6841b2e628a31e86d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2090cab02ad74ddfd6bb57ff50eba35d18f56b0cde1c230fe011c01994b693bc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 08:11:21.171011 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 08:11:21.177674 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-644352688/tls.crt::/tmp/serving-cert-644352688/tls.key\\\\\\\"\\\\nI1211 08:11:26.508606 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 08:11:26.512741 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 08:11:26.512778 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 08:11:26.512839 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 08:11:26.512858 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 08:11:26.521684 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1211 08:11:26.521700 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1211 08:11:26.521722 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:11:26.521732 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:11:26.521738 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 08:11:26.521746 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 08:11:26.521749 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 08:11:26.521753 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1211 08:11:26.524999 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2274fb298ca7d2b37b060b4dbf9b166d39ba77a108b08edeeb4d67196e72cb3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf0aff8a6ab816b6617b1feeca9146d41bb5e8808bf1db96f8d493500c135d85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf0aff8a6ab816b6617b1feeca9146d41bb5e8808bf1db96f8d493500c135d85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:07Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:07 crc kubenswrapper[4860]: I1211 08:12:07.721097 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:07Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:07 crc kubenswrapper[4860]: I1211 08:12:07.737369 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a41f35db087418d1be16953f10377042abd3f4488e5ac1bc29ca5dcfc00a37e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:07Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:07 crc kubenswrapper[4860]: I1211 08:12:07.750099 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gjhlj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5cfd640a-f3d2-4b28-87e1-5f49f6341e21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knq7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knq7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:42Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gjhlj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:07Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:07 crc kubenswrapper[4860]: I1211 08:12:07.761620 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9aa0aed-62b3-4ea2-82ae-9a653c1101a5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ad4ba1b769c6cb2b34073798f8e3c25eb0177e6d3059484bf365e6981cab710\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://077aa1dd2fe5189ff60a0f2365e47a6a545cca0de7aa6379370908db4ea5838a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d7c919899b74a2b0fcd9221f3f769a9d013267354b200a282669d342d6413d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5483384f1d2ebc1fefa9706c9e67a3dc5a00ff643f803e3f2707c252e27ad94\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:07Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:07 crc kubenswrapper[4860]: I1211 08:12:07.766782 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:07 crc kubenswrapper[4860]: I1211 08:12:07.766824 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:07 crc kubenswrapper[4860]: I1211 08:12:07.766840 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:07 crc kubenswrapper[4860]: I1211 08:12:07.766864 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:07 crc kubenswrapper[4860]: I1211 08:12:07.766879 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:07Z","lastTransitionTime":"2025-12-11T08:12:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:07 crc kubenswrapper[4860]: I1211 08:12:07.775314 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31c30642-6e60-41b4-a477-0d802424e0aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79969402f5c33852a26fc839f37a45febf4f2377d77823284fbb01769e5585fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2d4d5de7106175087d91acb1f0c7552014f0c08bfcd65cbbf184d3f8d7826ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-99qgp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:07Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:07 crc kubenswrapper[4860]: I1211 08:12:07.790966 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-26nc6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1faa733-f408-41fd-b7aa-0cd86283976a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fce5b642b843ed30823c94e0b0c3d8bb1e5d2d93d691f7b8ef901ece03b0ac6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ac45db4f858a1c18a6b87472fc1cea2300f5845f1ad62c22260374995548a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ac45db4f858a1c18a6b87472fc1cea2300f5845f1ad62c22260374995548a0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b35296a5eb17dd2bb5e72c86c1cafdb30b7712cb330264f935d3b4fff05cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c6b35296a5eb17dd2bb5e72c86c1cafdb30b7712cb330264f935d3b4fff05cd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b9e82d6d3a247309ef5bdbf5b3a86ec083add2a138b0b237801e77b09ee96a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b9e82d6d3a247309ef5bdbf5b3a86ec083add2a138b0b237801e77b09ee96a1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://908ee695e78145ef4375c4097d9f1e772502922d02a2fcb4566172a6ad9caaf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://908ee695e78145ef4375c4097d9f1e772502922d02a2fcb4566172a6ad9caaf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afb0a097f3556a8f9ed7cf8f762cb99c082fe0a36b8d8d3038bf188b08fe9a84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afb0a097f3556a8f9ed7cf8f762cb99c082fe0a36b8d8d3038bf188b08fe9a84\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce18f4f06db091e658684fa0bbc858e19b9f03010acf13b3ffccb1a6e3f88021\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce18f4f06db091e658684fa0bbc858e19b9f03010acf13b3ffccb1a6e3f88021\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-26nc6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:07Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:07 crc kubenswrapper[4860]: I1211 08:12:07.817446 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0589a204-c98c-417d-8256-bfe3bf747660\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3333b4346f59b04017e32d61f457ba7ff457166731128f339d7987ae6b56b06d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52bb8ade314afb9494f4e026d0af2aea7b1970d64dc66102f55e257ae4678154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dfcc4370edaae36a5a32b0a9c29290c50f8e4f427e4e147dc3c779beaa8b780\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d856ae95e9077524c518d82d10fb6b93b7fbcaaf2b720ffafe6f7fae58bd7caa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1426084ac580895c42354630910dd8c8e3a48ddd25f22abce1d28dc3de332a63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9984d5d242bb258405efe91fc9caf9e7ea592834445df2e78add9553a921c76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://769a74d7b125e7e981e5b0541324a777d021651182d900a57685c9db458395f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://769a74d7b125e7e981e5b0541324a777d021651182d900a57685c9db458395f1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-11T08:11:59Z\\\",\\\"message\\\":\\\"multus-pznwl\\\\nI1211 08:11:59.565441 6486 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1211 08:11:59.565459 6486 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-pznwl in node crc\\\\nI1211 08:11:59.565418 6486 ovn.go:134] Ensuring zone local for Pod openshift-kube-controller-manager/kube-controller-manager-crc in node crc\\\\nI1211 08:11:59.565467 6486 obj_retry.go:386] Retry successful for *v1.Pod openshift-multus/multus-pznwl after 0 failed attempt(s)\\\\nI1211 08:11:59.565474 6486 obj_retry.go:386] Retry successful for *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc after 0 failed attempt(s)\\\\nI1211 08:11:59.565476 6486 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-pznwl\\\\nI1211 08:11:59.565481 6486 default_network_controller.go:776] Recording success event on pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1211 08:11:59.565201 6486 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/network-metrics-daemon-gjhlj\\\\nI1211 08:11:59.565512 6486 ovn.go:134] Ensuring zone local for Pod openshift-multus/network-metrics-daemon-gjhlj in node crc\\\\nF1211 08:11:59.565540 6486 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:58Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-xvnqp_openshift-ovn-kubernetes(0589a204-c98c-417d-8256-bfe3bf747660)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb80d9928e8349fc5963227d135355ad7ba820d4fd89877af601d005ca03334b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17cbcb49c06f3a9059cd5a0f05eedb8c785da30a3e502e31bae2f57b3486cb14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17cbcb49c06f3a9059cd5a0f05eedb8c785da30a3e502e31bae2f57b3486cb14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xvnqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:07Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:07 crc kubenswrapper[4860]: I1211 08:12:07.829749 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gsx9l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aba5b6b9-f88d-4763-9562-6b4dec606df9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://735ef012c523ddc952d46a431f27b2582a2c301d9e9da0c697a20ff6e2fb5d86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6zw9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gsx9l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:07Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:07 crc kubenswrapper[4860]: I1211 08:12:07.852996 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9872be1e-b995-4a00-904b-875dff96ad56\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f74a32a8fefea704cc73ffd1114083dab68aa44dc33f8d9d432d2d24216546c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbe00b46063d3968d59c393b87178e135f982abb7d0f32fa798b212d4b83df6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eedc9988e1a4a3c8cae730dc4a352103a6d665535b825e157d0f87c331b137c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2137fee4b7abc9765f9517d7b3b6ea796601a31d03151bf167d17136ced8b25a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f683e48f3577a68862eb6302383538376e41f49b20b1e0ea3f1b3b143f24872a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5de794187456b3576b969069a3b839d75c7918418a97fa35dff8bb40bb76e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5de794187456b3576b969069a3b839d75c7918418a97fa35dff8bb40bb76e39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08ce42b974fba00c409c0af971f0de351e50f18b2f06f8e722150da3239bb908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08ce42b974fba00c409c0af971f0de351e50f18b2f06f8e722150da3239bb908\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cfac6be03e3cfa0b0f3450b48b67957f2d5172b9c65e27f14c2a260bda5d423f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cfac6be03e3cfa0b0f3450b48b67957f2d5172b9c65e27f14c2a260bda5d423f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:07Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:07 crc kubenswrapper[4860]: I1211 08:12:07.868950 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:07 crc kubenswrapper[4860]: I1211 08:12:07.869136 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:07 crc kubenswrapper[4860]: I1211 08:12:07.869250 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:07 crc kubenswrapper[4860]: I1211 08:12:07.869332 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:07 crc kubenswrapper[4860]: I1211 08:12:07.869411 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:07Z","lastTransitionTime":"2025-12-11T08:12:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:07 crc kubenswrapper[4860]: I1211 08:12:07.869386 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-pznwl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5310c1fc-66c6-40aa-b1bf-5a59a2410f9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2d6ff00ce322a28607bc3f6db05c5e4fb14acb6777ea315786ef53963ca504a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hkt48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-pznwl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:07Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:07 crc kubenswrapper[4860]: I1211 08:12:07.882057 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-854v6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e6529ba7-2341-43ac-a187-95ab530fe775\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8986538a421f0184f52f288b0e171da695591caf2b10c67ad2dce1073a6e6016\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kh22b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0d4e3f761ad15ae1424204e40b9d3f60188ab4a9047bd4b5ac6a8a8ca1fb385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kh22b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-854v6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:07Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:07 crc kubenswrapper[4860]: I1211 08:12:07.972087 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:07 crc kubenswrapper[4860]: I1211 08:12:07.972150 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:07 crc kubenswrapper[4860]: I1211 08:12:07.972169 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:07 crc kubenswrapper[4860]: I1211 08:12:07.972192 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:07 crc kubenswrapper[4860]: I1211 08:12:07.972212 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:07Z","lastTransitionTime":"2025-12-11T08:12:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:08 crc kubenswrapper[4860]: I1211 08:12:08.075081 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:08 crc kubenswrapper[4860]: I1211 08:12:08.075132 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:08 crc kubenswrapper[4860]: I1211 08:12:08.075146 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:08 crc kubenswrapper[4860]: I1211 08:12:08.075168 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:08 crc kubenswrapper[4860]: I1211 08:12:08.075182 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:08Z","lastTransitionTime":"2025-12-11T08:12:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:08 crc kubenswrapper[4860]: I1211 08:12:08.178266 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:08 crc kubenswrapper[4860]: I1211 08:12:08.178345 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:08 crc kubenswrapper[4860]: I1211 08:12:08.178383 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:08 crc kubenswrapper[4860]: I1211 08:12:08.178414 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:08 crc kubenswrapper[4860]: I1211 08:12:08.178440 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:08Z","lastTransitionTime":"2025-12-11T08:12:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:08 crc kubenswrapper[4860]: I1211 08:12:08.282914 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:08 crc kubenswrapper[4860]: I1211 08:12:08.282993 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:08 crc kubenswrapper[4860]: I1211 08:12:08.283021 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:08 crc kubenswrapper[4860]: I1211 08:12:08.283047 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:08 crc kubenswrapper[4860]: I1211 08:12:08.283065 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:08Z","lastTransitionTime":"2025-12-11T08:12:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:08 crc kubenswrapper[4860]: I1211 08:12:08.385042 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:08 crc kubenswrapper[4860]: I1211 08:12:08.385104 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:08 crc kubenswrapper[4860]: I1211 08:12:08.385121 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:08 crc kubenswrapper[4860]: I1211 08:12:08.385143 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:08 crc kubenswrapper[4860]: I1211 08:12:08.385161 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:08Z","lastTransitionTime":"2025-12-11T08:12:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:08 crc kubenswrapper[4860]: I1211 08:12:08.488260 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:08 crc kubenswrapper[4860]: I1211 08:12:08.488332 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:08 crc kubenswrapper[4860]: I1211 08:12:08.488343 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:08 crc kubenswrapper[4860]: I1211 08:12:08.488365 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:08 crc kubenswrapper[4860]: I1211 08:12:08.488377 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:08Z","lastTransitionTime":"2025-12-11T08:12:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:08 crc kubenswrapper[4860]: I1211 08:12:08.578485 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:12:08 crc kubenswrapper[4860]: E1211 08:12:08.578670 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:12:08 crc kubenswrapper[4860]: I1211 08:12:08.578882 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gjhlj" Dec 11 08:12:08 crc kubenswrapper[4860]: I1211 08:12:08.578873 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:12:08 crc kubenswrapper[4860]: E1211 08:12:08.578961 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gjhlj" podUID="5cfd640a-f3d2-4b28-87e1-5f49f6341e21" Dec 11 08:12:08 crc kubenswrapper[4860]: I1211 08:12:08.578914 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:12:08 crc kubenswrapper[4860]: E1211 08:12:08.579094 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:12:08 crc kubenswrapper[4860]: E1211 08:12:08.579556 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:12:08 crc kubenswrapper[4860]: I1211 08:12:08.591703 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:08 crc kubenswrapper[4860]: I1211 08:12:08.591770 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:08 crc kubenswrapper[4860]: I1211 08:12:08.591787 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:08 crc kubenswrapper[4860]: I1211 08:12:08.591812 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:08 crc kubenswrapper[4860]: I1211 08:12:08.591826 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:08Z","lastTransitionTime":"2025-12-11T08:12:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:08 crc kubenswrapper[4860]: I1211 08:12:08.693808 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:08 crc kubenswrapper[4860]: I1211 08:12:08.693886 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:08 crc kubenswrapper[4860]: I1211 08:12:08.693910 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:08 crc kubenswrapper[4860]: I1211 08:12:08.693940 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:08 crc kubenswrapper[4860]: I1211 08:12:08.693962 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:08Z","lastTransitionTime":"2025-12-11T08:12:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:08 crc kubenswrapper[4860]: I1211 08:12:08.796598 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:08 crc kubenswrapper[4860]: I1211 08:12:08.796841 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:08 crc kubenswrapper[4860]: I1211 08:12:08.796871 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:08 crc kubenswrapper[4860]: I1211 08:12:08.796957 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:08 crc kubenswrapper[4860]: I1211 08:12:08.797033 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:08Z","lastTransitionTime":"2025-12-11T08:12:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:08 crc kubenswrapper[4860]: I1211 08:12:08.899227 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:08 crc kubenswrapper[4860]: I1211 08:12:08.899298 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:08 crc kubenswrapper[4860]: I1211 08:12:08.899317 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:08 crc kubenswrapper[4860]: I1211 08:12:08.899342 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:08 crc kubenswrapper[4860]: I1211 08:12:08.899360 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:08Z","lastTransitionTime":"2025-12-11T08:12:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:09 crc kubenswrapper[4860]: I1211 08:12:09.002989 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:09 crc kubenswrapper[4860]: I1211 08:12:09.003054 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:09 crc kubenswrapper[4860]: I1211 08:12:09.003072 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:09 crc kubenswrapper[4860]: I1211 08:12:09.003096 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:09 crc kubenswrapper[4860]: I1211 08:12:09.003115 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:09Z","lastTransitionTime":"2025-12-11T08:12:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:09 crc kubenswrapper[4860]: I1211 08:12:09.105692 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:09 crc kubenswrapper[4860]: I1211 08:12:09.105771 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:09 crc kubenswrapper[4860]: I1211 08:12:09.105783 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:09 crc kubenswrapper[4860]: I1211 08:12:09.105822 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:09 crc kubenswrapper[4860]: I1211 08:12:09.105837 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:09Z","lastTransitionTime":"2025-12-11T08:12:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:09 crc kubenswrapper[4860]: I1211 08:12:09.208759 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:09 crc kubenswrapper[4860]: I1211 08:12:09.208826 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:09 crc kubenswrapper[4860]: I1211 08:12:09.208843 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:09 crc kubenswrapper[4860]: I1211 08:12:09.208872 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:09 crc kubenswrapper[4860]: I1211 08:12:09.208889 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:09Z","lastTransitionTime":"2025-12-11T08:12:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:09 crc kubenswrapper[4860]: I1211 08:12:09.310732 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:09 crc kubenswrapper[4860]: I1211 08:12:09.310769 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:09 crc kubenswrapper[4860]: I1211 08:12:09.310782 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:09 crc kubenswrapper[4860]: I1211 08:12:09.310799 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:09 crc kubenswrapper[4860]: I1211 08:12:09.310810 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:09Z","lastTransitionTime":"2025-12-11T08:12:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:09 crc kubenswrapper[4860]: I1211 08:12:09.413192 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:09 crc kubenswrapper[4860]: I1211 08:12:09.413234 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:09 crc kubenswrapper[4860]: I1211 08:12:09.413251 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:09 crc kubenswrapper[4860]: I1211 08:12:09.413269 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:09 crc kubenswrapper[4860]: I1211 08:12:09.413283 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:09Z","lastTransitionTime":"2025-12-11T08:12:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:09 crc kubenswrapper[4860]: I1211 08:12:09.515971 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:09 crc kubenswrapper[4860]: I1211 08:12:09.516027 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:09 crc kubenswrapper[4860]: I1211 08:12:09.516038 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:09 crc kubenswrapper[4860]: I1211 08:12:09.516055 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:09 crc kubenswrapper[4860]: I1211 08:12:09.516066 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:09Z","lastTransitionTime":"2025-12-11T08:12:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:09 crc kubenswrapper[4860]: I1211 08:12:09.618021 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:09 crc kubenswrapper[4860]: I1211 08:12:09.618053 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:09 crc kubenswrapper[4860]: I1211 08:12:09.618065 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:09 crc kubenswrapper[4860]: I1211 08:12:09.618080 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:09 crc kubenswrapper[4860]: I1211 08:12:09.618091 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:09Z","lastTransitionTime":"2025-12-11T08:12:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:09 crc kubenswrapper[4860]: I1211 08:12:09.720584 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:09 crc kubenswrapper[4860]: I1211 08:12:09.720630 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:09 crc kubenswrapper[4860]: I1211 08:12:09.720674 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:09 crc kubenswrapper[4860]: I1211 08:12:09.720697 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:09 crc kubenswrapper[4860]: I1211 08:12:09.720711 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:09Z","lastTransitionTime":"2025-12-11T08:12:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:09 crc kubenswrapper[4860]: I1211 08:12:09.823717 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:09 crc kubenswrapper[4860]: I1211 08:12:09.823797 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:09 crc kubenswrapper[4860]: I1211 08:12:09.823815 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:09 crc kubenswrapper[4860]: I1211 08:12:09.823839 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:09 crc kubenswrapper[4860]: I1211 08:12:09.823858 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:09Z","lastTransitionTime":"2025-12-11T08:12:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:09 crc kubenswrapper[4860]: I1211 08:12:09.925896 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:09 crc kubenswrapper[4860]: I1211 08:12:09.925955 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:09 crc kubenswrapper[4860]: I1211 08:12:09.925975 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:09 crc kubenswrapper[4860]: I1211 08:12:09.926004 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:09 crc kubenswrapper[4860]: I1211 08:12:09.926025 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:09Z","lastTransitionTime":"2025-12-11T08:12:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:10 crc kubenswrapper[4860]: I1211 08:12:10.028625 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:10 crc kubenswrapper[4860]: I1211 08:12:10.028725 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:10 crc kubenswrapper[4860]: I1211 08:12:10.028739 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:10 crc kubenswrapper[4860]: I1211 08:12:10.028759 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:10 crc kubenswrapper[4860]: I1211 08:12:10.028771 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:10Z","lastTransitionTime":"2025-12-11T08:12:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:10 crc kubenswrapper[4860]: I1211 08:12:10.131292 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:10 crc kubenswrapper[4860]: I1211 08:12:10.131329 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:10 crc kubenswrapper[4860]: I1211 08:12:10.131341 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:10 crc kubenswrapper[4860]: I1211 08:12:10.131356 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:10 crc kubenswrapper[4860]: I1211 08:12:10.131368 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:10Z","lastTransitionTime":"2025-12-11T08:12:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:10 crc kubenswrapper[4860]: I1211 08:12:10.234926 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:10 crc kubenswrapper[4860]: I1211 08:12:10.234983 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:10 crc kubenswrapper[4860]: I1211 08:12:10.234995 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:10 crc kubenswrapper[4860]: I1211 08:12:10.235015 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:10 crc kubenswrapper[4860]: I1211 08:12:10.235030 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:10Z","lastTransitionTime":"2025-12-11T08:12:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:10 crc kubenswrapper[4860]: I1211 08:12:10.337279 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:10 crc kubenswrapper[4860]: I1211 08:12:10.337357 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:10 crc kubenswrapper[4860]: I1211 08:12:10.337373 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:10 crc kubenswrapper[4860]: I1211 08:12:10.337440 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:10 crc kubenswrapper[4860]: I1211 08:12:10.337456 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:10Z","lastTransitionTime":"2025-12-11T08:12:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:10 crc kubenswrapper[4860]: I1211 08:12:10.441536 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:10 crc kubenswrapper[4860]: I1211 08:12:10.441601 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:10 crc kubenswrapper[4860]: I1211 08:12:10.441618 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:10 crc kubenswrapper[4860]: I1211 08:12:10.441667 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:10 crc kubenswrapper[4860]: I1211 08:12:10.441694 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:10Z","lastTransitionTime":"2025-12-11T08:12:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:10 crc kubenswrapper[4860]: I1211 08:12:10.544543 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:10 crc kubenswrapper[4860]: I1211 08:12:10.544947 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:10 crc kubenswrapper[4860]: I1211 08:12:10.544958 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:10 crc kubenswrapper[4860]: I1211 08:12:10.545006 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:10 crc kubenswrapper[4860]: I1211 08:12:10.545018 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:10Z","lastTransitionTime":"2025-12-11T08:12:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:10 crc kubenswrapper[4860]: I1211 08:12:10.578398 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:12:10 crc kubenswrapper[4860]: I1211 08:12:10.578450 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:12:10 crc kubenswrapper[4860]: I1211 08:12:10.578477 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:12:10 crc kubenswrapper[4860]: I1211 08:12:10.578428 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gjhlj" Dec 11 08:12:10 crc kubenswrapper[4860]: E1211 08:12:10.578568 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:12:10 crc kubenswrapper[4860]: E1211 08:12:10.578711 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:12:10 crc kubenswrapper[4860]: E1211 08:12:10.578797 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gjhlj" podUID="5cfd640a-f3d2-4b28-87e1-5f49f6341e21" Dec 11 08:12:10 crc kubenswrapper[4860]: E1211 08:12:10.578950 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:12:10 crc kubenswrapper[4860]: I1211 08:12:10.647945 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:10 crc kubenswrapper[4860]: I1211 08:12:10.647998 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:10 crc kubenswrapper[4860]: I1211 08:12:10.648009 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:10 crc kubenswrapper[4860]: I1211 08:12:10.648026 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:10 crc kubenswrapper[4860]: I1211 08:12:10.648041 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:10Z","lastTransitionTime":"2025-12-11T08:12:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:10 crc kubenswrapper[4860]: I1211 08:12:10.750463 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:10 crc kubenswrapper[4860]: I1211 08:12:10.750521 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:10 crc kubenswrapper[4860]: I1211 08:12:10.750537 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:10 crc kubenswrapper[4860]: I1211 08:12:10.750561 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:10 crc kubenswrapper[4860]: I1211 08:12:10.750576 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:10Z","lastTransitionTime":"2025-12-11T08:12:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:10 crc kubenswrapper[4860]: I1211 08:12:10.853170 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:10 crc kubenswrapper[4860]: I1211 08:12:10.853245 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:10 crc kubenswrapper[4860]: I1211 08:12:10.853268 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:10 crc kubenswrapper[4860]: I1211 08:12:10.853298 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:10 crc kubenswrapper[4860]: I1211 08:12:10.853322 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:10Z","lastTransitionTime":"2025-12-11T08:12:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:10 crc kubenswrapper[4860]: I1211 08:12:10.955675 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:10 crc kubenswrapper[4860]: I1211 08:12:10.955742 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:10 crc kubenswrapper[4860]: I1211 08:12:10.955759 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:10 crc kubenswrapper[4860]: I1211 08:12:10.955783 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:10 crc kubenswrapper[4860]: I1211 08:12:10.955799 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:10Z","lastTransitionTime":"2025-12-11T08:12:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:11 crc kubenswrapper[4860]: I1211 08:12:11.058220 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:11 crc kubenswrapper[4860]: I1211 08:12:11.058282 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:11 crc kubenswrapper[4860]: I1211 08:12:11.058298 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:11 crc kubenswrapper[4860]: I1211 08:12:11.058315 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:11 crc kubenswrapper[4860]: I1211 08:12:11.058325 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:11Z","lastTransitionTime":"2025-12-11T08:12:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:11 crc kubenswrapper[4860]: I1211 08:12:11.161438 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:11 crc kubenswrapper[4860]: I1211 08:12:11.161505 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:11 crc kubenswrapper[4860]: I1211 08:12:11.161532 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:11 crc kubenswrapper[4860]: I1211 08:12:11.161562 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:11 crc kubenswrapper[4860]: I1211 08:12:11.161587 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:11Z","lastTransitionTime":"2025-12-11T08:12:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:11 crc kubenswrapper[4860]: I1211 08:12:11.264028 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:11 crc kubenswrapper[4860]: I1211 08:12:11.264090 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:11 crc kubenswrapper[4860]: I1211 08:12:11.264106 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:11 crc kubenswrapper[4860]: I1211 08:12:11.264133 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:11 crc kubenswrapper[4860]: I1211 08:12:11.264150 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:11Z","lastTransitionTime":"2025-12-11T08:12:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:11 crc kubenswrapper[4860]: I1211 08:12:11.367907 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:11 crc kubenswrapper[4860]: I1211 08:12:11.367996 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:11 crc kubenswrapper[4860]: I1211 08:12:11.368022 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:11 crc kubenswrapper[4860]: I1211 08:12:11.368059 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:11 crc kubenswrapper[4860]: I1211 08:12:11.368086 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:11Z","lastTransitionTime":"2025-12-11T08:12:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:11 crc kubenswrapper[4860]: I1211 08:12:11.386525 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:11 crc kubenswrapper[4860]: I1211 08:12:11.386578 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:11 crc kubenswrapper[4860]: I1211 08:12:11.386590 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:11 crc kubenswrapper[4860]: I1211 08:12:11.386608 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:11 crc kubenswrapper[4860]: I1211 08:12:11.386619 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:11Z","lastTransitionTime":"2025-12-11T08:12:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:11 crc kubenswrapper[4860]: E1211 08:12:11.398014 4860 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:12:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:12:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:12:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:12:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"52631a1a-bc27-4dd5-8309-5dc39b5dcc41\\\",\\\"systemUUID\\\":\\\"299d3cd2-e1ad-4f5c-8af4-f12c0a5d76e7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:11Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:11 crc kubenswrapper[4860]: I1211 08:12:11.402032 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:11 crc kubenswrapper[4860]: I1211 08:12:11.402088 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:11 crc kubenswrapper[4860]: I1211 08:12:11.402102 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:11 crc kubenswrapper[4860]: I1211 08:12:11.402126 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:11 crc kubenswrapper[4860]: I1211 08:12:11.402137 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:11Z","lastTransitionTime":"2025-12-11T08:12:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:11 crc kubenswrapper[4860]: E1211 08:12:11.415583 4860 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:12:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:12:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:12:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:12:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"52631a1a-bc27-4dd5-8309-5dc39b5dcc41\\\",\\\"systemUUID\\\":\\\"299d3cd2-e1ad-4f5c-8af4-f12c0a5d76e7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:11Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:11 crc kubenswrapper[4860]: I1211 08:12:11.419613 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:11 crc kubenswrapper[4860]: I1211 08:12:11.419687 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:11 crc kubenswrapper[4860]: I1211 08:12:11.419699 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:11 crc kubenswrapper[4860]: I1211 08:12:11.419725 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:11 crc kubenswrapper[4860]: I1211 08:12:11.419738 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:11Z","lastTransitionTime":"2025-12-11T08:12:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:11 crc kubenswrapper[4860]: E1211 08:12:11.433742 4860 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:12:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:12:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:12:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:12:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"52631a1a-bc27-4dd5-8309-5dc39b5dcc41\\\",\\\"systemUUID\\\":\\\"299d3cd2-e1ad-4f5c-8af4-f12c0a5d76e7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:11Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:11 crc kubenswrapper[4860]: I1211 08:12:11.437954 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:11 crc kubenswrapper[4860]: I1211 08:12:11.437989 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:11 crc kubenswrapper[4860]: I1211 08:12:11.438004 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:11 crc kubenswrapper[4860]: I1211 08:12:11.438028 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:11 crc kubenswrapper[4860]: I1211 08:12:11.438045 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:11Z","lastTransitionTime":"2025-12-11T08:12:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:11 crc kubenswrapper[4860]: E1211 08:12:11.454770 4860 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:12:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:12:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:12:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:12:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"52631a1a-bc27-4dd5-8309-5dc39b5dcc41\\\",\\\"systemUUID\\\":\\\"299d3cd2-e1ad-4f5c-8af4-f12c0a5d76e7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:11Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:11 crc kubenswrapper[4860]: I1211 08:12:11.458678 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:11 crc kubenswrapper[4860]: I1211 08:12:11.458719 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:11 crc kubenswrapper[4860]: I1211 08:12:11.458731 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:11 crc kubenswrapper[4860]: I1211 08:12:11.458756 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:11 crc kubenswrapper[4860]: I1211 08:12:11.458772 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:11Z","lastTransitionTime":"2025-12-11T08:12:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:11 crc kubenswrapper[4860]: E1211 08:12:11.472346 4860 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:12:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:12:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:12:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:12:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"52631a1a-bc27-4dd5-8309-5dc39b5dcc41\\\",\\\"systemUUID\\\":\\\"299d3cd2-e1ad-4f5c-8af4-f12c0a5d76e7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:11Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:11 crc kubenswrapper[4860]: E1211 08:12:11.472479 4860 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 11 08:12:11 crc kubenswrapper[4860]: I1211 08:12:11.474295 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:11 crc kubenswrapper[4860]: I1211 08:12:11.474337 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:11 crc kubenswrapper[4860]: I1211 08:12:11.474346 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:11 crc kubenswrapper[4860]: I1211 08:12:11.474366 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:11 crc kubenswrapper[4860]: I1211 08:12:11.474376 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:11Z","lastTransitionTime":"2025-12-11T08:12:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:11 crc kubenswrapper[4860]: I1211 08:12:11.576864 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:11 crc kubenswrapper[4860]: I1211 08:12:11.576916 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:11 crc kubenswrapper[4860]: I1211 08:12:11.576927 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:11 crc kubenswrapper[4860]: I1211 08:12:11.576946 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:11 crc kubenswrapper[4860]: I1211 08:12:11.576962 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:11Z","lastTransitionTime":"2025-12-11T08:12:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:11 crc kubenswrapper[4860]: I1211 08:12:11.680223 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:11 crc kubenswrapper[4860]: I1211 08:12:11.680274 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:11 crc kubenswrapper[4860]: I1211 08:12:11.680286 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:11 crc kubenswrapper[4860]: I1211 08:12:11.680302 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:11 crc kubenswrapper[4860]: I1211 08:12:11.680313 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:11Z","lastTransitionTime":"2025-12-11T08:12:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:11 crc kubenswrapper[4860]: I1211 08:12:11.783242 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:11 crc kubenswrapper[4860]: I1211 08:12:11.783289 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:11 crc kubenswrapper[4860]: I1211 08:12:11.783302 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:11 crc kubenswrapper[4860]: I1211 08:12:11.783320 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:11 crc kubenswrapper[4860]: I1211 08:12:11.783336 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:11Z","lastTransitionTime":"2025-12-11T08:12:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:11 crc kubenswrapper[4860]: I1211 08:12:11.886888 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:11 crc kubenswrapper[4860]: I1211 08:12:11.887000 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:11 crc kubenswrapper[4860]: I1211 08:12:11.887013 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:11 crc kubenswrapper[4860]: I1211 08:12:11.887033 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:11 crc kubenswrapper[4860]: I1211 08:12:11.887048 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:11Z","lastTransitionTime":"2025-12-11T08:12:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:11 crc kubenswrapper[4860]: I1211 08:12:11.989847 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:11 crc kubenswrapper[4860]: I1211 08:12:11.989899 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:11 crc kubenswrapper[4860]: I1211 08:12:11.989913 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:11 crc kubenswrapper[4860]: I1211 08:12:11.989962 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:11 crc kubenswrapper[4860]: I1211 08:12:11.989977 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:11Z","lastTransitionTime":"2025-12-11T08:12:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:12 crc kubenswrapper[4860]: I1211 08:12:12.093056 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:12 crc kubenswrapper[4860]: I1211 08:12:12.093117 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:12 crc kubenswrapper[4860]: I1211 08:12:12.093129 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:12 crc kubenswrapper[4860]: I1211 08:12:12.093150 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:12 crc kubenswrapper[4860]: I1211 08:12:12.093163 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:12Z","lastTransitionTime":"2025-12-11T08:12:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:12 crc kubenswrapper[4860]: I1211 08:12:12.195835 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:12 crc kubenswrapper[4860]: I1211 08:12:12.195900 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:12 crc kubenswrapper[4860]: I1211 08:12:12.195913 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:12 crc kubenswrapper[4860]: I1211 08:12:12.195938 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:12 crc kubenswrapper[4860]: I1211 08:12:12.195959 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:12Z","lastTransitionTime":"2025-12-11T08:12:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:12 crc kubenswrapper[4860]: I1211 08:12:12.299284 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:12 crc kubenswrapper[4860]: I1211 08:12:12.299343 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:12 crc kubenswrapper[4860]: I1211 08:12:12.299352 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:12 crc kubenswrapper[4860]: I1211 08:12:12.299372 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:12 crc kubenswrapper[4860]: I1211 08:12:12.299382 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:12Z","lastTransitionTime":"2025-12-11T08:12:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:12 crc kubenswrapper[4860]: I1211 08:12:12.401824 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:12 crc kubenswrapper[4860]: I1211 08:12:12.401896 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:12 crc kubenswrapper[4860]: I1211 08:12:12.401914 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:12 crc kubenswrapper[4860]: I1211 08:12:12.401937 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:12 crc kubenswrapper[4860]: I1211 08:12:12.401954 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:12Z","lastTransitionTime":"2025-12-11T08:12:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:12 crc kubenswrapper[4860]: I1211 08:12:12.504446 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:12 crc kubenswrapper[4860]: I1211 08:12:12.504487 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:12 crc kubenswrapper[4860]: I1211 08:12:12.504496 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:12 crc kubenswrapper[4860]: I1211 08:12:12.504510 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:12 crc kubenswrapper[4860]: I1211 08:12:12.504519 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:12Z","lastTransitionTime":"2025-12-11T08:12:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:12 crc kubenswrapper[4860]: I1211 08:12:12.578480 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:12:12 crc kubenswrapper[4860]: I1211 08:12:12.578510 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:12:12 crc kubenswrapper[4860]: I1211 08:12:12.578496 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gjhlj" Dec 11 08:12:12 crc kubenswrapper[4860]: I1211 08:12:12.578489 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:12:12 crc kubenswrapper[4860]: E1211 08:12:12.578636 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:12:12 crc kubenswrapper[4860]: E1211 08:12:12.578748 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:12:12 crc kubenswrapper[4860]: E1211 08:12:12.578884 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:12:12 crc kubenswrapper[4860]: E1211 08:12:12.579035 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gjhlj" podUID="5cfd640a-f3d2-4b28-87e1-5f49f6341e21" Dec 11 08:12:12 crc kubenswrapper[4860]: I1211 08:12:12.607577 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:12 crc kubenswrapper[4860]: I1211 08:12:12.607635 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:12 crc kubenswrapper[4860]: I1211 08:12:12.607685 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:12 crc kubenswrapper[4860]: I1211 08:12:12.607710 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:12 crc kubenswrapper[4860]: I1211 08:12:12.607726 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:12Z","lastTransitionTime":"2025-12-11T08:12:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:12 crc kubenswrapper[4860]: I1211 08:12:12.709834 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:12 crc kubenswrapper[4860]: I1211 08:12:12.709875 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:12 crc kubenswrapper[4860]: I1211 08:12:12.709885 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:12 crc kubenswrapper[4860]: I1211 08:12:12.709901 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:12 crc kubenswrapper[4860]: I1211 08:12:12.709912 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:12Z","lastTransitionTime":"2025-12-11T08:12:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:12 crc kubenswrapper[4860]: I1211 08:12:12.811762 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:12 crc kubenswrapper[4860]: I1211 08:12:12.811819 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:12 crc kubenswrapper[4860]: I1211 08:12:12.811836 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:12 crc kubenswrapper[4860]: I1211 08:12:12.811860 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:12 crc kubenswrapper[4860]: I1211 08:12:12.811881 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:12Z","lastTransitionTime":"2025-12-11T08:12:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:12 crc kubenswrapper[4860]: I1211 08:12:12.913987 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:12 crc kubenswrapper[4860]: I1211 08:12:12.914046 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:12 crc kubenswrapper[4860]: I1211 08:12:12.914064 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:12 crc kubenswrapper[4860]: I1211 08:12:12.914088 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:12 crc kubenswrapper[4860]: I1211 08:12:12.914105 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:12Z","lastTransitionTime":"2025-12-11T08:12:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:13 crc kubenswrapper[4860]: I1211 08:12:13.016968 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:13 crc kubenswrapper[4860]: I1211 08:12:13.017019 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:13 crc kubenswrapper[4860]: I1211 08:12:13.017030 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:13 crc kubenswrapper[4860]: I1211 08:12:13.017049 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:13 crc kubenswrapper[4860]: I1211 08:12:13.017062 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:13Z","lastTransitionTime":"2025-12-11T08:12:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:13 crc kubenswrapper[4860]: I1211 08:12:13.119700 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:13 crc kubenswrapper[4860]: I1211 08:12:13.119740 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:13 crc kubenswrapper[4860]: I1211 08:12:13.119749 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:13 crc kubenswrapper[4860]: I1211 08:12:13.119763 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:13 crc kubenswrapper[4860]: I1211 08:12:13.119775 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:13Z","lastTransitionTime":"2025-12-11T08:12:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:13 crc kubenswrapper[4860]: I1211 08:12:13.222307 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:13 crc kubenswrapper[4860]: I1211 08:12:13.222352 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:13 crc kubenswrapper[4860]: I1211 08:12:13.222362 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:13 crc kubenswrapper[4860]: I1211 08:12:13.222379 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:13 crc kubenswrapper[4860]: I1211 08:12:13.222390 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:13Z","lastTransitionTime":"2025-12-11T08:12:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:13 crc kubenswrapper[4860]: I1211 08:12:13.324715 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:13 crc kubenswrapper[4860]: I1211 08:12:13.324792 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:13 crc kubenswrapper[4860]: I1211 08:12:13.324810 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:13 crc kubenswrapper[4860]: I1211 08:12:13.324834 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:13 crc kubenswrapper[4860]: I1211 08:12:13.324852 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:13Z","lastTransitionTime":"2025-12-11T08:12:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:13 crc kubenswrapper[4860]: I1211 08:12:13.427003 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:13 crc kubenswrapper[4860]: I1211 08:12:13.427046 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:13 crc kubenswrapper[4860]: I1211 08:12:13.427060 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:13 crc kubenswrapper[4860]: I1211 08:12:13.427075 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:13 crc kubenswrapper[4860]: I1211 08:12:13.427085 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:13Z","lastTransitionTime":"2025-12-11T08:12:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:13 crc kubenswrapper[4860]: I1211 08:12:13.529733 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:13 crc kubenswrapper[4860]: I1211 08:12:13.529778 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:13 crc kubenswrapper[4860]: I1211 08:12:13.529790 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:13 crc kubenswrapper[4860]: I1211 08:12:13.529813 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:13 crc kubenswrapper[4860]: I1211 08:12:13.529825 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:13Z","lastTransitionTime":"2025-12-11T08:12:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:13 crc kubenswrapper[4860]: I1211 08:12:13.579817 4860 scope.go:117] "RemoveContainer" containerID="769a74d7b125e7e981e5b0541324a777d021651182d900a57685c9db458395f1" Dec 11 08:12:13 crc kubenswrapper[4860]: E1211 08:12:13.580011 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-xvnqp_openshift-ovn-kubernetes(0589a204-c98c-417d-8256-bfe3bf747660)\"" pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" podUID="0589a204-c98c-417d-8256-bfe3bf747660" Dec 11 08:12:13 crc kubenswrapper[4860]: I1211 08:12:13.631999 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:13 crc kubenswrapper[4860]: I1211 08:12:13.632047 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:13 crc kubenswrapper[4860]: I1211 08:12:13.632058 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:13 crc kubenswrapper[4860]: I1211 08:12:13.632074 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:13 crc kubenswrapper[4860]: I1211 08:12:13.632085 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:13Z","lastTransitionTime":"2025-12-11T08:12:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:13 crc kubenswrapper[4860]: I1211 08:12:13.734513 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:13 crc kubenswrapper[4860]: I1211 08:12:13.734569 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:13 crc kubenswrapper[4860]: I1211 08:12:13.734586 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:13 crc kubenswrapper[4860]: I1211 08:12:13.734607 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:13 crc kubenswrapper[4860]: I1211 08:12:13.734622 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:13Z","lastTransitionTime":"2025-12-11T08:12:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:13 crc kubenswrapper[4860]: I1211 08:12:13.837922 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:13 crc kubenswrapper[4860]: I1211 08:12:13.837967 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:13 crc kubenswrapper[4860]: I1211 08:12:13.837978 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:13 crc kubenswrapper[4860]: I1211 08:12:13.837995 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:13 crc kubenswrapper[4860]: I1211 08:12:13.838006 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:13Z","lastTransitionTime":"2025-12-11T08:12:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:13 crc kubenswrapper[4860]: I1211 08:12:13.940694 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:13 crc kubenswrapper[4860]: I1211 08:12:13.940750 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:13 crc kubenswrapper[4860]: I1211 08:12:13.940768 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:13 crc kubenswrapper[4860]: I1211 08:12:13.940794 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:13 crc kubenswrapper[4860]: I1211 08:12:13.940813 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:13Z","lastTransitionTime":"2025-12-11T08:12:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:14 crc kubenswrapper[4860]: I1211 08:12:14.043945 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:14 crc kubenswrapper[4860]: I1211 08:12:14.043989 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:14 crc kubenswrapper[4860]: I1211 08:12:14.043999 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:14 crc kubenswrapper[4860]: I1211 08:12:14.044015 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:14 crc kubenswrapper[4860]: I1211 08:12:14.044025 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:14Z","lastTransitionTime":"2025-12-11T08:12:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:14 crc kubenswrapper[4860]: I1211 08:12:14.147101 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:14 crc kubenswrapper[4860]: I1211 08:12:14.147150 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:14 crc kubenswrapper[4860]: I1211 08:12:14.147169 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:14 crc kubenswrapper[4860]: I1211 08:12:14.147194 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:14 crc kubenswrapper[4860]: I1211 08:12:14.147212 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:14Z","lastTransitionTime":"2025-12-11T08:12:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:14 crc kubenswrapper[4860]: I1211 08:12:14.249778 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:14 crc kubenswrapper[4860]: I1211 08:12:14.249823 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:14 crc kubenswrapper[4860]: I1211 08:12:14.249836 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:14 crc kubenswrapper[4860]: I1211 08:12:14.249859 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:14 crc kubenswrapper[4860]: I1211 08:12:14.249875 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:14Z","lastTransitionTime":"2025-12-11T08:12:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:14 crc kubenswrapper[4860]: I1211 08:12:14.352550 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:14 crc kubenswrapper[4860]: I1211 08:12:14.352594 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:14 crc kubenswrapper[4860]: I1211 08:12:14.352607 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:14 crc kubenswrapper[4860]: I1211 08:12:14.352626 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:14 crc kubenswrapper[4860]: I1211 08:12:14.352637 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:14Z","lastTransitionTime":"2025-12-11T08:12:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:14 crc kubenswrapper[4860]: I1211 08:12:14.441752 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5cfd640a-f3d2-4b28-87e1-5f49f6341e21-metrics-certs\") pod \"network-metrics-daemon-gjhlj\" (UID: \"5cfd640a-f3d2-4b28-87e1-5f49f6341e21\") " pod="openshift-multus/network-metrics-daemon-gjhlj" Dec 11 08:12:14 crc kubenswrapper[4860]: E1211 08:12:14.441970 4860 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 11 08:12:14 crc kubenswrapper[4860]: E1211 08:12:14.442077 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5cfd640a-f3d2-4b28-87e1-5f49f6341e21-metrics-certs podName:5cfd640a-f3d2-4b28-87e1-5f49f6341e21 nodeName:}" failed. No retries permitted until 2025-12-11 08:12:46.442054921 +0000 UTC m=+99.170573986 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5cfd640a-f3d2-4b28-87e1-5f49f6341e21-metrics-certs") pod "network-metrics-daemon-gjhlj" (UID: "5cfd640a-f3d2-4b28-87e1-5f49f6341e21") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 11 08:12:14 crc kubenswrapper[4860]: I1211 08:12:14.455250 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:14 crc kubenswrapper[4860]: I1211 08:12:14.455279 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:14 crc kubenswrapper[4860]: I1211 08:12:14.455287 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:14 crc kubenswrapper[4860]: I1211 08:12:14.455299 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:14 crc kubenswrapper[4860]: I1211 08:12:14.455308 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:14Z","lastTransitionTime":"2025-12-11T08:12:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:14 crc kubenswrapper[4860]: I1211 08:12:14.558107 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:14 crc kubenswrapper[4860]: I1211 08:12:14.558150 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:14 crc kubenswrapper[4860]: I1211 08:12:14.558167 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:14 crc kubenswrapper[4860]: I1211 08:12:14.558190 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:14 crc kubenswrapper[4860]: I1211 08:12:14.558207 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:14Z","lastTransitionTime":"2025-12-11T08:12:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:14 crc kubenswrapper[4860]: I1211 08:12:14.578238 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gjhlj" Dec 11 08:12:14 crc kubenswrapper[4860]: E1211 08:12:14.578434 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gjhlj" podUID="5cfd640a-f3d2-4b28-87e1-5f49f6341e21" Dec 11 08:12:14 crc kubenswrapper[4860]: I1211 08:12:14.578506 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:12:14 crc kubenswrapper[4860]: E1211 08:12:14.578587 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:12:14 crc kubenswrapper[4860]: I1211 08:12:14.578635 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:12:14 crc kubenswrapper[4860]: E1211 08:12:14.578742 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:12:14 crc kubenswrapper[4860]: I1211 08:12:14.578793 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:12:14 crc kubenswrapper[4860]: E1211 08:12:14.578866 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:12:14 crc kubenswrapper[4860]: I1211 08:12:14.660629 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:14 crc kubenswrapper[4860]: I1211 08:12:14.660694 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:14 crc kubenswrapper[4860]: I1211 08:12:14.660703 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:14 crc kubenswrapper[4860]: I1211 08:12:14.660721 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:14 crc kubenswrapper[4860]: I1211 08:12:14.660731 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:14Z","lastTransitionTime":"2025-12-11T08:12:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:14 crc kubenswrapper[4860]: I1211 08:12:14.763404 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:14 crc kubenswrapper[4860]: I1211 08:12:14.763437 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:14 crc kubenswrapper[4860]: I1211 08:12:14.763445 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:14 crc kubenswrapper[4860]: I1211 08:12:14.763458 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:14 crc kubenswrapper[4860]: I1211 08:12:14.763469 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:14Z","lastTransitionTime":"2025-12-11T08:12:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:14 crc kubenswrapper[4860]: I1211 08:12:14.865870 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:14 crc kubenswrapper[4860]: I1211 08:12:14.866267 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:14 crc kubenswrapper[4860]: I1211 08:12:14.866474 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:14 crc kubenswrapper[4860]: I1211 08:12:14.866636 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:14 crc kubenswrapper[4860]: I1211 08:12:14.866850 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:14Z","lastTransitionTime":"2025-12-11T08:12:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:14 crc kubenswrapper[4860]: I1211 08:12:14.969779 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:14 crc kubenswrapper[4860]: I1211 08:12:14.969851 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:14 crc kubenswrapper[4860]: I1211 08:12:14.969872 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:14 crc kubenswrapper[4860]: I1211 08:12:14.969897 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:14 crc kubenswrapper[4860]: I1211 08:12:14.969914 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:14Z","lastTransitionTime":"2025-12-11T08:12:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:15 crc kubenswrapper[4860]: I1211 08:12:15.072338 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:15 crc kubenswrapper[4860]: I1211 08:12:15.072401 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:15 crc kubenswrapper[4860]: I1211 08:12:15.072419 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:15 crc kubenswrapper[4860]: I1211 08:12:15.072443 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:15 crc kubenswrapper[4860]: I1211 08:12:15.072459 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:15Z","lastTransitionTime":"2025-12-11T08:12:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:15 crc kubenswrapper[4860]: I1211 08:12:15.174294 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:15 crc kubenswrapper[4860]: I1211 08:12:15.174330 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:15 crc kubenswrapper[4860]: I1211 08:12:15.174340 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:15 crc kubenswrapper[4860]: I1211 08:12:15.174354 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:15 crc kubenswrapper[4860]: I1211 08:12:15.174363 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:15Z","lastTransitionTime":"2025-12-11T08:12:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:15 crc kubenswrapper[4860]: I1211 08:12:15.276175 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:15 crc kubenswrapper[4860]: I1211 08:12:15.276455 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:15 crc kubenswrapper[4860]: I1211 08:12:15.276529 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:15 crc kubenswrapper[4860]: I1211 08:12:15.276603 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:15 crc kubenswrapper[4860]: I1211 08:12:15.276684 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:15Z","lastTransitionTime":"2025-12-11T08:12:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:15 crc kubenswrapper[4860]: I1211 08:12:15.378703 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:15 crc kubenswrapper[4860]: I1211 08:12:15.378762 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:15 crc kubenswrapper[4860]: I1211 08:12:15.378777 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:15 crc kubenswrapper[4860]: I1211 08:12:15.378793 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:15 crc kubenswrapper[4860]: I1211 08:12:15.378804 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:15Z","lastTransitionTime":"2025-12-11T08:12:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:15 crc kubenswrapper[4860]: I1211 08:12:15.482015 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:15 crc kubenswrapper[4860]: I1211 08:12:15.482065 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:15 crc kubenswrapper[4860]: I1211 08:12:15.482076 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:15 crc kubenswrapper[4860]: I1211 08:12:15.482098 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:15 crc kubenswrapper[4860]: I1211 08:12:15.482112 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:15Z","lastTransitionTime":"2025-12-11T08:12:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:15 crc kubenswrapper[4860]: I1211 08:12:15.584345 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:15 crc kubenswrapper[4860]: I1211 08:12:15.584664 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:15 crc kubenswrapper[4860]: I1211 08:12:15.584762 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:15 crc kubenswrapper[4860]: I1211 08:12:15.584896 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:15 crc kubenswrapper[4860]: I1211 08:12:15.585055 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:15Z","lastTransitionTime":"2025-12-11T08:12:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:15 crc kubenswrapper[4860]: I1211 08:12:15.687699 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:15 crc kubenswrapper[4860]: I1211 08:12:15.687753 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:15 crc kubenswrapper[4860]: I1211 08:12:15.687766 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:15 crc kubenswrapper[4860]: I1211 08:12:15.687785 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:15 crc kubenswrapper[4860]: I1211 08:12:15.687799 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:15Z","lastTransitionTime":"2025-12-11T08:12:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:15 crc kubenswrapper[4860]: I1211 08:12:15.790589 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:15 crc kubenswrapper[4860]: I1211 08:12:15.790633 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:15 crc kubenswrapper[4860]: I1211 08:12:15.790667 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:15 crc kubenswrapper[4860]: I1211 08:12:15.790686 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:15 crc kubenswrapper[4860]: I1211 08:12:15.790698 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:15Z","lastTransitionTime":"2025-12-11T08:12:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:15 crc kubenswrapper[4860]: I1211 08:12:15.893084 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:15 crc kubenswrapper[4860]: I1211 08:12:15.893117 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:15 crc kubenswrapper[4860]: I1211 08:12:15.893125 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:15 crc kubenswrapper[4860]: I1211 08:12:15.893138 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:15 crc kubenswrapper[4860]: I1211 08:12:15.893147 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:15Z","lastTransitionTime":"2025-12-11T08:12:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:15 crc kubenswrapper[4860]: I1211 08:12:15.995515 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:15 crc kubenswrapper[4860]: I1211 08:12:15.995557 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:15 crc kubenswrapper[4860]: I1211 08:12:15.995565 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:15 crc kubenswrapper[4860]: I1211 08:12:15.995577 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:15 crc kubenswrapper[4860]: I1211 08:12:15.995586 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:15Z","lastTransitionTime":"2025-12-11T08:12:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:16 crc kubenswrapper[4860]: I1211 08:12:16.097530 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:16 crc kubenswrapper[4860]: I1211 08:12:16.097580 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:16 crc kubenswrapper[4860]: I1211 08:12:16.097594 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:16 crc kubenswrapper[4860]: I1211 08:12:16.097610 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:16 crc kubenswrapper[4860]: I1211 08:12:16.097623 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:16Z","lastTransitionTime":"2025-12-11T08:12:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:16 crc kubenswrapper[4860]: I1211 08:12:16.200544 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:16 crc kubenswrapper[4860]: I1211 08:12:16.200591 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:16 crc kubenswrapper[4860]: I1211 08:12:16.200599 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:16 crc kubenswrapper[4860]: I1211 08:12:16.200612 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:16 crc kubenswrapper[4860]: I1211 08:12:16.200621 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:16Z","lastTransitionTime":"2025-12-11T08:12:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:16 crc kubenswrapper[4860]: I1211 08:12:16.303402 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:16 crc kubenswrapper[4860]: I1211 08:12:16.303463 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:16 crc kubenswrapper[4860]: I1211 08:12:16.303471 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:16 crc kubenswrapper[4860]: I1211 08:12:16.303487 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:16 crc kubenswrapper[4860]: I1211 08:12:16.303496 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:16Z","lastTransitionTime":"2025-12-11T08:12:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:16 crc kubenswrapper[4860]: I1211 08:12:16.406188 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:16 crc kubenswrapper[4860]: I1211 08:12:16.406276 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:16 crc kubenswrapper[4860]: I1211 08:12:16.406300 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:16 crc kubenswrapper[4860]: I1211 08:12:16.406332 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:16 crc kubenswrapper[4860]: I1211 08:12:16.406354 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:16Z","lastTransitionTime":"2025-12-11T08:12:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:16 crc kubenswrapper[4860]: I1211 08:12:16.508732 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:16 crc kubenswrapper[4860]: I1211 08:12:16.508794 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:16 crc kubenswrapper[4860]: I1211 08:12:16.508814 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:16 crc kubenswrapper[4860]: I1211 08:12:16.508839 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:16 crc kubenswrapper[4860]: I1211 08:12:16.508857 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:16Z","lastTransitionTime":"2025-12-11T08:12:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:16 crc kubenswrapper[4860]: I1211 08:12:16.578276 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gjhlj" Dec 11 08:12:16 crc kubenswrapper[4860]: I1211 08:12:16.578302 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:12:16 crc kubenswrapper[4860]: I1211 08:12:16.578339 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:12:16 crc kubenswrapper[4860]: I1211 08:12:16.578391 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:12:16 crc kubenswrapper[4860]: E1211 08:12:16.578447 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gjhlj" podUID="5cfd640a-f3d2-4b28-87e1-5f49f6341e21" Dec 11 08:12:16 crc kubenswrapper[4860]: E1211 08:12:16.578538 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:12:16 crc kubenswrapper[4860]: E1211 08:12:16.578655 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:12:16 crc kubenswrapper[4860]: E1211 08:12:16.578768 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:12:16 crc kubenswrapper[4860]: I1211 08:12:16.610944 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:16 crc kubenswrapper[4860]: I1211 08:12:16.610980 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:16 crc kubenswrapper[4860]: I1211 08:12:16.610991 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:16 crc kubenswrapper[4860]: I1211 08:12:16.611008 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:16 crc kubenswrapper[4860]: I1211 08:12:16.611020 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:16Z","lastTransitionTime":"2025-12-11T08:12:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:16 crc kubenswrapper[4860]: I1211 08:12:16.713142 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:16 crc kubenswrapper[4860]: I1211 08:12:16.713174 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:16 crc kubenswrapper[4860]: I1211 08:12:16.713184 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:16 crc kubenswrapper[4860]: I1211 08:12:16.713201 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:16 crc kubenswrapper[4860]: I1211 08:12:16.713211 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:16Z","lastTransitionTime":"2025-12-11T08:12:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:16 crc kubenswrapper[4860]: I1211 08:12:16.815982 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:16 crc kubenswrapper[4860]: I1211 08:12:16.816024 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:16 crc kubenswrapper[4860]: I1211 08:12:16.816040 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:16 crc kubenswrapper[4860]: I1211 08:12:16.816060 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:16 crc kubenswrapper[4860]: I1211 08:12:16.816075 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:16Z","lastTransitionTime":"2025-12-11T08:12:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:16 crc kubenswrapper[4860]: I1211 08:12:16.918800 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:16 crc kubenswrapper[4860]: I1211 08:12:16.918871 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:16 crc kubenswrapper[4860]: I1211 08:12:16.918887 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:16 crc kubenswrapper[4860]: I1211 08:12:16.918911 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:16 crc kubenswrapper[4860]: I1211 08:12:16.918930 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:16Z","lastTransitionTime":"2025-12-11T08:12:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:17 crc kubenswrapper[4860]: I1211 08:12:17.021561 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:17 crc kubenswrapper[4860]: I1211 08:12:17.021730 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:17 crc kubenswrapper[4860]: I1211 08:12:17.021755 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:17 crc kubenswrapper[4860]: I1211 08:12:17.021839 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:17 crc kubenswrapper[4860]: I1211 08:12:17.021906 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:17Z","lastTransitionTime":"2025-12-11T08:12:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:17 crc kubenswrapper[4860]: I1211 08:12:17.124155 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:17 crc kubenswrapper[4860]: I1211 08:12:17.124191 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:17 crc kubenswrapper[4860]: I1211 08:12:17.124201 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:17 crc kubenswrapper[4860]: I1211 08:12:17.124216 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:17 crc kubenswrapper[4860]: I1211 08:12:17.124228 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:17Z","lastTransitionTime":"2025-12-11T08:12:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:17 crc kubenswrapper[4860]: I1211 08:12:17.226869 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:17 crc kubenswrapper[4860]: I1211 08:12:17.226915 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:17 crc kubenswrapper[4860]: I1211 08:12:17.226927 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:17 crc kubenswrapper[4860]: I1211 08:12:17.226947 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:17 crc kubenswrapper[4860]: I1211 08:12:17.226958 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:17Z","lastTransitionTime":"2025-12-11T08:12:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:17 crc kubenswrapper[4860]: I1211 08:12:17.328850 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:17 crc kubenswrapper[4860]: I1211 08:12:17.328894 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:17 crc kubenswrapper[4860]: I1211 08:12:17.328905 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:17 crc kubenswrapper[4860]: I1211 08:12:17.328920 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:17 crc kubenswrapper[4860]: I1211 08:12:17.328931 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:17Z","lastTransitionTime":"2025-12-11T08:12:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:17 crc kubenswrapper[4860]: I1211 08:12:17.430719 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:17 crc kubenswrapper[4860]: I1211 08:12:17.430805 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:17 crc kubenswrapper[4860]: I1211 08:12:17.430818 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:17 crc kubenswrapper[4860]: I1211 08:12:17.430839 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:17 crc kubenswrapper[4860]: I1211 08:12:17.430850 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:17Z","lastTransitionTime":"2025-12-11T08:12:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:17 crc kubenswrapper[4860]: I1211 08:12:17.534388 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:17 crc kubenswrapper[4860]: I1211 08:12:17.534435 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:17 crc kubenswrapper[4860]: I1211 08:12:17.534449 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:17 crc kubenswrapper[4860]: I1211 08:12:17.534466 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:17 crc kubenswrapper[4860]: I1211 08:12:17.534478 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:17Z","lastTransitionTime":"2025-12-11T08:12:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:17 crc kubenswrapper[4860]: I1211 08:12:17.586700 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Dec 11 08:12:17 crc kubenswrapper[4860]: I1211 08:12:17.594543 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25a6a154da2b74952951ae94efc914ba078ea4bbe5893360b3c84d46be6bf973\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:17Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:17 crc kubenswrapper[4860]: I1211 08:12:17.605951 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6xq7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a75ce3c-5e1e-4b9f-bbde-7a3b0cce5eae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe8d4d675567b41728e476351811a93c47b19868ccbadab8c613a4eb8eca8562\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5c6nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6xq7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:17Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:17 crc kubenswrapper[4860]: I1211 08:12:17.620871 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"910ec852-e5d2-4370-806d-77f16a36e630\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://122eb7e14a042027de3aafaa8db89562d80565e7df20f302f872e48d412a26b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1348b83ed7b1e8c25f9dfb1b2d2d035b5adc96dc23854ee90236f06fb155f155\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://867cdac386b0576ee2958a520afd522a2558a8878ea6b5af977b136dc7bba4a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4602fc6635e6188bf4cb75c5c7cf1e71a64fc0c5420626bfafd87beae150722d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4602fc6635e6188bf4cb75c5c7cf1e71a64fc0c5420626bfafd87beae150722d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:17Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:17 crc kubenswrapper[4860]: I1211 08:12:17.632013 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:17Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:17 crc kubenswrapper[4860]: I1211 08:12:17.638053 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:17 crc kubenswrapper[4860]: I1211 08:12:17.638081 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:17 crc kubenswrapper[4860]: I1211 08:12:17.638094 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:17 crc kubenswrapper[4860]: I1211 08:12:17.638109 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:17 crc kubenswrapper[4860]: I1211 08:12:17.638118 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:17Z","lastTransitionTime":"2025-12-11T08:12:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:17 crc kubenswrapper[4860]: I1211 08:12:17.641341 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d8c2f28628dc2f1ffbdd98a55670e22cf6cdc8cf2c6442e9d1f8e5466b2f12e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://074ea63beac52bd196507ee7cfb1c3d7fc2d37d5ec6faa4d47dde62b922e1c43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:17Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:17 crc kubenswrapper[4860]: I1211 08:12:17.652202 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:17Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:17 crc kubenswrapper[4860]: I1211 08:12:17.663071 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ec3758-85f1-4a22-a4ec-c805472607a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e831db1a15976d3967d56479689572c26f2599a766baa0402897a3008428c56f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a498aa90f914e1f0370df01dc5ef174df113a2a16d86b2e3f24b19a1b99b4c0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8caebd2f4dfcbb89c597029a07c335171889761d2d200206987537fb0a228db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34e91f615b76445ff3241a304bb2964fde59de7e8a8dfff6841b2e628a31e86d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2090cab02ad74ddfd6bb57ff50eba35d18f56b0cde1c230fe011c01994b693bc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 08:11:21.171011 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 08:11:21.177674 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-644352688/tls.crt::/tmp/serving-cert-644352688/tls.key\\\\\\\"\\\\nI1211 08:11:26.508606 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 08:11:26.512741 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 08:11:26.512778 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 08:11:26.512839 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 08:11:26.512858 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 08:11:26.521684 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1211 08:11:26.521700 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1211 08:11:26.521722 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:11:26.521732 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:11:26.521738 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 08:11:26.521746 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 08:11:26.521749 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 08:11:26.521753 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1211 08:11:26.524999 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2274fb298ca7d2b37b060b4dbf9b166d39ba77a108b08edeeb4d67196e72cb3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf0aff8a6ab816b6617b1feeca9146d41bb5e8808bf1db96f8d493500c135d85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf0aff8a6ab816b6617b1feeca9146d41bb5e8808bf1db96f8d493500c135d85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:17Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:17 crc kubenswrapper[4860]: I1211 08:12:17.673607 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:17Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:17 crc kubenswrapper[4860]: I1211 08:12:17.684706 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a41f35db087418d1be16953f10377042abd3f4488e5ac1bc29ca5dcfc00a37e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:17Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:17 crc kubenswrapper[4860]: I1211 08:12:17.695305 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gjhlj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5cfd640a-f3d2-4b28-87e1-5f49f6341e21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knq7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knq7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:42Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gjhlj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:17Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:17 crc kubenswrapper[4860]: I1211 08:12:17.714802 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0589a204-c98c-417d-8256-bfe3bf747660\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3333b4346f59b04017e32d61f457ba7ff457166731128f339d7987ae6b56b06d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52bb8ade314afb9494f4e026d0af2aea7b1970d64dc66102f55e257ae4678154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dfcc4370edaae36a5a32b0a9c29290c50f8e4f427e4e147dc3c779beaa8b780\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d856ae95e9077524c518d82d10fb6b93b7fbcaaf2b720ffafe6f7fae58bd7caa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1426084ac580895c42354630910dd8c8e3a48ddd25f22abce1d28dc3de332a63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9984d5d242bb258405efe91fc9caf9e7ea592834445df2e78add9553a921c76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://769a74d7b125e7e981e5b0541324a777d021651182d900a57685c9db458395f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://769a74d7b125e7e981e5b0541324a777d021651182d900a57685c9db458395f1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-11T08:11:59Z\\\",\\\"message\\\":\\\"multus-pznwl\\\\nI1211 08:11:59.565441 6486 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1211 08:11:59.565459 6486 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-pznwl in node crc\\\\nI1211 08:11:59.565418 6486 ovn.go:134] Ensuring zone local for Pod openshift-kube-controller-manager/kube-controller-manager-crc in node crc\\\\nI1211 08:11:59.565467 6486 obj_retry.go:386] Retry successful for *v1.Pod openshift-multus/multus-pznwl after 0 failed attempt(s)\\\\nI1211 08:11:59.565474 6486 obj_retry.go:386] Retry successful for *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc after 0 failed attempt(s)\\\\nI1211 08:11:59.565476 6486 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-pznwl\\\\nI1211 08:11:59.565481 6486 default_network_controller.go:776] Recording success event on pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1211 08:11:59.565201 6486 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/network-metrics-daemon-gjhlj\\\\nI1211 08:11:59.565512 6486 ovn.go:134] Ensuring zone local for Pod openshift-multus/network-metrics-daemon-gjhlj in node crc\\\\nF1211 08:11:59.565540 6486 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:58Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-xvnqp_openshift-ovn-kubernetes(0589a204-c98c-417d-8256-bfe3bf747660)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb80d9928e8349fc5963227d135355ad7ba820d4fd89877af601d005ca03334b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17cbcb49c06f3a9059cd5a0f05eedb8c785da30a3e502e31bae2f57b3486cb14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17cbcb49c06f3a9059cd5a0f05eedb8c785da30a3e502e31bae2f57b3486cb14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xvnqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:17Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:17 crc kubenswrapper[4860]: I1211 08:12:17.722761 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gsx9l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aba5b6b9-f88d-4763-9562-6b4dec606df9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://735ef012c523ddc952d46a431f27b2582a2c301d9e9da0c697a20ff6e2fb5d86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6zw9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gsx9l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:17Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:17 crc kubenswrapper[4860]: I1211 08:12:17.738198 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9aa0aed-62b3-4ea2-82ae-9a653c1101a5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ad4ba1b769c6cb2b34073798f8e3c25eb0177e6d3059484bf365e6981cab710\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://077aa1dd2fe5189ff60a0f2365e47a6a545cca0de7aa6379370908db4ea5838a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d7c919899b74a2b0fcd9221f3f769a9d013267354b200a282669d342d6413d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5483384f1d2ebc1fefa9706c9e67a3dc5a00ff643f803e3f2707c252e27ad94\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:17Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:17 crc kubenswrapper[4860]: I1211 08:12:17.739602 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:17 crc kubenswrapper[4860]: I1211 08:12:17.739708 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:17 crc kubenswrapper[4860]: I1211 08:12:17.739734 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:17 crc kubenswrapper[4860]: I1211 08:12:17.739767 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:17 crc kubenswrapper[4860]: I1211 08:12:17.739799 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:17Z","lastTransitionTime":"2025-12-11T08:12:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:17 crc kubenswrapper[4860]: I1211 08:12:17.749769 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31c30642-6e60-41b4-a477-0d802424e0aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79969402f5c33852a26fc839f37a45febf4f2377d77823284fbb01769e5585fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2d4d5de7106175087d91acb1f0c7552014f0c08bfcd65cbbf184d3f8d7826ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-99qgp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:17Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:17 crc kubenswrapper[4860]: I1211 08:12:17.769708 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-26nc6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1faa733-f408-41fd-b7aa-0cd86283976a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fce5b642b843ed30823c94e0b0c3d8bb1e5d2d93d691f7b8ef901ece03b0ac6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ac45db4f858a1c18a6b87472fc1cea2300f5845f1ad62c22260374995548a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ac45db4f858a1c18a6b87472fc1cea2300f5845f1ad62c22260374995548a0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b35296a5eb17dd2bb5e72c86c1cafdb30b7712cb330264f935d3b4fff05cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c6b35296a5eb17dd2bb5e72c86c1cafdb30b7712cb330264f935d3b4fff05cd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b9e82d6d3a247309ef5bdbf5b3a86ec083add2a138b0b237801e77b09ee96a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b9e82d6d3a247309ef5bdbf5b3a86ec083add2a138b0b237801e77b09ee96a1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://908ee695e78145ef4375c4097d9f1e772502922d02a2fcb4566172a6ad9caaf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://908ee695e78145ef4375c4097d9f1e772502922d02a2fcb4566172a6ad9caaf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afb0a097f3556a8f9ed7cf8f762cb99c082fe0a36b8d8d3038bf188b08fe9a84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afb0a097f3556a8f9ed7cf8f762cb99c082fe0a36b8d8d3038bf188b08fe9a84\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce18f4f06db091e658684fa0bbc858e19b9f03010acf13b3ffccb1a6e3f88021\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce18f4f06db091e658684fa0bbc858e19b9f03010acf13b3ffccb1a6e3f88021\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-26nc6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:17Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:17 crc kubenswrapper[4860]: I1211 08:12:17.792575 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9872be1e-b995-4a00-904b-875dff96ad56\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f74a32a8fefea704cc73ffd1114083dab68aa44dc33f8d9d432d2d24216546c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbe00b46063d3968d59c393b87178e135f982abb7d0f32fa798b212d4b83df6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eedc9988e1a4a3c8cae730dc4a352103a6d665535b825e157d0f87c331b137c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2137fee4b7abc9765f9517d7b3b6ea796601a31d03151bf167d17136ced8b25a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f683e48f3577a68862eb6302383538376e41f49b20b1e0ea3f1b3b143f24872a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5de794187456b3576b969069a3b839d75c7918418a97fa35dff8bb40bb76e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5de794187456b3576b969069a3b839d75c7918418a97fa35dff8bb40bb76e39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08ce42b974fba00c409c0af971f0de351e50f18b2f06f8e722150da3239bb908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08ce42b974fba00c409c0af971f0de351e50f18b2f06f8e722150da3239bb908\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cfac6be03e3cfa0b0f3450b48b67957f2d5172b9c65e27f14c2a260bda5d423f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cfac6be03e3cfa0b0f3450b48b67957f2d5172b9c65e27f14c2a260bda5d423f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:17Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:17 crc kubenswrapper[4860]: I1211 08:12:17.803973 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-pznwl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5310c1fc-66c6-40aa-b1bf-5a59a2410f9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2d6ff00ce322a28607bc3f6db05c5e4fb14acb6777ea315786ef53963ca504a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hkt48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-pznwl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:17Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:17 crc kubenswrapper[4860]: I1211 08:12:17.816058 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-854v6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e6529ba7-2341-43ac-a187-95ab530fe775\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8986538a421f0184f52f288b0e171da695591caf2b10c67ad2dce1073a6e6016\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kh22b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0d4e3f761ad15ae1424204e40b9d3f60188ab4a9047bd4b5ac6a8a8ca1fb385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kh22b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-854v6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:17Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:17 crc kubenswrapper[4860]: I1211 08:12:17.841708 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:17 crc kubenswrapper[4860]: I1211 08:12:17.841749 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:17 crc kubenswrapper[4860]: I1211 08:12:17.841758 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:17 crc kubenswrapper[4860]: I1211 08:12:17.841776 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:17 crc kubenswrapper[4860]: I1211 08:12:17.841786 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:17Z","lastTransitionTime":"2025-12-11T08:12:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:17 crc kubenswrapper[4860]: I1211 08:12:17.944498 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:17 crc kubenswrapper[4860]: I1211 08:12:17.944538 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:17 crc kubenswrapper[4860]: I1211 08:12:17.944549 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:17 crc kubenswrapper[4860]: I1211 08:12:17.944565 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:17 crc kubenswrapper[4860]: I1211 08:12:17.944575 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:17Z","lastTransitionTime":"2025-12-11T08:12:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:18 crc kubenswrapper[4860]: I1211 08:12:18.046442 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:18 crc kubenswrapper[4860]: I1211 08:12:18.046474 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:18 crc kubenswrapper[4860]: I1211 08:12:18.046483 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:18 crc kubenswrapper[4860]: I1211 08:12:18.046496 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:18 crc kubenswrapper[4860]: I1211 08:12:18.046505 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:18Z","lastTransitionTime":"2025-12-11T08:12:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:18 crc kubenswrapper[4860]: I1211 08:12:18.064594 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-pznwl_5310c1fc-66c6-40aa-b1bf-5a59a2410f9c/kube-multus/0.log" Dec 11 08:12:18 crc kubenswrapper[4860]: I1211 08:12:18.064664 4860 generic.go:334] "Generic (PLEG): container finished" podID="5310c1fc-66c6-40aa-b1bf-5a59a2410f9c" containerID="d2d6ff00ce322a28607bc3f6db05c5e4fb14acb6777ea315786ef53963ca504a" exitCode=1 Dec 11 08:12:18 crc kubenswrapper[4860]: I1211 08:12:18.064810 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-pznwl" event={"ID":"5310c1fc-66c6-40aa-b1bf-5a59a2410f9c","Type":"ContainerDied","Data":"d2d6ff00ce322a28607bc3f6db05c5e4fb14acb6777ea315786ef53963ca504a"} Dec 11 08:12:18 crc kubenswrapper[4860]: I1211 08:12:18.067054 4860 scope.go:117] "RemoveContainer" containerID="d2d6ff00ce322a28607bc3f6db05c5e4fb14acb6777ea315786ef53963ca504a" Dec 11 08:12:18 crc kubenswrapper[4860]: I1211 08:12:18.079261 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gjhlj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5cfd640a-f3d2-4b28-87e1-5f49f6341e21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knq7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knq7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:42Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gjhlj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:18Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:18 crc kubenswrapper[4860]: I1211 08:12:18.100312 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ec3758-85f1-4a22-a4ec-c805472607a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e831db1a15976d3967d56479689572c26f2599a766baa0402897a3008428c56f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a498aa90f914e1f0370df01dc5ef174df113a2a16d86b2e3f24b19a1b99b4c0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8caebd2f4dfcbb89c597029a07c335171889761d2d200206987537fb0a228db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34e91f615b76445ff3241a304bb2964fde59de7e8a8dfff6841b2e628a31e86d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2090cab02ad74ddfd6bb57ff50eba35d18f56b0cde1c230fe011c01994b693bc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 08:11:21.171011 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 08:11:21.177674 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-644352688/tls.crt::/tmp/serving-cert-644352688/tls.key\\\\\\\"\\\\nI1211 08:11:26.508606 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 08:11:26.512741 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 08:11:26.512778 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 08:11:26.512839 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 08:11:26.512858 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 08:11:26.521684 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1211 08:11:26.521700 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1211 08:11:26.521722 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:11:26.521732 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:11:26.521738 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 08:11:26.521746 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 08:11:26.521749 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 08:11:26.521753 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1211 08:11:26.524999 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2274fb298ca7d2b37b060b4dbf9b166d39ba77a108b08edeeb4d67196e72cb3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf0aff8a6ab816b6617b1feeca9146d41bb5e8808bf1db96f8d493500c135d85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf0aff8a6ab816b6617b1feeca9146d41bb5e8808bf1db96f8d493500c135d85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:18Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:18 crc kubenswrapper[4860]: I1211 08:12:18.115574 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:18Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:18 crc kubenswrapper[4860]: I1211 08:12:18.130059 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a41f35db087418d1be16953f10377042abd3f4488e5ac1bc29ca5dcfc00a37e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:18Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:18 crc kubenswrapper[4860]: I1211 08:12:18.143108 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-26nc6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1faa733-f408-41fd-b7aa-0cd86283976a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fce5b642b843ed30823c94e0b0c3d8bb1e5d2d93d691f7b8ef901ece03b0ac6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ac45db4f858a1c18a6b87472fc1cea2300f5845f1ad62c22260374995548a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ac45db4f858a1c18a6b87472fc1cea2300f5845f1ad62c22260374995548a0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b35296a5eb17dd2bb5e72c86c1cafdb30b7712cb330264f935d3b4fff05cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c6b35296a5eb17dd2bb5e72c86c1cafdb30b7712cb330264f935d3b4fff05cd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b9e82d6d3a247309ef5bdbf5b3a86ec083add2a138b0b237801e77b09ee96a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b9e82d6d3a247309ef5bdbf5b3a86ec083add2a138b0b237801e77b09ee96a1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://908ee695e78145ef4375c4097d9f1e772502922d02a2fcb4566172a6ad9caaf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://908ee695e78145ef4375c4097d9f1e772502922d02a2fcb4566172a6ad9caaf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afb0a097f3556a8f9ed7cf8f762cb99c082fe0a36b8d8d3038bf188b08fe9a84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afb0a097f3556a8f9ed7cf8f762cb99c082fe0a36b8d8d3038bf188b08fe9a84\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce18f4f06db091e658684fa0bbc858e19b9f03010acf13b3ffccb1a6e3f88021\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce18f4f06db091e658684fa0bbc858e19b9f03010acf13b3ffccb1a6e3f88021\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-26nc6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:18Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:18 crc kubenswrapper[4860]: I1211 08:12:18.149456 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:18 crc kubenswrapper[4860]: I1211 08:12:18.149487 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:18 crc kubenswrapper[4860]: I1211 08:12:18.149495 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:18 crc kubenswrapper[4860]: I1211 08:12:18.149527 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:18 crc kubenswrapper[4860]: I1211 08:12:18.149538 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:18Z","lastTransitionTime":"2025-12-11T08:12:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:18 crc kubenswrapper[4860]: I1211 08:12:18.160351 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0589a204-c98c-417d-8256-bfe3bf747660\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3333b4346f59b04017e32d61f457ba7ff457166731128f339d7987ae6b56b06d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52bb8ade314afb9494f4e026d0af2aea7b1970d64dc66102f55e257ae4678154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dfcc4370edaae36a5a32b0a9c29290c50f8e4f427e4e147dc3c779beaa8b780\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d856ae95e9077524c518d82d10fb6b93b7fbcaaf2b720ffafe6f7fae58bd7caa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1426084ac580895c42354630910dd8c8e3a48ddd25f22abce1d28dc3de332a63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9984d5d242bb258405efe91fc9caf9e7ea592834445df2e78add9553a921c76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://769a74d7b125e7e981e5b0541324a777d021651182d900a57685c9db458395f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://769a74d7b125e7e981e5b0541324a777d021651182d900a57685c9db458395f1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-11T08:11:59Z\\\",\\\"message\\\":\\\"multus-pznwl\\\\nI1211 08:11:59.565441 6486 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1211 08:11:59.565459 6486 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-pznwl in node crc\\\\nI1211 08:11:59.565418 6486 ovn.go:134] Ensuring zone local for Pod openshift-kube-controller-manager/kube-controller-manager-crc in node crc\\\\nI1211 08:11:59.565467 6486 obj_retry.go:386] Retry successful for *v1.Pod openshift-multus/multus-pznwl after 0 failed attempt(s)\\\\nI1211 08:11:59.565474 6486 obj_retry.go:386] Retry successful for *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc after 0 failed attempt(s)\\\\nI1211 08:11:59.565476 6486 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-pznwl\\\\nI1211 08:11:59.565481 6486 default_network_controller.go:776] Recording success event on pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1211 08:11:59.565201 6486 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/network-metrics-daemon-gjhlj\\\\nI1211 08:11:59.565512 6486 ovn.go:134] Ensuring zone local for Pod openshift-multus/network-metrics-daemon-gjhlj in node crc\\\\nF1211 08:11:59.565540 6486 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:58Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-xvnqp_openshift-ovn-kubernetes(0589a204-c98c-417d-8256-bfe3bf747660)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb80d9928e8349fc5963227d135355ad7ba820d4fd89877af601d005ca03334b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17cbcb49c06f3a9059cd5a0f05eedb8c785da30a3e502e31bae2f57b3486cb14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17cbcb49c06f3a9059cd5a0f05eedb8c785da30a3e502e31bae2f57b3486cb14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xvnqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:18Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:18 crc kubenswrapper[4860]: I1211 08:12:18.171797 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gsx9l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aba5b6b9-f88d-4763-9562-6b4dec606df9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://735ef012c523ddc952d46a431f27b2582a2c301d9e9da0c697a20ff6e2fb5d86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6zw9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gsx9l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:18Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:18 crc kubenswrapper[4860]: I1211 08:12:18.183415 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8bf0e9df-4e9d-4f45-a5b9-0020fdab2ea6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77c0c831f046149002cd12564aa9b1544a813da1fcd81e61451eeb504f0450b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3498078e25d1156a89ddc17f4823182b1aeb8b283f8b46bf66df67d67c5c5e44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3498078e25d1156a89ddc17f4823182b1aeb8b283f8b46bf66df67d67c5c5e44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:18Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:18 crc kubenswrapper[4860]: I1211 08:12:18.198055 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9aa0aed-62b3-4ea2-82ae-9a653c1101a5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ad4ba1b769c6cb2b34073798f8e3c25eb0177e6d3059484bf365e6981cab710\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://077aa1dd2fe5189ff60a0f2365e47a6a545cca0de7aa6379370908db4ea5838a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d7c919899b74a2b0fcd9221f3f769a9d013267354b200a282669d342d6413d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5483384f1d2ebc1fefa9706c9e67a3dc5a00ff643f803e3f2707c252e27ad94\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:18Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:18 crc kubenswrapper[4860]: I1211 08:12:18.210950 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31c30642-6e60-41b4-a477-0d802424e0aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79969402f5c33852a26fc839f37a45febf4f2377d77823284fbb01769e5585fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2d4d5de7106175087d91acb1f0c7552014f0c08bfcd65cbbf184d3f8d7826ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-99qgp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:18Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:18 crc kubenswrapper[4860]: I1211 08:12:18.231554 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9872be1e-b995-4a00-904b-875dff96ad56\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f74a32a8fefea704cc73ffd1114083dab68aa44dc33f8d9d432d2d24216546c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbe00b46063d3968d59c393b87178e135f982abb7d0f32fa798b212d4b83df6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eedc9988e1a4a3c8cae730dc4a352103a6d665535b825e157d0f87c331b137c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2137fee4b7abc9765f9517d7b3b6ea796601a31d03151bf167d17136ced8b25a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f683e48f3577a68862eb6302383538376e41f49b20b1e0ea3f1b3b143f24872a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5de794187456b3576b969069a3b839d75c7918418a97fa35dff8bb40bb76e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5de794187456b3576b969069a3b839d75c7918418a97fa35dff8bb40bb76e39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08ce42b974fba00c409c0af971f0de351e50f18b2f06f8e722150da3239bb908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08ce42b974fba00c409c0af971f0de351e50f18b2f06f8e722150da3239bb908\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cfac6be03e3cfa0b0f3450b48b67957f2d5172b9c65e27f14c2a260bda5d423f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cfac6be03e3cfa0b0f3450b48b67957f2d5172b9c65e27f14c2a260bda5d423f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:18Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:18 crc kubenswrapper[4860]: I1211 08:12:18.244201 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-pznwl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5310c1fc-66c6-40aa-b1bf-5a59a2410f9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2d6ff00ce322a28607bc3f6db05c5e4fb14acb6777ea315786ef53963ca504a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2d6ff00ce322a28607bc3f6db05c5e4fb14acb6777ea315786ef53963ca504a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-11T08:12:17Z\\\",\\\"message\\\":\\\"2025-12-11T08:11:31+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_f37a703f-1609-4581-ab99-50337287ff33\\\\n2025-12-11T08:11:31+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_f37a703f-1609-4581-ab99-50337287ff33 to /host/opt/cni/bin/\\\\n2025-12-11T08:11:32Z [verbose] multus-daemon started\\\\n2025-12-11T08:11:32Z [verbose] Readiness Indicator file check\\\\n2025-12-11T08:12:17Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hkt48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-pznwl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:18Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:18 crc kubenswrapper[4860]: I1211 08:12:18.252022 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:18 crc kubenswrapper[4860]: I1211 08:12:18.252054 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:18 crc kubenswrapper[4860]: I1211 08:12:18.252063 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:18 crc kubenswrapper[4860]: I1211 08:12:18.252079 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:18 crc kubenswrapper[4860]: I1211 08:12:18.252091 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:18Z","lastTransitionTime":"2025-12-11T08:12:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:18 crc kubenswrapper[4860]: I1211 08:12:18.257001 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-854v6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e6529ba7-2341-43ac-a187-95ab530fe775\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8986538a421f0184f52f288b0e171da695591caf2b10c67ad2dce1073a6e6016\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kh22b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0d4e3f761ad15ae1424204e40b9d3f60188ab4a9047bd4b5ac6a8a8ca1fb385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kh22b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-854v6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:18Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:18 crc kubenswrapper[4860]: I1211 08:12:18.269593 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:18Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:18 crc kubenswrapper[4860]: I1211 08:12:18.281673 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25a6a154da2b74952951ae94efc914ba078ea4bbe5893360b3c84d46be6bf973\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:18Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:18 crc kubenswrapper[4860]: I1211 08:12:18.292343 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6xq7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a75ce3c-5e1e-4b9f-bbde-7a3b0cce5eae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe8d4d675567b41728e476351811a93c47b19868ccbadab8c613a4eb8eca8562\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5c6nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6xq7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:18Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:18 crc kubenswrapper[4860]: I1211 08:12:18.304475 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"910ec852-e5d2-4370-806d-77f16a36e630\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://122eb7e14a042027de3aafaa8db89562d80565e7df20f302f872e48d412a26b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1348b83ed7b1e8c25f9dfb1b2d2d035b5adc96dc23854ee90236f06fb155f155\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://867cdac386b0576ee2958a520afd522a2558a8878ea6b5af977b136dc7bba4a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4602fc6635e6188bf4cb75c5c7cf1e71a64fc0c5420626bfafd87beae150722d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4602fc6635e6188bf4cb75c5c7cf1e71a64fc0c5420626bfafd87beae150722d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:18Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:18 crc kubenswrapper[4860]: I1211 08:12:18.316263 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:18Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:18 crc kubenswrapper[4860]: I1211 08:12:18.330253 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d8c2f28628dc2f1ffbdd98a55670e22cf6cdc8cf2c6442e9d1f8e5466b2f12e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://074ea63beac52bd196507ee7cfb1c3d7fc2d37d5ec6faa4d47dde62b922e1c43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:18Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:18 crc kubenswrapper[4860]: I1211 08:12:18.354900 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:18 crc kubenswrapper[4860]: I1211 08:12:18.354960 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:18 crc kubenswrapper[4860]: I1211 08:12:18.354970 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:18 crc kubenswrapper[4860]: I1211 08:12:18.354987 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:18 crc kubenswrapper[4860]: I1211 08:12:18.354998 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:18Z","lastTransitionTime":"2025-12-11T08:12:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:18 crc kubenswrapper[4860]: I1211 08:12:18.458713 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:18 crc kubenswrapper[4860]: I1211 08:12:18.458758 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:18 crc kubenswrapper[4860]: I1211 08:12:18.458767 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:18 crc kubenswrapper[4860]: I1211 08:12:18.458783 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:18 crc kubenswrapper[4860]: I1211 08:12:18.458795 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:18Z","lastTransitionTime":"2025-12-11T08:12:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:18 crc kubenswrapper[4860]: I1211 08:12:18.561400 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:18 crc kubenswrapper[4860]: I1211 08:12:18.561866 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:18 crc kubenswrapper[4860]: I1211 08:12:18.562039 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:18 crc kubenswrapper[4860]: I1211 08:12:18.562200 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:18 crc kubenswrapper[4860]: I1211 08:12:18.562367 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:18Z","lastTransitionTime":"2025-12-11T08:12:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:18 crc kubenswrapper[4860]: I1211 08:12:18.577849 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:12:18 crc kubenswrapper[4860]: I1211 08:12:18.577924 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gjhlj" Dec 11 08:12:18 crc kubenswrapper[4860]: E1211 08:12:18.577973 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:12:18 crc kubenswrapper[4860]: E1211 08:12:18.578064 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gjhlj" podUID="5cfd640a-f3d2-4b28-87e1-5f49f6341e21" Dec 11 08:12:18 crc kubenswrapper[4860]: I1211 08:12:18.578146 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:12:18 crc kubenswrapper[4860]: E1211 08:12:18.578204 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:12:18 crc kubenswrapper[4860]: I1211 08:12:18.578621 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:12:18 crc kubenswrapper[4860]: E1211 08:12:18.579006 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:12:18 crc kubenswrapper[4860]: I1211 08:12:18.665169 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:18 crc kubenswrapper[4860]: I1211 08:12:18.665248 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:18 crc kubenswrapper[4860]: I1211 08:12:18.665273 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:18 crc kubenswrapper[4860]: I1211 08:12:18.665299 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:18 crc kubenswrapper[4860]: I1211 08:12:18.665315 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:18Z","lastTransitionTime":"2025-12-11T08:12:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:18 crc kubenswrapper[4860]: I1211 08:12:18.768283 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:18 crc kubenswrapper[4860]: I1211 08:12:18.768338 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:18 crc kubenswrapper[4860]: I1211 08:12:18.768355 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:18 crc kubenswrapper[4860]: I1211 08:12:18.768378 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:18 crc kubenswrapper[4860]: I1211 08:12:18.768397 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:18Z","lastTransitionTime":"2025-12-11T08:12:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:18 crc kubenswrapper[4860]: I1211 08:12:18.871394 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:18 crc kubenswrapper[4860]: I1211 08:12:18.871459 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:18 crc kubenswrapper[4860]: I1211 08:12:18.871482 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:18 crc kubenswrapper[4860]: I1211 08:12:18.871510 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:18 crc kubenswrapper[4860]: I1211 08:12:18.871532 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:18Z","lastTransitionTime":"2025-12-11T08:12:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:18 crc kubenswrapper[4860]: I1211 08:12:18.974018 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:18 crc kubenswrapper[4860]: I1211 08:12:18.974049 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:18 crc kubenswrapper[4860]: I1211 08:12:18.974058 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:18 crc kubenswrapper[4860]: I1211 08:12:18.974070 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:18 crc kubenswrapper[4860]: I1211 08:12:18.974081 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:18Z","lastTransitionTime":"2025-12-11T08:12:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:19 crc kubenswrapper[4860]: I1211 08:12:19.068779 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-pznwl_5310c1fc-66c6-40aa-b1bf-5a59a2410f9c/kube-multus/0.log" Dec 11 08:12:19 crc kubenswrapper[4860]: I1211 08:12:19.068832 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-pznwl" event={"ID":"5310c1fc-66c6-40aa-b1bf-5a59a2410f9c","Type":"ContainerStarted","Data":"34119a6367261d7e12174f914ee8cfaa55703a55210f83c58b0e4694c54a5b03"} Dec 11 08:12:19 crc kubenswrapper[4860]: I1211 08:12:19.075978 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:19 crc kubenswrapper[4860]: I1211 08:12:19.076017 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:19 crc kubenswrapper[4860]: I1211 08:12:19.076030 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:19 crc kubenswrapper[4860]: I1211 08:12:19.076045 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:19 crc kubenswrapper[4860]: I1211 08:12:19.076057 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:19Z","lastTransitionTime":"2025-12-11T08:12:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:19 crc kubenswrapper[4860]: I1211 08:12:19.079981 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"910ec852-e5d2-4370-806d-77f16a36e630\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://122eb7e14a042027de3aafaa8db89562d80565e7df20f302f872e48d412a26b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1348b83ed7b1e8c25f9dfb1b2d2d035b5adc96dc23854ee90236f06fb155f155\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://867cdac386b0576ee2958a520afd522a2558a8878ea6b5af977b136dc7bba4a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4602fc6635e6188bf4cb75c5c7cf1e71a64fc0c5420626bfafd87beae150722d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4602fc6635e6188bf4cb75c5c7cf1e71a64fc0c5420626bfafd87beae150722d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:19Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:19 crc kubenswrapper[4860]: I1211 08:12:19.090854 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:19Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:19 crc kubenswrapper[4860]: I1211 08:12:19.102701 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d8c2f28628dc2f1ffbdd98a55670e22cf6cdc8cf2c6442e9d1f8e5466b2f12e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://074ea63beac52bd196507ee7cfb1c3d7fc2d37d5ec6faa4d47dde62b922e1c43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:19Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:19 crc kubenswrapper[4860]: I1211 08:12:19.112690 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:19Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:19 crc kubenswrapper[4860]: I1211 08:12:19.124199 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25a6a154da2b74952951ae94efc914ba078ea4bbe5893360b3c84d46be6bf973\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:19Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:19 crc kubenswrapper[4860]: I1211 08:12:19.134879 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6xq7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a75ce3c-5e1e-4b9f-bbde-7a3b0cce5eae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe8d4d675567b41728e476351811a93c47b19868ccbadab8c613a4eb8eca8562\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5c6nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6xq7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:19Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:19 crc kubenswrapper[4860]: I1211 08:12:19.146484 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ec3758-85f1-4a22-a4ec-c805472607a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e831db1a15976d3967d56479689572c26f2599a766baa0402897a3008428c56f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a498aa90f914e1f0370df01dc5ef174df113a2a16d86b2e3f24b19a1b99b4c0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8caebd2f4dfcbb89c597029a07c335171889761d2d200206987537fb0a228db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34e91f615b76445ff3241a304bb2964fde59de7e8a8dfff6841b2e628a31e86d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2090cab02ad74ddfd6bb57ff50eba35d18f56b0cde1c230fe011c01994b693bc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 08:11:21.171011 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 08:11:21.177674 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-644352688/tls.crt::/tmp/serving-cert-644352688/tls.key\\\\\\\"\\\\nI1211 08:11:26.508606 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 08:11:26.512741 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 08:11:26.512778 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 08:11:26.512839 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 08:11:26.512858 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 08:11:26.521684 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1211 08:11:26.521700 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1211 08:11:26.521722 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:11:26.521732 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:11:26.521738 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 08:11:26.521746 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 08:11:26.521749 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 08:11:26.521753 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1211 08:11:26.524999 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2274fb298ca7d2b37b060b4dbf9b166d39ba77a108b08edeeb4d67196e72cb3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf0aff8a6ab816b6617b1feeca9146d41bb5e8808bf1db96f8d493500c135d85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf0aff8a6ab816b6617b1feeca9146d41bb5e8808bf1db96f8d493500c135d85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:19Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:19 crc kubenswrapper[4860]: I1211 08:12:19.155994 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:19Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:19 crc kubenswrapper[4860]: I1211 08:12:19.167198 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a41f35db087418d1be16953f10377042abd3f4488e5ac1bc29ca5dcfc00a37e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:19Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:19 crc kubenswrapper[4860]: I1211 08:12:19.176241 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gjhlj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5cfd640a-f3d2-4b28-87e1-5f49f6341e21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knq7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knq7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:42Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gjhlj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:19Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:19 crc kubenswrapper[4860]: I1211 08:12:19.178034 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:19 crc kubenswrapper[4860]: I1211 08:12:19.178062 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:19 crc kubenswrapper[4860]: I1211 08:12:19.178072 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:19 crc kubenswrapper[4860]: I1211 08:12:19.178086 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:19 crc kubenswrapper[4860]: I1211 08:12:19.178095 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:19Z","lastTransitionTime":"2025-12-11T08:12:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:19 crc kubenswrapper[4860]: I1211 08:12:19.183864 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8bf0e9df-4e9d-4f45-a5b9-0020fdab2ea6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77c0c831f046149002cd12564aa9b1544a813da1fcd81e61451eeb504f0450b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3498078e25d1156a89ddc17f4823182b1aeb8b283f8b46bf66df67d67c5c5e44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3498078e25d1156a89ddc17f4823182b1aeb8b283f8b46bf66df67d67c5c5e44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:19Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:19 crc kubenswrapper[4860]: I1211 08:12:19.195342 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9aa0aed-62b3-4ea2-82ae-9a653c1101a5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ad4ba1b769c6cb2b34073798f8e3c25eb0177e6d3059484bf365e6981cab710\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://077aa1dd2fe5189ff60a0f2365e47a6a545cca0de7aa6379370908db4ea5838a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d7c919899b74a2b0fcd9221f3f769a9d013267354b200a282669d342d6413d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5483384f1d2ebc1fefa9706c9e67a3dc5a00ff643f803e3f2707c252e27ad94\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:19Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:19 crc kubenswrapper[4860]: I1211 08:12:19.204261 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31c30642-6e60-41b4-a477-0d802424e0aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79969402f5c33852a26fc839f37a45febf4f2377d77823284fbb01769e5585fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2d4d5de7106175087d91acb1f0c7552014f0c08bfcd65cbbf184d3f8d7826ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-99qgp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:19Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:19 crc kubenswrapper[4860]: I1211 08:12:19.217339 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-26nc6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1faa733-f408-41fd-b7aa-0cd86283976a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fce5b642b843ed30823c94e0b0c3d8bb1e5d2d93d691f7b8ef901ece03b0ac6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ac45db4f858a1c18a6b87472fc1cea2300f5845f1ad62c22260374995548a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ac45db4f858a1c18a6b87472fc1cea2300f5845f1ad62c22260374995548a0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b35296a5eb17dd2bb5e72c86c1cafdb30b7712cb330264f935d3b4fff05cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c6b35296a5eb17dd2bb5e72c86c1cafdb30b7712cb330264f935d3b4fff05cd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b9e82d6d3a247309ef5bdbf5b3a86ec083add2a138b0b237801e77b09ee96a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b9e82d6d3a247309ef5bdbf5b3a86ec083add2a138b0b237801e77b09ee96a1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://908ee695e78145ef4375c4097d9f1e772502922d02a2fcb4566172a6ad9caaf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://908ee695e78145ef4375c4097d9f1e772502922d02a2fcb4566172a6ad9caaf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afb0a097f3556a8f9ed7cf8f762cb99c082fe0a36b8d8d3038bf188b08fe9a84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afb0a097f3556a8f9ed7cf8f762cb99c082fe0a36b8d8d3038bf188b08fe9a84\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce18f4f06db091e658684fa0bbc858e19b9f03010acf13b3ffccb1a6e3f88021\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce18f4f06db091e658684fa0bbc858e19b9f03010acf13b3ffccb1a6e3f88021\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-26nc6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:19Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:19 crc kubenswrapper[4860]: I1211 08:12:19.235932 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0589a204-c98c-417d-8256-bfe3bf747660\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3333b4346f59b04017e32d61f457ba7ff457166731128f339d7987ae6b56b06d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52bb8ade314afb9494f4e026d0af2aea7b1970d64dc66102f55e257ae4678154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dfcc4370edaae36a5a32b0a9c29290c50f8e4f427e4e147dc3c779beaa8b780\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d856ae95e9077524c518d82d10fb6b93b7fbcaaf2b720ffafe6f7fae58bd7caa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1426084ac580895c42354630910dd8c8e3a48ddd25f22abce1d28dc3de332a63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9984d5d242bb258405efe91fc9caf9e7ea592834445df2e78add9553a921c76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://769a74d7b125e7e981e5b0541324a777d021651182d900a57685c9db458395f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://769a74d7b125e7e981e5b0541324a777d021651182d900a57685c9db458395f1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-11T08:11:59Z\\\",\\\"message\\\":\\\"multus-pznwl\\\\nI1211 08:11:59.565441 6486 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1211 08:11:59.565459 6486 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-pznwl in node crc\\\\nI1211 08:11:59.565418 6486 ovn.go:134] Ensuring zone local for Pod openshift-kube-controller-manager/kube-controller-manager-crc in node crc\\\\nI1211 08:11:59.565467 6486 obj_retry.go:386] Retry successful for *v1.Pod openshift-multus/multus-pznwl after 0 failed attempt(s)\\\\nI1211 08:11:59.565474 6486 obj_retry.go:386] Retry successful for *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc after 0 failed attempt(s)\\\\nI1211 08:11:59.565476 6486 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-pznwl\\\\nI1211 08:11:59.565481 6486 default_network_controller.go:776] Recording success event on pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1211 08:11:59.565201 6486 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/network-metrics-daemon-gjhlj\\\\nI1211 08:11:59.565512 6486 ovn.go:134] Ensuring zone local for Pod openshift-multus/network-metrics-daemon-gjhlj in node crc\\\\nF1211 08:11:59.565540 6486 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:58Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-xvnqp_openshift-ovn-kubernetes(0589a204-c98c-417d-8256-bfe3bf747660)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb80d9928e8349fc5963227d135355ad7ba820d4fd89877af601d005ca03334b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17cbcb49c06f3a9059cd5a0f05eedb8c785da30a3e502e31bae2f57b3486cb14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17cbcb49c06f3a9059cd5a0f05eedb8c785da30a3e502e31bae2f57b3486cb14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xvnqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:19Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:19 crc kubenswrapper[4860]: I1211 08:12:19.244196 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gsx9l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aba5b6b9-f88d-4763-9562-6b4dec606df9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://735ef012c523ddc952d46a431f27b2582a2c301d9e9da0c697a20ff6e2fb5d86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6zw9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gsx9l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:19Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:19 crc kubenswrapper[4860]: I1211 08:12:19.260222 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9872be1e-b995-4a00-904b-875dff96ad56\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f74a32a8fefea704cc73ffd1114083dab68aa44dc33f8d9d432d2d24216546c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbe00b46063d3968d59c393b87178e135f982abb7d0f32fa798b212d4b83df6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eedc9988e1a4a3c8cae730dc4a352103a6d665535b825e157d0f87c331b137c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2137fee4b7abc9765f9517d7b3b6ea796601a31d03151bf167d17136ced8b25a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f683e48f3577a68862eb6302383538376e41f49b20b1e0ea3f1b3b143f24872a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5de794187456b3576b969069a3b839d75c7918418a97fa35dff8bb40bb76e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5de794187456b3576b969069a3b839d75c7918418a97fa35dff8bb40bb76e39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08ce42b974fba00c409c0af971f0de351e50f18b2f06f8e722150da3239bb908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08ce42b974fba00c409c0af971f0de351e50f18b2f06f8e722150da3239bb908\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cfac6be03e3cfa0b0f3450b48b67957f2d5172b9c65e27f14c2a260bda5d423f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cfac6be03e3cfa0b0f3450b48b67957f2d5172b9c65e27f14c2a260bda5d423f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:19Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:19 crc kubenswrapper[4860]: I1211 08:12:19.274025 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-pznwl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5310c1fc-66c6-40aa-b1bf-5a59a2410f9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34119a6367261d7e12174f914ee8cfaa55703a55210f83c58b0e4694c54a5b03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2d6ff00ce322a28607bc3f6db05c5e4fb14acb6777ea315786ef53963ca504a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-11T08:12:17Z\\\",\\\"message\\\":\\\"2025-12-11T08:11:31+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_f37a703f-1609-4581-ab99-50337287ff33\\\\n2025-12-11T08:11:31+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_f37a703f-1609-4581-ab99-50337287ff33 to /host/opt/cni/bin/\\\\n2025-12-11T08:11:32Z [verbose] multus-daemon started\\\\n2025-12-11T08:11:32Z [verbose] Readiness Indicator file check\\\\n2025-12-11T08:12:17Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:12:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hkt48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-pznwl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:19Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:19 crc kubenswrapper[4860]: I1211 08:12:19.281083 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:19 crc kubenswrapper[4860]: I1211 08:12:19.281412 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:19 crc kubenswrapper[4860]: I1211 08:12:19.281508 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:19 crc kubenswrapper[4860]: I1211 08:12:19.281608 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:19 crc kubenswrapper[4860]: I1211 08:12:19.281711 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:19Z","lastTransitionTime":"2025-12-11T08:12:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:19 crc kubenswrapper[4860]: I1211 08:12:19.288128 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-854v6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e6529ba7-2341-43ac-a187-95ab530fe775\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8986538a421f0184f52f288b0e171da695591caf2b10c67ad2dce1073a6e6016\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kh22b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0d4e3f761ad15ae1424204e40b9d3f60188ab4a9047bd4b5ac6a8a8ca1fb385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kh22b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-854v6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:19Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:19 crc kubenswrapper[4860]: I1211 08:12:19.384091 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:19 crc kubenswrapper[4860]: I1211 08:12:19.384126 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:19 crc kubenswrapper[4860]: I1211 08:12:19.384134 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:19 crc kubenswrapper[4860]: I1211 08:12:19.384146 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:19 crc kubenswrapper[4860]: I1211 08:12:19.384154 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:19Z","lastTransitionTime":"2025-12-11T08:12:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:19 crc kubenswrapper[4860]: I1211 08:12:19.486055 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:19 crc kubenswrapper[4860]: I1211 08:12:19.486096 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:19 crc kubenswrapper[4860]: I1211 08:12:19.486104 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:19 crc kubenswrapper[4860]: I1211 08:12:19.486117 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:19 crc kubenswrapper[4860]: I1211 08:12:19.486126 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:19Z","lastTransitionTime":"2025-12-11T08:12:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:19 crc kubenswrapper[4860]: I1211 08:12:19.587876 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:19 crc kubenswrapper[4860]: I1211 08:12:19.587926 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:19 crc kubenswrapper[4860]: I1211 08:12:19.587938 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:19 crc kubenswrapper[4860]: I1211 08:12:19.587955 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:19 crc kubenswrapper[4860]: I1211 08:12:19.587968 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:19Z","lastTransitionTime":"2025-12-11T08:12:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:19 crc kubenswrapper[4860]: I1211 08:12:19.690425 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:19 crc kubenswrapper[4860]: I1211 08:12:19.690610 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:19 crc kubenswrapper[4860]: I1211 08:12:19.690713 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:19 crc kubenswrapper[4860]: I1211 08:12:19.690806 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:19 crc kubenswrapper[4860]: I1211 08:12:19.690909 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:19Z","lastTransitionTime":"2025-12-11T08:12:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:19 crc kubenswrapper[4860]: I1211 08:12:19.793283 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:19 crc kubenswrapper[4860]: I1211 08:12:19.793323 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:19 crc kubenswrapper[4860]: I1211 08:12:19.793334 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:19 crc kubenswrapper[4860]: I1211 08:12:19.793349 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:19 crc kubenswrapper[4860]: I1211 08:12:19.793360 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:19Z","lastTransitionTime":"2025-12-11T08:12:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:19 crc kubenswrapper[4860]: I1211 08:12:19.895245 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:19 crc kubenswrapper[4860]: I1211 08:12:19.895302 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:19 crc kubenswrapper[4860]: I1211 08:12:19.895327 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:19 crc kubenswrapper[4860]: I1211 08:12:19.895355 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:19 crc kubenswrapper[4860]: I1211 08:12:19.895375 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:19Z","lastTransitionTime":"2025-12-11T08:12:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:19 crc kubenswrapper[4860]: I1211 08:12:19.997304 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:19 crc kubenswrapper[4860]: I1211 08:12:19.997339 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:19 crc kubenswrapper[4860]: I1211 08:12:19.997351 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:19 crc kubenswrapper[4860]: I1211 08:12:19.997367 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:19 crc kubenswrapper[4860]: I1211 08:12:19.997378 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:19Z","lastTransitionTime":"2025-12-11T08:12:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:20 crc kubenswrapper[4860]: I1211 08:12:20.100617 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:20 crc kubenswrapper[4860]: I1211 08:12:20.100672 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:20 crc kubenswrapper[4860]: I1211 08:12:20.100684 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:20 crc kubenswrapper[4860]: I1211 08:12:20.100699 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:20 crc kubenswrapper[4860]: I1211 08:12:20.100713 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:20Z","lastTransitionTime":"2025-12-11T08:12:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:20 crc kubenswrapper[4860]: I1211 08:12:20.203483 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:20 crc kubenswrapper[4860]: I1211 08:12:20.203540 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:20 crc kubenswrapper[4860]: I1211 08:12:20.203553 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:20 crc kubenswrapper[4860]: I1211 08:12:20.203572 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:20 crc kubenswrapper[4860]: I1211 08:12:20.203586 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:20Z","lastTransitionTime":"2025-12-11T08:12:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:20 crc kubenswrapper[4860]: I1211 08:12:20.307860 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:20 crc kubenswrapper[4860]: I1211 08:12:20.308226 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:20 crc kubenswrapper[4860]: I1211 08:12:20.308372 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:20 crc kubenswrapper[4860]: I1211 08:12:20.308514 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:20 crc kubenswrapper[4860]: I1211 08:12:20.308672 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:20Z","lastTransitionTime":"2025-12-11T08:12:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:20 crc kubenswrapper[4860]: I1211 08:12:20.411676 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:20 crc kubenswrapper[4860]: I1211 08:12:20.411785 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:20 crc kubenswrapper[4860]: I1211 08:12:20.411816 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:20 crc kubenswrapper[4860]: I1211 08:12:20.411980 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:20 crc kubenswrapper[4860]: I1211 08:12:20.412033 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:20Z","lastTransitionTime":"2025-12-11T08:12:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:20 crc kubenswrapper[4860]: I1211 08:12:20.514608 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:20 crc kubenswrapper[4860]: I1211 08:12:20.514659 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:20 crc kubenswrapper[4860]: I1211 08:12:20.514668 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:20 crc kubenswrapper[4860]: I1211 08:12:20.514682 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:20 crc kubenswrapper[4860]: I1211 08:12:20.514692 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:20Z","lastTransitionTime":"2025-12-11T08:12:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:20 crc kubenswrapper[4860]: I1211 08:12:20.578354 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:12:20 crc kubenswrapper[4860]: I1211 08:12:20.578388 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gjhlj" Dec 11 08:12:20 crc kubenswrapper[4860]: I1211 08:12:20.578468 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:12:20 crc kubenswrapper[4860]: I1211 08:12:20.578515 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:12:20 crc kubenswrapper[4860]: E1211 08:12:20.578818 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:12:20 crc kubenswrapper[4860]: E1211 08:12:20.578984 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:12:20 crc kubenswrapper[4860]: E1211 08:12:20.579121 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gjhlj" podUID="5cfd640a-f3d2-4b28-87e1-5f49f6341e21" Dec 11 08:12:20 crc kubenswrapper[4860]: E1211 08:12:20.579319 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:12:20 crc kubenswrapper[4860]: I1211 08:12:20.617222 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:20 crc kubenswrapper[4860]: I1211 08:12:20.617261 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:20 crc kubenswrapper[4860]: I1211 08:12:20.617269 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:20 crc kubenswrapper[4860]: I1211 08:12:20.617283 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:20 crc kubenswrapper[4860]: I1211 08:12:20.617300 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:20Z","lastTransitionTime":"2025-12-11T08:12:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:20 crc kubenswrapper[4860]: I1211 08:12:20.721032 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:20 crc kubenswrapper[4860]: I1211 08:12:20.721074 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:20 crc kubenswrapper[4860]: I1211 08:12:20.721087 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:20 crc kubenswrapper[4860]: I1211 08:12:20.721107 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:20 crc kubenswrapper[4860]: I1211 08:12:20.721121 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:20Z","lastTransitionTime":"2025-12-11T08:12:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:20 crc kubenswrapper[4860]: I1211 08:12:20.823705 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:20 crc kubenswrapper[4860]: I1211 08:12:20.823746 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:20 crc kubenswrapper[4860]: I1211 08:12:20.823755 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:20 crc kubenswrapper[4860]: I1211 08:12:20.823771 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:20 crc kubenswrapper[4860]: I1211 08:12:20.823781 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:20Z","lastTransitionTime":"2025-12-11T08:12:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:20 crc kubenswrapper[4860]: I1211 08:12:20.926114 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:20 crc kubenswrapper[4860]: I1211 08:12:20.926168 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:20 crc kubenswrapper[4860]: I1211 08:12:20.926186 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:20 crc kubenswrapper[4860]: I1211 08:12:20.926209 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:20 crc kubenswrapper[4860]: I1211 08:12:20.926226 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:20Z","lastTransitionTime":"2025-12-11T08:12:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:21 crc kubenswrapper[4860]: I1211 08:12:21.028456 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:21 crc kubenswrapper[4860]: I1211 08:12:21.028531 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:21 crc kubenswrapper[4860]: I1211 08:12:21.028548 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:21 crc kubenswrapper[4860]: I1211 08:12:21.028573 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:21 crc kubenswrapper[4860]: I1211 08:12:21.028591 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:21Z","lastTransitionTime":"2025-12-11T08:12:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:21 crc kubenswrapper[4860]: I1211 08:12:21.131247 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:21 crc kubenswrapper[4860]: I1211 08:12:21.131292 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:21 crc kubenswrapper[4860]: I1211 08:12:21.131308 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:21 crc kubenswrapper[4860]: I1211 08:12:21.131330 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:21 crc kubenswrapper[4860]: I1211 08:12:21.131345 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:21Z","lastTransitionTime":"2025-12-11T08:12:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:21 crc kubenswrapper[4860]: I1211 08:12:21.233588 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:21 crc kubenswrapper[4860]: I1211 08:12:21.233714 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:21 crc kubenswrapper[4860]: I1211 08:12:21.233739 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:21 crc kubenswrapper[4860]: I1211 08:12:21.233763 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:21 crc kubenswrapper[4860]: I1211 08:12:21.233780 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:21Z","lastTransitionTime":"2025-12-11T08:12:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:21 crc kubenswrapper[4860]: I1211 08:12:21.336044 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:21 crc kubenswrapper[4860]: I1211 08:12:21.336116 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:21 crc kubenswrapper[4860]: I1211 08:12:21.336130 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:21 crc kubenswrapper[4860]: I1211 08:12:21.336151 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:21 crc kubenswrapper[4860]: I1211 08:12:21.336163 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:21Z","lastTransitionTime":"2025-12-11T08:12:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:21 crc kubenswrapper[4860]: I1211 08:12:21.438748 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:21 crc kubenswrapper[4860]: I1211 08:12:21.438812 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:21 crc kubenswrapper[4860]: I1211 08:12:21.438826 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:21 crc kubenswrapper[4860]: I1211 08:12:21.438849 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:21 crc kubenswrapper[4860]: I1211 08:12:21.438864 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:21Z","lastTransitionTime":"2025-12-11T08:12:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:21 crc kubenswrapper[4860]: I1211 08:12:21.541681 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:21 crc kubenswrapper[4860]: I1211 08:12:21.541743 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:21 crc kubenswrapper[4860]: I1211 08:12:21.541760 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:21 crc kubenswrapper[4860]: I1211 08:12:21.541784 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:21 crc kubenswrapper[4860]: I1211 08:12:21.541803 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:21Z","lastTransitionTime":"2025-12-11T08:12:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:21 crc kubenswrapper[4860]: I1211 08:12:21.644223 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:21 crc kubenswrapper[4860]: I1211 08:12:21.644300 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:21 crc kubenswrapper[4860]: I1211 08:12:21.644320 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:21 crc kubenswrapper[4860]: I1211 08:12:21.644347 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:21 crc kubenswrapper[4860]: I1211 08:12:21.644368 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:21Z","lastTransitionTime":"2025-12-11T08:12:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:21 crc kubenswrapper[4860]: I1211 08:12:21.746612 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:21 crc kubenswrapper[4860]: I1211 08:12:21.746702 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:21 crc kubenswrapper[4860]: I1211 08:12:21.746721 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:21 crc kubenswrapper[4860]: I1211 08:12:21.746745 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:21 crc kubenswrapper[4860]: I1211 08:12:21.746764 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:21Z","lastTransitionTime":"2025-12-11T08:12:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:21 crc kubenswrapper[4860]: I1211 08:12:21.765109 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:21 crc kubenswrapper[4860]: I1211 08:12:21.765358 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:21 crc kubenswrapper[4860]: I1211 08:12:21.765380 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:21 crc kubenswrapper[4860]: I1211 08:12:21.765405 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:21 crc kubenswrapper[4860]: I1211 08:12:21.765429 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:21Z","lastTransitionTime":"2025-12-11T08:12:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:21 crc kubenswrapper[4860]: E1211 08:12:21.787125 4860 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:12:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:12:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:12:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:12:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"52631a1a-bc27-4dd5-8309-5dc39b5dcc41\\\",\\\"systemUUID\\\":\\\"299d3cd2-e1ad-4f5c-8af4-f12c0a5d76e7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:21Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:21 crc kubenswrapper[4860]: I1211 08:12:21.792516 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:21 crc kubenswrapper[4860]: I1211 08:12:21.792601 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:21 crc kubenswrapper[4860]: I1211 08:12:21.792626 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:21 crc kubenswrapper[4860]: I1211 08:12:21.792704 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:21 crc kubenswrapper[4860]: I1211 08:12:21.792732 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:21Z","lastTransitionTime":"2025-12-11T08:12:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:21 crc kubenswrapper[4860]: E1211 08:12:21.814252 4860 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:12:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:12:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:12:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:12:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"52631a1a-bc27-4dd5-8309-5dc39b5dcc41\\\",\\\"systemUUID\\\":\\\"299d3cd2-e1ad-4f5c-8af4-f12c0a5d76e7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:21Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:21 crc kubenswrapper[4860]: I1211 08:12:21.817593 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:21 crc kubenswrapper[4860]: I1211 08:12:21.817637 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:21 crc kubenswrapper[4860]: I1211 08:12:21.817672 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:21 crc kubenswrapper[4860]: I1211 08:12:21.817692 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:21 crc kubenswrapper[4860]: I1211 08:12:21.817706 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:21Z","lastTransitionTime":"2025-12-11T08:12:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:21 crc kubenswrapper[4860]: E1211 08:12:21.837324 4860 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:12:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:12:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:12:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:12:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"52631a1a-bc27-4dd5-8309-5dc39b5dcc41\\\",\\\"systemUUID\\\":\\\"299d3cd2-e1ad-4f5c-8af4-f12c0a5d76e7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:21Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:21 crc kubenswrapper[4860]: I1211 08:12:21.841118 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:21 crc kubenswrapper[4860]: I1211 08:12:21.841173 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:21 crc kubenswrapper[4860]: I1211 08:12:21.841191 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:21 crc kubenswrapper[4860]: I1211 08:12:21.841217 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:21 crc kubenswrapper[4860]: I1211 08:12:21.841235 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:21Z","lastTransitionTime":"2025-12-11T08:12:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:21 crc kubenswrapper[4860]: E1211 08:12:21.856405 4860 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:12:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:12:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:12:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:12:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"52631a1a-bc27-4dd5-8309-5dc39b5dcc41\\\",\\\"systemUUID\\\":\\\"299d3cd2-e1ad-4f5c-8af4-f12c0a5d76e7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:21Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:21 crc kubenswrapper[4860]: I1211 08:12:21.860165 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:21 crc kubenswrapper[4860]: I1211 08:12:21.860207 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:21 crc kubenswrapper[4860]: I1211 08:12:21.860219 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:21 crc kubenswrapper[4860]: I1211 08:12:21.860236 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:21 crc kubenswrapper[4860]: I1211 08:12:21.860249 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:21Z","lastTransitionTime":"2025-12-11T08:12:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:21 crc kubenswrapper[4860]: E1211 08:12:21.876031 4860 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:12:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:12:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:12:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:12:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"52631a1a-bc27-4dd5-8309-5dc39b5dcc41\\\",\\\"systemUUID\\\":\\\"299d3cd2-e1ad-4f5c-8af4-f12c0a5d76e7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:21Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:21 crc kubenswrapper[4860]: E1211 08:12:21.876390 4860 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 11 08:12:21 crc kubenswrapper[4860]: I1211 08:12:21.878196 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:21 crc kubenswrapper[4860]: I1211 08:12:21.878228 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:21 crc kubenswrapper[4860]: I1211 08:12:21.878239 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:21 crc kubenswrapper[4860]: I1211 08:12:21.878257 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:21 crc kubenswrapper[4860]: I1211 08:12:21.878271 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:21Z","lastTransitionTime":"2025-12-11T08:12:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:21 crc kubenswrapper[4860]: I1211 08:12:21.980707 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:21 crc kubenswrapper[4860]: I1211 08:12:21.980750 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:21 crc kubenswrapper[4860]: I1211 08:12:21.980785 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:21 crc kubenswrapper[4860]: I1211 08:12:21.980819 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:21 crc kubenswrapper[4860]: I1211 08:12:21.980833 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:21Z","lastTransitionTime":"2025-12-11T08:12:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:22 crc kubenswrapper[4860]: I1211 08:12:22.083876 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:22 crc kubenswrapper[4860]: I1211 08:12:22.083948 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:22 crc kubenswrapper[4860]: I1211 08:12:22.083968 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:22 crc kubenswrapper[4860]: I1211 08:12:22.083992 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:22 crc kubenswrapper[4860]: I1211 08:12:22.084010 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:22Z","lastTransitionTime":"2025-12-11T08:12:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:22 crc kubenswrapper[4860]: I1211 08:12:22.186892 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:22 crc kubenswrapper[4860]: I1211 08:12:22.186927 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:22 crc kubenswrapper[4860]: I1211 08:12:22.186937 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:22 crc kubenswrapper[4860]: I1211 08:12:22.186951 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:22 crc kubenswrapper[4860]: I1211 08:12:22.186960 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:22Z","lastTransitionTime":"2025-12-11T08:12:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:22 crc kubenswrapper[4860]: I1211 08:12:22.290356 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:22 crc kubenswrapper[4860]: I1211 08:12:22.290440 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:22 crc kubenswrapper[4860]: I1211 08:12:22.290450 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:22 crc kubenswrapper[4860]: I1211 08:12:22.290470 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:22 crc kubenswrapper[4860]: I1211 08:12:22.290490 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:22Z","lastTransitionTime":"2025-12-11T08:12:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:22 crc kubenswrapper[4860]: I1211 08:12:22.393386 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:22 crc kubenswrapper[4860]: I1211 08:12:22.393424 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:22 crc kubenswrapper[4860]: I1211 08:12:22.393436 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:22 crc kubenswrapper[4860]: I1211 08:12:22.393452 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:22 crc kubenswrapper[4860]: I1211 08:12:22.393466 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:22Z","lastTransitionTime":"2025-12-11T08:12:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:22 crc kubenswrapper[4860]: I1211 08:12:22.496456 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:22 crc kubenswrapper[4860]: I1211 08:12:22.496849 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:22 crc kubenswrapper[4860]: I1211 08:12:22.496941 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:22 crc kubenswrapper[4860]: I1211 08:12:22.497031 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:22 crc kubenswrapper[4860]: I1211 08:12:22.497161 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:22Z","lastTransitionTime":"2025-12-11T08:12:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:22 crc kubenswrapper[4860]: I1211 08:12:22.578787 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:12:22 crc kubenswrapper[4860]: I1211 08:12:22.578833 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gjhlj" Dec 11 08:12:22 crc kubenswrapper[4860]: I1211 08:12:22.578897 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:12:22 crc kubenswrapper[4860]: I1211 08:12:22.578982 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:12:22 crc kubenswrapper[4860]: E1211 08:12:22.579257 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:12:22 crc kubenswrapper[4860]: E1211 08:12:22.579508 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gjhlj" podUID="5cfd640a-f3d2-4b28-87e1-5f49f6341e21" Dec 11 08:12:22 crc kubenswrapper[4860]: E1211 08:12:22.579582 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:12:22 crc kubenswrapper[4860]: E1211 08:12:22.579704 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:12:22 crc kubenswrapper[4860]: I1211 08:12:22.600817 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:22 crc kubenswrapper[4860]: I1211 08:12:22.600910 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:22 crc kubenswrapper[4860]: I1211 08:12:22.600928 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:22 crc kubenswrapper[4860]: I1211 08:12:22.600953 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:22 crc kubenswrapper[4860]: I1211 08:12:22.600970 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:22Z","lastTransitionTime":"2025-12-11T08:12:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:22 crc kubenswrapper[4860]: I1211 08:12:22.704315 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:22 crc kubenswrapper[4860]: I1211 08:12:22.704611 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:22 crc kubenswrapper[4860]: I1211 08:12:22.704858 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:22 crc kubenswrapper[4860]: I1211 08:12:22.705049 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:22 crc kubenswrapper[4860]: I1211 08:12:22.705304 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:22Z","lastTransitionTime":"2025-12-11T08:12:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:22 crc kubenswrapper[4860]: I1211 08:12:22.808081 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:22 crc kubenswrapper[4860]: I1211 08:12:22.808203 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:22 crc kubenswrapper[4860]: I1211 08:12:22.808227 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:22 crc kubenswrapper[4860]: I1211 08:12:22.808255 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:22 crc kubenswrapper[4860]: I1211 08:12:22.808275 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:22Z","lastTransitionTime":"2025-12-11T08:12:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:22 crc kubenswrapper[4860]: I1211 08:12:22.912746 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:22 crc kubenswrapper[4860]: I1211 08:12:22.912811 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:22 crc kubenswrapper[4860]: I1211 08:12:22.912832 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:22 crc kubenswrapper[4860]: I1211 08:12:22.912862 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:22 crc kubenswrapper[4860]: I1211 08:12:22.912884 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:22Z","lastTransitionTime":"2025-12-11T08:12:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:23 crc kubenswrapper[4860]: I1211 08:12:23.015606 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:23 crc kubenswrapper[4860]: I1211 08:12:23.015695 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:23 crc kubenswrapper[4860]: I1211 08:12:23.015712 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:23 crc kubenswrapper[4860]: I1211 08:12:23.015735 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:23 crc kubenswrapper[4860]: I1211 08:12:23.015752 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:23Z","lastTransitionTime":"2025-12-11T08:12:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:23 crc kubenswrapper[4860]: I1211 08:12:23.118764 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:23 crc kubenswrapper[4860]: I1211 08:12:23.118872 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:23 crc kubenswrapper[4860]: I1211 08:12:23.118891 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:23 crc kubenswrapper[4860]: I1211 08:12:23.118918 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:23 crc kubenswrapper[4860]: I1211 08:12:23.118940 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:23Z","lastTransitionTime":"2025-12-11T08:12:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:23 crc kubenswrapper[4860]: I1211 08:12:23.222290 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:23 crc kubenswrapper[4860]: I1211 08:12:23.222364 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:23 crc kubenswrapper[4860]: I1211 08:12:23.222381 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:23 crc kubenswrapper[4860]: I1211 08:12:23.222405 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:23 crc kubenswrapper[4860]: I1211 08:12:23.222422 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:23Z","lastTransitionTime":"2025-12-11T08:12:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:23 crc kubenswrapper[4860]: I1211 08:12:23.325895 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:23 crc kubenswrapper[4860]: I1211 08:12:23.325954 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:23 crc kubenswrapper[4860]: I1211 08:12:23.325974 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:23 crc kubenswrapper[4860]: I1211 08:12:23.325998 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:23 crc kubenswrapper[4860]: I1211 08:12:23.326015 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:23Z","lastTransitionTime":"2025-12-11T08:12:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:23 crc kubenswrapper[4860]: I1211 08:12:23.429107 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:23 crc kubenswrapper[4860]: I1211 08:12:23.429211 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:23 crc kubenswrapper[4860]: I1211 08:12:23.429233 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:23 crc kubenswrapper[4860]: I1211 08:12:23.429260 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:23 crc kubenswrapper[4860]: I1211 08:12:23.429279 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:23Z","lastTransitionTime":"2025-12-11T08:12:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:23 crc kubenswrapper[4860]: I1211 08:12:23.531828 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:23 crc kubenswrapper[4860]: I1211 08:12:23.531935 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:23 crc kubenswrapper[4860]: I1211 08:12:23.531949 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:23 crc kubenswrapper[4860]: I1211 08:12:23.531963 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:23 crc kubenswrapper[4860]: I1211 08:12:23.531973 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:23Z","lastTransitionTime":"2025-12-11T08:12:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:23 crc kubenswrapper[4860]: I1211 08:12:23.634247 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:23 crc kubenswrapper[4860]: I1211 08:12:23.634291 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:23 crc kubenswrapper[4860]: I1211 08:12:23.634300 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:23 crc kubenswrapper[4860]: I1211 08:12:23.634313 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:23 crc kubenswrapper[4860]: I1211 08:12:23.634322 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:23Z","lastTransitionTime":"2025-12-11T08:12:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:23 crc kubenswrapper[4860]: I1211 08:12:23.736806 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:23 crc kubenswrapper[4860]: I1211 08:12:23.736861 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:23 crc kubenswrapper[4860]: I1211 08:12:23.736872 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:23 crc kubenswrapper[4860]: I1211 08:12:23.736889 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:23 crc kubenswrapper[4860]: I1211 08:12:23.736900 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:23Z","lastTransitionTime":"2025-12-11T08:12:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:23 crc kubenswrapper[4860]: I1211 08:12:23.838789 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:23 crc kubenswrapper[4860]: I1211 08:12:23.838843 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:23 crc kubenswrapper[4860]: I1211 08:12:23.838854 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:23 crc kubenswrapper[4860]: I1211 08:12:23.838872 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:23 crc kubenswrapper[4860]: I1211 08:12:23.838889 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:23Z","lastTransitionTime":"2025-12-11T08:12:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:23 crc kubenswrapper[4860]: I1211 08:12:23.941821 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:23 crc kubenswrapper[4860]: I1211 08:12:23.941890 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:23 crc kubenswrapper[4860]: I1211 08:12:23.941908 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:23 crc kubenswrapper[4860]: I1211 08:12:23.941932 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:23 crc kubenswrapper[4860]: I1211 08:12:23.941951 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:23Z","lastTransitionTime":"2025-12-11T08:12:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:24 crc kubenswrapper[4860]: I1211 08:12:24.044836 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:24 crc kubenswrapper[4860]: I1211 08:12:24.044870 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:24 crc kubenswrapper[4860]: I1211 08:12:24.044881 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:24 crc kubenswrapper[4860]: I1211 08:12:24.044896 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:24 crc kubenswrapper[4860]: I1211 08:12:24.044907 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:24Z","lastTransitionTime":"2025-12-11T08:12:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:24 crc kubenswrapper[4860]: I1211 08:12:24.147805 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:24 crc kubenswrapper[4860]: I1211 08:12:24.147889 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:24 crc kubenswrapper[4860]: I1211 08:12:24.147913 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:24 crc kubenswrapper[4860]: I1211 08:12:24.147947 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:24 crc kubenswrapper[4860]: I1211 08:12:24.148003 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:24Z","lastTransitionTime":"2025-12-11T08:12:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:24 crc kubenswrapper[4860]: I1211 08:12:24.251345 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:24 crc kubenswrapper[4860]: I1211 08:12:24.251392 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:24 crc kubenswrapper[4860]: I1211 08:12:24.251402 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:24 crc kubenswrapper[4860]: I1211 08:12:24.251421 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:24 crc kubenswrapper[4860]: I1211 08:12:24.251432 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:24Z","lastTransitionTime":"2025-12-11T08:12:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:24 crc kubenswrapper[4860]: I1211 08:12:24.354990 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:24 crc kubenswrapper[4860]: I1211 08:12:24.355038 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:24 crc kubenswrapper[4860]: I1211 08:12:24.355049 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:24 crc kubenswrapper[4860]: I1211 08:12:24.355070 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:24 crc kubenswrapper[4860]: I1211 08:12:24.355083 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:24Z","lastTransitionTime":"2025-12-11T08:12:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:24 crc kubenswrapper[4860]: I1211 08:12:24.458154 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:24 crc kubenswrapper[4860]: I1211 08:12:24.458253 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:24 crc kubenswrapper[4860]: I1211 08:12:24.458282 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:24 crc kubenswrapper[4860]: I1211 08:12:24.458313 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:24 crc kubenswrapper[4860]: I1211 08:12:24.458336 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:24Z","lastTransitionTime":"2025-12-11T08:12:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:24 crc kubenswrapper[4860]: I1211 08:12:24.561523 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:24 crc kubenswrapper[4860]: I1211 08:12:24.561602 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:24 crc kubenswrapper[4860]: I1211 08:12:24.561626 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:24 crc kubenswrapper[4860]: I1211 08:12:24.561708 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:24 crc kubenswrapper[4860]: I1211 08:12:24.561736 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:24Z","lastTransitionTime":"2025-12-11T08:12:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:24 crc kubenswrapper[4860]: I1211 08:12:24.577789 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gjhlj" Dec 11 08:12:24 crc kubenswrapper[4860]: I1211 08:12:24.577809 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:12:24 crc kubenswrapper[4860]: I1211 08:12:24.577863 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:12:24 crc kubenswrapper[4860]: I1211 08:12:24.578117 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:12:24 crc kubenswrapper[4860]: E1211 08:12:24.578124 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gjhlj" podUID="5cfd640a-f3d2-4b28-87e1-5f49f6341e21" Dec 11 08:12:24 crc kubenswrapper[4860]: E1211 08:12:24.578233 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:12:24 crc kubenswrapper[4860]: E1211 08:12:24.578310 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:12:24 crc kubenswrapper[4860]: E1211 08:12:24.578387 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:12:24 crc kubenswrapper[4860]: I1211 08:12:24.664775 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:24 crc kubenswrapper[4860]: I1211 08:12:24.664840 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:24 crc kubenswrapper[4860]: I1211 08:12:24.664864 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:24 crc kubenswrapper[4860]: I1211 08:12:24.664906 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:24 crc kubenswrapper[4860]: I1211 08:12:24.664973 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:24Z","lastTransitionTime":"2025-12-11T08:12:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:24 crc kubenswrapper[4860]: I1211 08:12:24.769098 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:24 crc kubenswrapper[4860]: I1211 08:12:24.769194 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:24 crc kubenswrapper[4860]: I1211 08:12:24.769220 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:24 crc kubenswrapper[4860]: I1211 08:12:24.769251 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:24 crc kubenswrapper[4860]: I1211 08:12:24.769274 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:24Z","lastTransitionTime":"2025-12-11T08:12:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:24 crc kubenswrapper[4860]: I1211 08:12:24.872027 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:24 crc kubenswrapper[4860]: I1211 08:12:24.872084 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:24 crc kubenswrapper[4860]: I1211 08:12:24.872103 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:24 crc kubenswrapper[4860]: I1211 08:12:24.872128 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:24 crc kubenswrapper[4860]: I1211 08:12:24.872144 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:24Z","lastTransitionTime":"2025-12-11T08:12:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:24 crc kubenswrapper[4860]: I1211 08:12:24.974635 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:24 crc kubenswrapper[4860]: I1211 08:12:24.974712 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:24 crc kubenswrapper[4860]: I1211 08:12:24.974728 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:24 crc kubenswrapper[4860]: I1211 08:12:24.974750 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:24 crc kubenswrapper[4860]: I1211 08:12:24.974768 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:24Z","lastTransitionTime":"2025-12-11T08:12:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:25 crc kubenswrapper[4860]: I1211 08:12:25.081804 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:25 crc kubenswrapper[4860]: I1211 08:12:25.081865 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:25 crc kubenswrapper[4860]: I1211 08:12:25.081885 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:25 crc kubenswrapper[4860]: I1211 08:12:25.081914 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:25 crc kubenswrapper[4860]: I1211 08:12:25.081933 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:25Z","lastTransitionTime":"2025-12-11T08:12:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:25 crc kubenswrapper[4860]: I1211 08:12:25.185637 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:25 crc kubenswrapper[4860]: I1211 08:12:25.185785 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:25 crc kubenswrapper[4860]: I1211 08:12:25.185808 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:25 crc kubenswrapper[4860]: I1211 08:12:25.185885 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:25 crc kubenswrapper[4860]: I1211 08:12:25.185910 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:25Z","lastTransitionTime":"2025-12-11T08:12:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:25 crc kubenswrapper[4860]: I1211 08:12:25.289545 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:25 crc kubenswrapper[4860]: I1211 08:12:25.289600 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:25 crc kubenswrapper[4860]: I1211 08:12:25.289611 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:25 crc kubenswrapper[4860]: I1211 08:12:25.289629 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:25 crc kubenswrapper[4860]: I1211 08:12:25.289659 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:25Z","lastTransitionTime":"2025-12-11T08:12:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:25 crc kubenswrapper[4860]: I1211 08:12:25.392416 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:25 crc kubenswrapper[4860]: I1211 08:12:25.392485 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:25 crc kubenswrapper[4860]: I1211 08:12:25.392506 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:25 crc kubenswrapper[4860]: I1211 08:12:25.392533 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:25 crc kubenswrapper[4860]: I1211 08:12:25.392555 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:25Z","lastTransitionTime":"2025-12-11T08:12:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:25 crc kubenswrapper[4860]: I1211 08:12:25.495708 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:25 crc kubenswrapper[4860]: I1211 08:12:25.495745 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:25 crc kubenswrapper[4860]: I1211 08:12:25.495755 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:25 crc kubenswrapper[4860]: I1211 08:12:25.495767 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:25 crc kubenswrapper[4860]: I1211 08:12:25.495776 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:25Z","lastTransitionTime":"2025-12-11T08:12:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:25 crc kubenswrapper[4860]: I1211 08:12:25.598397 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:25 crc kubenswrapper[4860]: I1211 08:12:25.598499 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:25 crc kubenswrapper[4860]: I1211 08:12:25.598550 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:25 crc kubenswrapper[4860]: I1211 08:12:25.598673 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:25 crc kubenswrapper[4860]: I1211 08:12:25.598704 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:25Z","lastTransitionTime":"2025-12-11T08:12:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:25 crc kubenswrapper[4860]: I1211 08:12:25.702069 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:25 crc kubenswrapper[4860]: I1211 08:12:25.702131 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:25 crc kubenswrapper[4860]: I1211 08:12:25.702148 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:25 crc kubenswrapper[4860]: I1211 08:12:25.702171 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:25 crc kubenswrapper[4860]: I1211 08:12:25.702189 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:25Z","lastTransitionTime":"2025-12-11T08:12:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:25 crc kubenswrapper[4860]: I1211 08:12:25.804725 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:25 crc kubenswrapper[4860]: I1211 08:12:25.804762 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:25 crc kubenswrapper[4860]: I1211 08:12:25.804772 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:25 crc kubenswrapper[4860]: I1211 08:12:25.804785 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:25 crc kubenswrapper[4860]: I1211 08:12:25.804794 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:25Z","lastTransitionTime":"2025-12-11T08:12:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:25 crc kubenswrapper[4860]: I1211 08:12:25.907624 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:25 crc kubenswrapper[4860]: I1211 08:12:25.907732 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:25 crc kubenswrapper[4860]: I1211 08:12:25.907749 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:25 crc kubenswrapper[4860]: I1211 08:12:25.907775 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:25 crc kubenswrapper[4860]: I1211 08:12:25.907792 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:25Z","lastTransitionTime":"2025-12-11T08:12:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:26 crc kubenswrapper[4860]: I1211 08:12:26.010965 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:26 crc kubenswrapper[4860]: I1211 08:12:26.011018 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:26 crc kubenswrapper[4860]: I1211 08:12:26.011030 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:26 crc kubenswrapper[4860]: I1211 08:12:26.011047 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:26 crc kubenswrapper[4860]: I1211 08:12:26.011064 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:26Z","lastTransitionTime":"2025-12-11T08:12:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:26 crc kubenswrapper[4860]: I1211 08:12:26.114389 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:26 crc kubenswrapper[4860]: I1211 08:12:26.114451 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:26 crc kubenswrapper[4860]: I1211 08:12:26.114470 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:26 crc kubenswrapper[4860]: I1211 08:12:26.114495 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:26 crc kubenswrapper[4860]: I1211 08:12:26.114513 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:26Z","lastTransitionTime":"2025-12-11T08:12:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:26 crc kubenswrapper[4860]: I1211 08:12:26.217585 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:26 crc kubenswrapper[4860]: I1211 08:12:26.217687 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:26 crc kubenswrapper[4860]: I1211 08:12:26.217711 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:26 crc kubenswrapper[4860]: I1211 08:12:26.217738 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:26 crc kubenswrapper[4860]: I1211 08:12:26.217760 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:26Z","lastTransitionTime":"2025-12-11T08:12:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:26 crc kubenswrapper[4860]: I1211 08:12:26.354178 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:26 crc kubenswrapper[4860]: I1211 08:12:26.354258 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:26 crc kubenswrapper[4860]: I1211 08:12:26.354300 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:26 crc kubenswrapper[4860]: I1211 08:12:26.354331 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:26 crc kubenswrapper[4860]: I1211 08:12:26.354350 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:26Z","lastTransitionTime":"2025-12-11T08:12:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:26 crc kubenswrapper[4860]: I1211 08:12:26.457844 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:26 crc kubenswrapper[4860]: I1211 08:12:26.457910 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:26 crc kubenswrapper[4860]: I1211 08:12:26.457929 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:26 crc kubenswrapper[4860]: I1211 08:12:26.457953 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:26 crc kubenswrapper[4860]: I1211 08:12:26.457970 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:26Z","lastTransitionTime":"2025-12-11T08:12:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:26 crc kubenswrapper[4860]: I1211 08:12:26.560790 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:26 crc kubenswrapper[4860]: I1211 08:12:26.560838 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:26 crc kubenswrapper[4860]: I1211 08:12:26.560854 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:26 crc kubenswrapper[4860]: I1211 08:12:26.560875 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:26 crc kubenswrapper[4860]: I1211 08:12:26.560888 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:26Z","lastTransitionTime":"2025-12-11T08:12:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:26 crc kubenswrapper[4860]: I1211 08:12:26.577898 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:12:26 crc kubenswrapper[4860]: I1211 08:12:26.577898 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gjhlj" Dec 11 08:12:26 crc kubenswrapper[4860]: I1211 08:12:26.577916 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:12:26 crc kubenswrapper[4860]: I1211 08:12:26.578042 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:12:26 crc kubenswrapper[4860]: E1211 08:12:26.578246 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:12:26 crc kubenswrapper[4860]: E1211 08:12:26.578485 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gjhlj" podUID="5cfd640a-f3d2-4b28-87e1-5f49f6341e21" Dec 11 08:12:26 crc kubenswrapper[4860]: E1211 08:12:26.578550 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:12:26 crc kubenswrapper[4860]: E1211 08:12:26.578628 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:12:26 crc kubenswrapper[4860]: I1211 08:12:26.663477 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:26 crc kubenswrapper[4860]: I1211 08:12:26.663556 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:26 crc kubenswrapper[4860]: I1211 08:12:26.663584 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:26 crc kubenswrapper[4860]: I1211 08:12:26.663611 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:26 crc kubenswrapper[4860]: I1211 08:12:26.663632 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:26Z","lastTransitionTime":"2025-12-11T08:12:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:26 crc kubenswrapper[4860]: I1211 08:12:26.766999 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:26 crc kubenswrapper[4860]: I1211 08:12:26.767071 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:26 crc kubenswrapper[4860]: I1211 08:12:26.767097 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:26 crc kubenswrapper[4860]: I1211 08:12:26.767125 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:26 crc kubenswrapper[4860]: I1211 08:12:26.767147 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:26Z","lastTransitionTime":"2025-12-11T08:12:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:26 crc kubenswrapper[4860]: I1211 08:12:26.870152 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:26 crc kubenswrapper[4860]: I1211 08:12:26.870203 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:26 crc kubenswrapper[4860]: I1211 08:12:26.870227 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:26 crc kubenswrapper[4860]: I1211 08:12:26.870258 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:26 crc kubenswrapper[4860]: I1211 08:12:26.870281 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:26Z","lastTransitionTime":"2025-12-11T08:12:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:26 crc kubenswrapper[4860]: I1211 08:12:26.972866 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:26 crc kubenswrapper[4860]: I1211 08:12:26.972912 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:26 crc kubenswrapper[4860]: I1211 08:12:26.972922 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:26 crc kubenswrapper[4860]: I1211 08:12:26.972938 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:26 crc kubenswrapper[4860]: I1211 08:12:26.972947 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:26Z","lastTransitionTime":"2025-12-11T08:12:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:27 crc kubenswrapper[4860]: I1211 08:12:27.075518 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:27 crc kubenswrapper[4860]: I1211 08:12:27.075571 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:27 crc kubenswrapper[4860]: I1211 08:12:27.075588 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:27 crc kubenswrapper[4860]: I1211 08:12:27.075612 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:27 crc kubenswrapper[4860]: I1211 08:12:27.075630 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:27Z","lastTransitionTime":"2025-12-11T08:12:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:27 crc kubenswrapper[4860]: I1211 08:12:27.179533 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:27 crc kubenswrapper[4860]: I1211 08:12:27.179585 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:27 crc kubenswrapper[4860]: I1211 08:12:27.179601 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:27 crc kubenswrapper[4860]: I1211 08:12:27.179625 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:27 crc kubenswrapper[4860]: I1211 08:12:27.179676 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:27Z","lastTransitionTime":"2025-12-11T08:12:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:27 crc kubenswrapper[4860]: I1211 08:12:27.283505 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:27 crc kubenswrapper[4860]: I1211 08:12:27.283569 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:27 crc kubenswrapper[4860]: I1211 08:12:27.283585 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:27 crc kubenswrapper[4860]: I1211 08:12:27.283614 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:27 crc kubenswrapper[4860]: I1211 08:12:27.283631 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:27Z","lastTransitionTime":"2025-12-11T08:12:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:27 crc kubenswrapper[4860]: I1211 08:12:27.387354 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:27 crc kubenswrapper[4860]: I1211 08:12:27.387415 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:27 crc kubenswrapper[4860]: I1211 08:12:27.387431 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:27 crc kubenswrapper[4860]: I1211 08:12:27.387455 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:27 crc kubenswrapper[4860]: I1211 08:12:27.387471 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:27Z","lastTransitionTime":"2025-12-11T08:12:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:27 crc kubenswrapper[4860]: I1211 08:12:27.490178 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:27 crc kubenswrapper[4860]: I1211 08:12:27.490244 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:27 crc kubenswrapper[4860]: I1211 08:12:27.490261 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:27 crc kubenswrapper[4860]: I1211 08:12:27.490285 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:27 crc kubenswrapper[4860]: I1211 08:12:27.490302 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:27Z","lastTransitionTime":"2025-12-11T08:12:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:27 crc kubenswrapper[4860]: I1211 08:12:27.593519 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:27 crc kubenswrapper[4860]: I1211 08:12:27.593628 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:27 crc kubenswrapper[4860]: I1211 08:12:27.593724 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:27 crc kubenswrapper[4860]: I1211 08:12:27.593757 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:27 crc kubenswrapper[4860]: I1211 08:12:27.593779 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:27Z","lastTransitionTime":"2025-12-11T08:12:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:27 crc kubenswrapper[4860]: I1211 08:12:27.602379 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ec3758-85f1-4a22-a4ec-c805472607a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e831db1a15976d3967d56479689572c26f2599a766baa0402897a3008428c56f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a498aa90f914e1f0370df01dc5ef174df113a2a16d86b2e3f24b19a1b99b4c0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8caebd2f4dfcbb89c597029a07c335171889761d2d200206987537fb0a228db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34e91f615b76445ff3241a304bb2964fde59de7e8a8dfff6841b2e628a31e86d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2090cab02ad74ddfd6bb57ff50eba35d18f56b0cde1c230fe011c01994b693bc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 08:11:21.171011 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 08:11:21.177674 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-644352688/tls.crt::/tmp/serving-cert-644352688/tls.key\\\\\\\"\\\\nI1211 08:11:26.508606 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 08:11:26.512741 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 08:11:26.512778 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 08:11:26.512839 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 08:11:26.512858 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 08:11:26.521684 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1211 08:11:26.521700 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1211 08:11:26.521722 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:11:26.521732 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:11:26.521738 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 08:11:26.521746 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 08:11:26.521749 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 08:11:26.521753 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1211 08:11:26.524999 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2274fb298ca7d2b37b060b4dbf9b166d39ba77a108b08edeeb4d67196e72cb3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf0aff8a6ab816b6617b1feeca9146d41bb5e8808bf1db96f8d493500c135d85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf0aff8a6ab816b6617b1feeca9146d41bb5e8808bf1db96f8d493500c135d85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:27Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:27 crc kubenswrapper[4860]: I1211 08:12:27.623339 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:27Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:27 crc kubenswrapper[4860]: I1211 08:12:27.643706 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a41f35db087418d1be16953f10377042abd3f4488e5ac1bc29ca5dcfc00a37e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:27Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:27 crc kubenswrapper[4860]: I1211 08:12:27.662199 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gjhlj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5cfd640a-f3d2-4b28-87e1-5f49f6341e21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knq7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knq7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:42Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gjhlj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:27Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:27 crc kubenswrapper[4860]: I1211 08:12:27.674587 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gsx9l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aba5b6b9-f88d-4763-9562-6b4dec606df9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://735ef012c523ddc952d46a431f27b2582a2c301d9e9da0c697a20ff6e2fb5d86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6zw9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gsx9l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:27Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:27 crc kubenswrapper[4860]: I1211 08:12:27.695018 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8bf0e9df-4e9d-4f45-a5b9-0020fdab2ea6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77c0c831f046149002cd12564aa9b1544a813da1fcd81e61451eeb504f0450b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3498078e25d1156a89ddc17f4823182b1aeb8b283f8b46bf66df67d67c5c5e44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3498078e25d1156a89ddc17f4823182b1aeb8b283f8b46bf66df67d67c5c5e44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:27Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:27 crc kubenswrapper[4860]: I1211 08:12:27.702965 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:27 crc kubenswrapper[4860]: I1211 08:12:27.703040 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:27 crc kubenswrapper[4860]: I1211 08:12:27.703068 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:27 crc kubenswrapper[4860]: I1211 08:12:27.703099 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:27 crc kubenswrapper[4860]: I1211 08:12:27.703124 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:27Z","lastTransitionTime":"2025-12-11T08:12:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:27 crc kubenswrapper[4860]: I1211 08:12:27.724470 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9aa0aed-62b3-4ea2-82ae-9a653c1101a5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ad4ba1b769c6cb2b34073798f8e3c25eb0177e6d3059484bf365e6981cab710\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://077aa1dd2fe5189ff60a0f2365e47a6a545cca0de7aa6379370908db4ea5838a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d7c919899b74a2b0fcd9221f3f769a9d013267354b200a282669d342d6413d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5483384f1d2ebc1fefa9706c9e67a3dc5a00ff643f803e3f2707c252e27ad94\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:27Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:27 crc kubenswrapper[4860]: I1211 08:12:27.739985 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31c30642-6e60-41b4-a477-0d802424e0aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79969402f5c33852a26fc839f37a45febf4f2377d77823284fbb01769e5585fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2d4d5de7106175087d91acb1f0c7552014f0c08bfcd65cbbf184d3f8d7826ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-99qgp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:27Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:27 crc kubenswrapper[4860]: I1211 08:12:27.755669 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-26nc6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1faa733-f408-41fd-b7aa-0cd86283976a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fce5b642b843ed30823c94e0b0c3d8bb1e5d2d93d691f7b8ef901ece03b0ac6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ac45db4f858a1c18a6b87472fc1cea2300f5845f1ad62c22260374995548a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ac45db4f858a1c18a6b87472fc1cea2300f5845f1ad62c22260374995548a0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b35296a5eb17dd2bb5e72c86c1cafdb30b7712cb330264f935d3b4fff05cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c6b35296a5eb17dd2bb5e72c86c1cafdb30b7712cb330264f935d3b4fff05cd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b9e82d6d3a247309ef5bdbf5b3a86ec083add2a138b0b237801e77b09ee96a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b9e82d6d3a247309ef5bdbf5b3a86ec083add2a138b0b237801e77b09ee96a1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://908ee695e78145ef4375c4097d9f1e772502922d02a2fcb4566172a6ad9caaf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://908ee695e78145ef4375c4097d9f1e772502922d02a2fcb4566172a6ad9caaf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afb0a097f3556a8f9ed7cf8f762cb99c082fe0a36b8d8d3038bf188b08fe9a84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afb0a097f3556a8f9ed7cf8f762cb99c082fe0a36b8d8d3038bf188b08fe9a84\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce18f4f06db091e658684fa0bbc858e19b9f03010acf13b3ffccb1a6e3f88021\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce18f4f06db091e658684fa0bbc858e19b9f03010acf13b3ffccb1a6e3f88021\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-26nc6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:27Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:27 crc kubenswrapper[4860]: I1211 08:12:27.777092 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0589a204-c98c-417d-8256-bfe3bf747660\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3333b4346f59b04017e32d61f457ba7ff457166731128f339d7987ae6b56b06d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52bb8ade314afb9494f4e026d0af2aea7b1970d64dc66102f55e257ae4678154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dfcc4370edaae36a5a32b0a9c29290c50f8e4f427e4e147dc3c779beaa8b780\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d856ae95e9077524c518d82d10fb6b93b7fbcaaf2b720ffafe6f7fae58bd7caa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1426084ac580895c42354630910dd8c8e3a48ddd25f22abce1d28dc3de332a63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9984d5d242bb258405efe91fc9caf9e7ea592834445df2e78add9553a921c76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://769a74d7b125e7e981e5b0541324a777d021651182d900a57685c9db458395f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://769a74d7b125e7e981e5b0541324a777d021651182d900a57685c9db458395f1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-11T08:11:59Z\\\",\\\"message\\\":\\\"multus-pznwl\\\\nI1211 08:11:59.565441 6486 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1211 08:11:59.565459 6486 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-pznwl in node crc\\\\nI1211 08:11:59.565418 6486 ovn.go:134] Ensuring zone local for Pod openshift-kube-controller-manager/kube-controller-manager-crc in node crc\\\\nI1211 08:11:59.565467 6486 obj_retry.go:386] Retry successful for *v1.Pod openshift-multus/multus-pznwl after 0 failed attempt(s)\\\\nI1211 08:11:59.565474 6486 obj_retry.go:386] Retry successful for *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc after 0 failed attempt(s)\\\\nI1211 08:11:59.565476 6486 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-pznwl\\\\nI1211 08:11:59.565481 6486 default_network_controller.go:776] Recording success event on pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1211 08:11:59.565201 6486 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/network-metrics-daemon-gjhlj\\\\nI1211 08:11:59.565512 6486 ovn.go:134] Ensuring zone local for Pod openshift-multus/network-metrics-daemon-gjhlj in node crc\\\\nF1211 08:11:59.565540 6486 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:58Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-xvnqp_openshift-ovn-kubernetes(0589a204-c98c-417d-8256-bfe3bf747660)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb80d9928e8349fc5963227d135355ad7ba820d4fd89877af601d005ca03334b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17cbcb49c06f3a9059cd5a0f05eedb8c785da30a3e502e31bae2f57b3486cb14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17cbcb49c06f3a9059cd5a0f05eedb8c785da30a3e502e31bae2f57b3486cb14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xvnqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:27Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:27 crc kubenswrapper[4860]: I1211 08:12:27.805342 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:27 crc kubenswrapper[4860]: I1211 08:12:27.805377 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:27 crc kubenswrapper[4860]: I1211 08:12:27.805387 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:27 crc kubenswrapper[4860]: I1211 08:12:27.805403 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:27 crc kubenswrapper[4860]: I1211 08:12:27.805415 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:27Z","lastTransitionTime":"2025-12-11T08:12:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:27 crc kubenswrapper[4860]: I1211 08:12:27.808336 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9872be1e-b995-4a00-904b-875dff96ad56\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f74a32a8fefea704cc73ffd1114083dab68aa44dc33f8d9d432d2d24216546c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbe00b46063d3968d59c393b87178e135f982abb7d0f32fa798b212d4b83df6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eedc9988e1a4a3c8cae730dc4a352103a6d665535b825e157d0f87c331b137c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2137fee4b7abc9765f9517d7b3b6ea796601a31d03151bf167d17136ced8b25a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f683e48f3577a68862eb6302383538376e41f49b20b1e0ea3f1b3b143f24872a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5de794187456b3576b969069a3b839d75c7918418a97fa35dff8bb40bb76e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5de794187456b3576b969069a3b839d75c7918418a97fa35dff8bb40bb76e39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08ce42b974fba00c409c0af971f0de351e50f18b2f06f8e722150da3239bb908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08ce42b974fba00c409c0af971f0de351e50f18b2f06f8e722150da3239bb908\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cfac6be03e3cfa0b0f3450b48b67957f2d5172b9c65e27f14c2a260bda5d423f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cfac6be03e3cfa0b0f3450b48b67957f2d5172b9c65e27f14c2a260bda5d423f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:27Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:27 crc kubenswrapper[4860]: I1211 08:12:27.830434 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-pznwl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5310c1fc-66c6-40aa-b1bf-5a59a2410f9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34119a6367261d7e12174f914ee8cfaa55703a55210f83c58b0e4694c54a5b03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2d6ff00ce322a28607bc3f6db05c5e4fb14acb6777ea315786ef53963ca504a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-11T08:12:17Z\\\",\\\"message\\\":\\\"2025-12-11T08:11:31+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_f37a703f-1609-4581-ab99-50337287ff33\\\\n2025-12-11T08:11:31+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_f37a703f-1609-4581-ab99-50337287ff33 to /host/opt/cni/bin/\\\\n2025-12-11T08:11:32Z [verbose] multus-daemon started\\\\n2025-12-11T08:11:32Z [verbose] Readiness Indicator file check\\\\n2025-12-11T08:12:17Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:12:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hkt48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-pznwl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:27Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:27 crc kubenswrapper[4860]: I1211 08:12:27.842126 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-854v6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e6529ba7-2341-43ac-a187-95ab530fe775\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8986538a421f0184f52f288b0e171da695591caf2b10c67ad2dce1073a6e6016\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kh22b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0d4e3f761ad15ae1424204e40b9d3f60188ab4a9047bd4b5ac6a8a8ca1fb385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kh22b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-854v6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:27Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:27 crc kubenswrapper[4860]: I1211 08:12:27.857710 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6xq7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a75ce3c-5e1e-4b9f-bbde-7a3b0cce5eae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe8d4d675567b41728e476351811a93c47b19868ccbadab8c613a4eb8eca8562\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5c6nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6xq7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:27Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:27 crc kubenswrapper[4860]: I1211 08:12:27.872406 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"910ec852-e5d2-4370-806d-77f16a36e630\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://122eb7e14a042027de3aafaa8db89562d80565e7df20f302f872e48d412a26b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1348b83ed7b1e8c25f9dfb1b2d2d035b5adc96dc23854ee90236f06fb155f155\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://867cdac386b0576ee2958a520afd522a2558a8878ea6b5af977b136dc7bba4a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4602fc6635e6188bf4cb75c5c7cf1e71a64fc0c5420626bfafd87beae150722d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4602fc6635e6188bf4cb75c5c7cf1e71a64fc0c5420626bfafd87beae150722d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:27Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:27 crc kubenswrapper[4860]: I1211 08:12:27.887303 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:27Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:27 crc kubenswrapper[4860]: I1211 08:12:27.903396 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d8c2f28628dc2f1ffbdd98a55670e22cf6cdc8cf2c6442e9d1f8e5466b2f12e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://074ea63beac52bd196507ee7cfb1c3d7fc2d37d5ec6faa4d47dde62b922e1c43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:27Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:27 crc kubenswrapper[4860]: I1211 08:12:27.907861 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:27 crc kubenswrapper[4860]: I1211 08:12:27.907912 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:27 crc kubenswrapper[4860]: I1211 08:12:27.907934 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:27 crc kubenswrapper[4860]: I1211 08:12:27.908001 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:27 crc kubenswrapper[4860]: I1211 08:12:27.908079 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:27Z","lastTransitionTime":"2025-12-11T08:12:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:27 crc kubenswrapper[4860]: I1211 08:12:27.919341 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:27Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:27 crc kubenswrapper[4860]: I1211 08:12:27.938844 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25a6a154da2b74952951ae94efc914ba078ea4bbe5893360b3c84d46be6bf973\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:27Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:28 crc kubenswrapper[4860]: I1211 08:12:28.011013 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:28 crc kubenswrapper[4860]: I1211 08:12:28.011055 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:28 crc kubenswrapper[4860]: I1211 08:12:28.011068 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:28 crc kubenswrapper[4860]: I1211 08:12:28.011085 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:28 crc kubenswrapper[4860]: I1211 08:12:28.011097 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:28Z","lastTransitionTime":"2025-12-11T08:12:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:28 crc kubenswrapper[4860]: I1211 08:12:28.113461 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:28 crc kubenswrapper[4860]: I1211 08:12:28.113495 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:28 crc kubenswrapper[4860]: I1211 08:12:28.113503 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:28 crc kubenswrapper[4860]: I1211 08:12:28.113517 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:28 crc kubenswrapper[4860]: I1211 08:12:28.113527 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:28Z","lastTransitionTime":"2025-12-11T08:12:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:28 crc kubenswrapper[4860]: I1211 08:12:28.216584 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:28 crc kubenswrapper[4860]: I1211 08:12:28.216688 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:28 crc kubenswrapper[4860]: I1211 08:12:28.216729 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:28 crc kubenswrapper[4860]: I1211 08:12:28.216763 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:28 crc kubenswrapper[4860]: I1211 08:12:28.216785 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:28Z","lastTransitionTime":"2025-12-11T08:12:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:28 crc kubenswrapper[4860]: I1211 08:12:28.319919 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:28 crc kubenswrapper[4860]: I1211 08:12:28.319986 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:28 crc kubenswrapper[4860]: I1211 08:12:28.320004 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:28 crc kubenswrapper[4860]: I1211 08:12:28.320028 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:28 crc kubenswrapper[4860]: I1211 08:12:28.320049 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:28Z","lastTransitionTime":"2025-12-11T08:12:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:28 crc kubenswrapper[4860]: I1211 08:12:28.422888 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:28 crc kubenswrapper[4860]: I1211 08:12:28.422958 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:28 crc kubenswrapper[4860]: I1211 08:12:28.422986 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:28 crc kubenswrapper[4860]: I1211 08:12:28.423204 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:28 crc kubenswrapper[4860]: I1211 08:12:28.423235 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:28Z","lastTransitionTime":"2025-12-11T08:12:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:28 crc kubenswrapper[4860]: I1211 08:12:28.525845 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:28 crc kubenswrapper[4860]: I1211 08:12:28.525908 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:28 crc kubenswrapper[4860]: I1211 08:12:28.525926 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:28 crc kubenswrapper[4860]: I1211 08:12:28.525949 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:28 crc kubenswrapper[4860]: I1211 08:12:28.525969 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:28Z","lastTransitionTime":"2025-12-11T08:12:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:28 crc kubenswrapper[4860]: I1211 08:12:28.578543 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:12:28 crc kubenswrapper[4860]: I1211 08:12:28.578571 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:12:28 crc kubenswrapper[4860]: E1211 08:12:28.578825 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:12:28 crc kubenswrapper[4860]: I1211 08:12:28.578963 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gjhlj" Dec 11 08:12:28 crc kubenswrapper[4860]: I1211 08:12:28.581705 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:12:28 crc kubenswrapper[4860]: E1211 08:12:28.581846 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gjhlj" podUID="5cfd640a-f3d2-4b28-87e1-5f49f6341e21" Dec 11 08:12:28 crc kubenswrapper[4860]: E1211 08:12:28.581891 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:12:28 crc kubenswrapper[4860]: E1211 08:12:28.581885 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:12:28 crc kubenswrapper[4860]: I1211 08:12:28.582344 4860 scope.go:117] "RemoveContainer" containerID="769a74d7b125e7e981e5b0541324a777d021651182d900a57685c9db458395f1" Dec 11 08:12:28 crc kubenswrapper[4860]: I1211 08:12:28.629024 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:28 crc kubenswrapper[4860]: I1211 08:12:28.629335 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:28 crc kubenswrapper[4860]: I1211 08:12:28.629354 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:28 crc kubenswrapper[4860]: I1211 08:12:28.629379 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:28 crc kubenswrapper[4860]: I1211 08:12:28.629397 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:28Z","lastTransitionTime":"2025-12-11T08:12:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:28 crc kubenswrapper[4860]: I1211 08:12:28.731526 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:28 crc kubenswrapper[4860]: I1211 08:12:28.731582 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:28 crc kubenswrapper[4860]: I1211 08:12:28.731596 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:28 crc kubenswrapper[4860]: I1211 08:12:28.731616 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:28 crc kubenswrapper[4860]: I1211 08:12:28.731628 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:28Z","lastTransitionTime":"2025-12-11T08:12:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:28 crc kubenswrapper[4860]: I1211 08:12:28.835125 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:28 crc kubenswrapper[4860]: I1211 08:12:28.835221 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:28 crc kubenswrapper[4860]: I1211 08:12:28.835240 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:28 crc kubenswrapper[4860]: I1211 08:12:28.835265 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:28 crc kubenswrapper[4860]: I1211 08:12:28.835283 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:28Z","lastTransitionTime":"2025-12-11T08:12:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:28 crc kubenswrapper[4860]: I1211 08:12:28.938377 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:28 crc kubenswrapper[4860]: I1211 08:12:28.938424 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:28 crc kubenswrapper[4860]: I1211 08:12:28.938436 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:28 crc kubenswrapper[4860]: I1211 08:12:28.938462 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:28 crc kubenswrapper[4860]: I1211 08:12:28.938475 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:28Z","lastTransitionTime":"2025-12-11T08:12:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:29 crc kubenswrapper[4860]: I1211 08:12:29.041518 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:29 crc kubenswrapper[4860]: I1211 08:12:29.041585 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:29 crc kubenswrapper[4860]: I1211 08:12:29.041611 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:29 crc kubenswrapper[4860]: I1211 08:12:29.041670 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:29 crc kubenswrapper[4860]: I1211 08:12:29.041694 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:29Z","lastTransitionTime":"2025-12-11T08:12:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:29 crc kubenswrapper[4860]: I1211 08:12:29.143672 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:29 crc kubenswrapper[4860]: I1211 08:12:29.143701 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:29 crc kubenswrapper[4860]: I1211 08:12:29.143710 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:29 crc kubenswrapper[4860]: I1211 08:12:29.143722 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:29 crc kubenswrapper[4860]: I1211 08:12:29.143731 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:29Z","lastTransitionTime":"2025-12-11T08:12:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:29 crc kubenswrapper[4860]: I1211 08:12:29.246760 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:29 crc kubenswrapper[4860]: I1211 08:12:29.246807 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:29 crc kubenswrapper[4860]: I1211 08:12:29.246952 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:29 crc kubenswrapper[4860]: I1211 08:12:29.246979 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:29 crc kubenswrapper[4860]: I1211 08:12:29.246997 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:29Z","lastTransitionTime":"2025-12-11T08:12:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:29 crc kubenswrapper[4860]: I1211 08:12:29.350576 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:29 crc kubenswrapper[4860]: I1211 08:12:29.350618 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:29 crc kubenswrapper[4860]: I1211 08:12:29.350629 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:29 crc kubenswrapper[4860]: I1211 08:12:29.350669 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:29 crc kubenswrapper[4860]: I1211 08:12:29.350681 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:29Z","lastTransitionTime":"2025-12-11T08:12:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:29 crc kubenswrapper[4860]: I1211 08:12:29.453011 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:29 crc kubenswrapper[4860]: I1211 08:12:29.453047 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:29 crc kubenswrapper[4860]: I1211 08:12:29.453057 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:29 crc kubenswrapper[4860]: I1211 08:12:29.453073 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:29 crc kubenswrapper[4860]: I1211 08:12:29.453083 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:29Z","lastTransitionTime":"2025-12-11T08:12:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:29 crc kubenswrapper[4860]: I1211 08:12:29.555663 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:29 crc kubenswrapper[4860]: I1211 08:12:29.555740 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:29 crc kubenswrapper[4860]: I1211 08:12:29.555754 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:29 crc kubenswrapper[4860]: I1211 08:12:29.555781 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:29 crc kubenswrapper[4860]: I1211 08:12:29.555795 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:29Z","lastTransitionTime":"2025-12-11T08:12:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:29 crc kubenswrapper[4860]: I1211 08:12:29.657873 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:29 crc kubenswrapper[4860]: I1211 08:12:29.657914 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:29 crc kubenswrapper[4860]: I1211 08:12:29.657927 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:29 crc kubenswrapper[4860]: I1211 08:12:29.657944 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:29 crc kubenswrapper[4860]: I1211 08:12:29.657956 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:29Z","lastTransitionTime":"2025-12-11T08:12:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:29 crc kubenswrapper[4860]: I1211 08:12:29.760791 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:29 crc kubenswrapper[4860]: I1211 08:12:29.760845 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:29 crc kubenswrapper[4860]: I1211 08:12:29.760867 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:29 crc kubenswrapper[4860]: I1211 08:12:29.760892 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:29 crc kubenswrapper[4860]: I1211 08:12:29.760910 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:29Z","lastTransitionTime":"2025-12-11T08:12:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:29 crc kubenswrapper[4860]: I1211 08:12:29.862354 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:29 crc kubenswrapper[4860]: I1211 08:12:29.862389 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:29 crc kubenswrapper[4860]: I1211 08:12:29.862401 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:29 crc kubenswrapper[4860]: I1211 08:12:29.862414 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:29 crc kubenswrapper[4860]: I1211 08:12:29.862424 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:29Z","lastTransitionTime":"2025-12-11T08:12:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:29 crc kubenswrapper[4860]: I1211 08:12:29.965009 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:29 crc kubenswrapper[4860]: I1211 08:12:29.965064 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:29 crc kubenswrapper[4860]: I1211 08:12:29.965075 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:29 crc kubenswrapper[4860]: I1211 08:12:29.965092 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:29 crc kubenswrapper[4860]: I1211 08:12:29.965104 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:29Z","lastTransitionTime":"2025-12-11T08:12:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:30 crc kubenswrapper[4860]: I1211 08:12:30.067802 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:30 crc kubenswrapper[4860]: I1211 08:12:30.067851 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:30 crc kubenswrapper[4860]: I1211 08:12:30.067868 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:30 crc kubenswrapper[4860]: I1211 08:12:30.067889 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:30 crc kubenswrapper[4860]: I1211 08:12:30.067904 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:30Z","lastTransitionTime":"2025-12-11T08:12:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:30 crc kubenswrapper[4860]: I1211 08:12:30.107327 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-xvnqp_0589a204-c98c-417d-8256-bfe3bf747660/ovnkube-controller/2.log" Dec 11 08:12:30 crc kubenswrapper[4860]: I1211 08:12:30.110474 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" event={"ID":"0589a204-c98c-417d-8256-bfe3bf747660","Type":"ContainerStarted","Data":"d9e83cda3e40d325604114a37fe6e8b707ac1f91a034b0d979c70b5f4f24f2e7"} Dec 11 08:12:30 crc kubenswrapper[4860]: I1211 08:12:30.110894 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" Dec 11 08:12:30 crc kubenswrapper[4860]: I1211 08:12:30.122276 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8bf0e9df-4e9d-4f45-a5b9-0020fdab2ea6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77c0c831f046149002cd12564aa9b1544a813da1fcd81e61451eeb504f0450b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3498078e25d1156a89ddc17f4823182b1aeb8b283f8b46bf66df67d67c5c5e44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3498078e25d1156a89ddc17f4823182b1aeb8b283f8b46bf66df67d67c5c5e44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:30Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:30 crc kubenswrapper[4860]: I1211 08:12:30.137726 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9aa0aed-62b3-4ea2-82ae-9a653c1101a5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ad4ba1b769c6cb2b34073798f8e3c25eb0177e6d3059484bf365e6981cab710\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://077aa1dd2fe5189ff60a0f2365e47a6a545cca0de7aa6379370908db4ea5838a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d7c919899b74a2b0fcd9221f3f769a9d013267354b200a282669d342d6413d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5483384f1d2ebc1fefa9706c9e67a3dc5a00ff643f803e3f2707c252e27ad94\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:30Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:30 crc kubenswrapper[4860]: I1211 08:12:30.152782 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31c30642-6e60-41b4-a477-0d802424e0aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79969402f5c33852a26fc839f37a45febf4f2377d77823284fbb01769e5585fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2d4d5de7106175087d91acb1f0c7552014f0c08bfcd65cbbf184d3f8d7826ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-99qgp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:30Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:30 crc kubenswrapper[4860]: I1211 08:12:30.168464 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-26nc6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1faa733-f408-41fd-b7aa-0cd86283976a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fce5b642b843ed30823c94e0b0c3d8bb1e5d2d93d691f7b8ef901ece03b0ac6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ac45db4f858a1c18a6b87472fc1cea2300f5845f1ad62c22260374995548a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ac45db4f858a1c18a6b87472fc1cea2300f5845f1ad62c22260374995548a0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b35296a5eb17dd2bb5e72c86c1cafdb30b7712cb330264f935d3b4fff05cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c6b35296a5eb17dd2bb5e72c86c1cafdb30b7712cb330264f935d3b4fff05cd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b9e82d6d3a247309ef5bdbf5b3a86ec083add2a138b0b237801e77b09ee96a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b9e82d6d3a247309ef5bdbf5b3a86ec083add2a138b0b237801e77b09ee96a1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://908ee695e78145ef4375c4097d9f1e772502922d02a2fcb4566172a6ad9caaf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://908ee695e78145ef4375c4097d9f1e772502922d02a2fcb4566172a6ad9caaf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afb0a097f3556a8f9ed7cf8f762cb99c082fe0a36b8d8d3038bf188b08fe9a84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afb0a097f3556a8f9ed7cf8f762cb99c082fe0a36b8d8d3038bf188b08fe9a84\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce18f4f06db091e658684fa0bbc858e19b9f03010acf13b3ffccb1a6e3f88021\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce18f4f06db091e658684fa0bbc858e19b9f03010acf13b3ffccb1a6e3f88021\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-26nc6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:30Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:30 crc kubenswrapper[4860]: I1211 08:12:30.170008 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:30 crc kubenswrapper[4860]: I1211 08:12:30.170043 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:30 crc kubenswrapper[4860]: I1211 08:12:30.170069 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:30 crc kubenswrapper[4860]: I1211 08:12:30.170084 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:30 crc kubenswrapper[4860]: I1211 08:12:30.170094 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:30Z","lastTransitionTime":"2025-12-11T08:12:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:30 crc kubenswrapper[4860]: I1211 08:12:30.187468 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0589a204-c98c-417d-8256-bfe3bf747660\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3333b4346f59b04017e32d61f457ba7ff457166731128f339d7987ae6b56b06d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52bb8ade314afb9494f4e026d0af2aea7b1970d64dc66102f55e257ae4678154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dfcc4370edaae36a5a32b0a9c29290c50f8e4f427e4e147dc3c779beaa8b780\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d856ae95e9077524c518d82d10fb6b93b7fbcaaf2b720ffafe6f7fae58bd7caa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1426084ac580895c42354630910dd8c8e3a48ddd25f22abce1d28dc3de332a63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9984d5d242bb258405efe91fc9caf9e7ea592834445df2e78add9553a921c76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9e83cda3e40d325604114a37fe6e8b707ac1f91a034b0d979c70b5f4f24f2e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://769a74d7b125e7e981e5b0541324a777d021651182d900a57685c9db458395f1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-11T08:11:59Z\\\",\\\"message\\\":\\\"multus-pznwl\\\\nI1211 08:11:59.565441 6486 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1211 08:11:59.565459 6486 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-pznwl in node crc\\\\nI1211 08:11:59.565418 6486 ovn.go:134] Ensuring zone local for Pod openshift-kube-controller-manager/kube-controller-manager-crc in node crc\\\\nI1211 08:11:59.565467 6486 obj_retry.go:386] Retry successful for *v1.Pod openshift-multus/multus-pznwl after 0 failed attempt(s)\\\\nI1211 08:11:59.565474 6486 obj_retry.go:386] Retry successful for *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc after 0 failed attempt(s)\\\\nI1211 08:11:59.565476 6486 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-pznwl\\\\nI1211 08:11:59.565481 6486 default_network_controller.go:776] Recording success event on pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1211 08:11:59.565201 6486 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/network-metrics-daemon-gjhlj\\\\nI1211 08:11:59.565512 6486 ovn.go:134] Ensuring zone local for Pod openshift-multus/network-metrics-daemon-gjhlj in node crc\\\\nF1211 08:11:59.565540 6486 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:58Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:12:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb80d9928e8349fc5963227d135355ad7ba820d4fd89877af601d005ca03334b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17cbcb49c06f3a9059cd5a0f05eedb8c785da30a3e502e31bae2f57b3486cb14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17cbcb49c06f3a9059cd5a0f05eedb8c785da30a3e502e31bae2f57b3486cb14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xvnqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:30Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:30 crc kubenswrapper[4860]: I1211 08:12:30.198630 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gsx9l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aba5b6b9-f88d-4763-9562-6b4dec606df9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://735ef012c523ddc952d46a431f27b2582a2c301d9e9da0c697a20ff6e2fb5d86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6zw9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gsx9l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:30Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:30 crc kubenswrapper[4860]: I1211 08:12:30.220150 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9872be1e-b995-4a00-904b-875dff96ad56\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f74a32a8fefea704cc73ffd1114083dab68aa44dc33f8d9d432d2d24216546c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbe00b46063d3968d59c393b87178e135f982abb7d0f32fa798b212d4b83df6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eedc9988e1a4a3c8cae730dc4a352103a6d665535b825e157d0f87c331b137c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2137fee4b7abc9765f9517d7b3b6ea796601a31d03151bf167d17136ced8b25a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f683e48f3577a68862eb6302383538376e41f49b20b1e0ea3f1b3b143f24872a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5de794187456b3576b969069a3b839d75c7918418a97fa35dff8bb40bb76e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5de794187456b3576b969069a3b839d75c7918418a97fa35dff8bb40bb76e39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08ce42b974fba00c409c0af971f0de351e50f18b2f06f8e722150da3239bb908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08ce42b974fba00c409c0af971f0de351e50f18b2f06f8e722150da3239bb908\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cfac6be03e3cfa0b0f3450b48b67957f2d5172b9c65e27f14c2a260bda5d423f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cfac6be03e3cfa0b0f3450b48b67957f2d5172b9c65e27f14c2a260bda5d423f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:30Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:30 crc kubenswrapper[4860]: I1211 08:12:30.233926 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-pznwl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5310c1fc-66c6-40aa-b1bf-5a59a2410f9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34119a6367261d7e12174f914ee8cfaa55703a55210f83c58b0e4694c54a5b03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2d6ff00ce322a28607bc3f6db05c5e4fb14acb6777ea315786ef53963ca504a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-11T08:12:17Z\\\",\\\"message\\\":\\\"2025-12-11T08:11:31+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_f37a703f-1609-4581-ab99-50337287ff33\\\\n2025-12-11T08:11:31+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_f37a703f-1609-4581-ab99-50337287ff33 to /host/opt/cni/bin/\\\\n2025-12-11T08:11:32Z [verbose] multus-daemon started\\\\n2025-12-11T08:11:32Z [verbose] Readiness Indicator file check\\\\n2025-12-11T08:12:17Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:12:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hkt48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-pznwl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:30Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:30 crc kubenswrapper[4860]: I1211 08:12:30.246816 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-854v6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e6529ba7-2341-43ac-a187-95ab530fe775\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8986538a421f0184f52f288b0e171da695591caf2b10c67ad2dce1073a6e6016\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kh22b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0d4e3f761ad15ae1424204e40b9d3f60188ab4a9047bd4b5ac6a8a8ca1fb385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kh22b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-854v6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:30Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:30 crc kubenswrapper[4860]: I1211 08:12:30.263655 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"910ec852-e5d2-4370-806d-77f16a36e630\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://122eb7e14a042027de3aafaa8db89562d80565e7df20f302f872e48d412a26b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1348b83ed7b1e8c25f9dfb1b2d2d035b5adc96dc23854ee90236f06fb155f155\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://867cdac386b0576ee2958a520afd522a2558a8878ea6b5af977b136dc7bba4a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4602fc6635e6188bf4cb75c5c7cf1e71a64fc0c5420626bfafd87beae150722d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4602fc6635e6188bf4cb75c5c7cf1e71a64fc0c5420626bfafd87beae150722d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:30Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:30 crc kubenswrapper[4860]: I1211 08:12:30.272380 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:30 crc kubenswrapper[4860]: I1211 08:12:30.272436 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:30 crc kubenswrapper[4860]: I1211 08:12:30.272450 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:30 crc kubenswrapper[4860]: I1211 08:12:30.272469 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:30 crc kubenswrapper[4860]: I1211 08:12:30.272485 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:30Z","lastTransitionTime":"2025-12-11T08:12:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:30 crc kubenswrapper[4860]: I1211 08:12:30.280341 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:30Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:30 crc kubenswrapper[4860]: I1211 08:12:30.298831 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d8c2f28628dc2f1ffbdd98a55670e22cf6cdc8cf2c6442e9d1f8e5466b2f12e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://074ea63beac52bd196507ee7cfb1c3d7fc2d37d5ec6faa4d47dde62b922e1c43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:30Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:30 crc kubenswrapper[4860]: I1211 08:12:30.316033 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:30Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:30 crc kubenswrapper[4860]: I1211 08:12:30.329894 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25a6a154da2b74952951ae94efc914ba078ea4bbe5893360b3c84d46be6bf973\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:30Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:30 crc kubenswrapper[4860]: I1211 08:12:30.342563 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6xq7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a75ce3c-5e1e-4b9f-bbde-7a3b0cce5eae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe8d4d675567b41728e476351811a93c47b19868ccbadab8c613a4eb8eca8562\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5c6nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6xq7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:30Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:30 crc kubenswrapper[4860]: I1211 08:12:30.360273 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ec3758-85f1-4a22-a4ec-c805472607a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e831db1a15976d3967d56479689572c26f2599a766baa0402897a3008428c56f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a498aa90f914e1f0370df01dc5ef174df113a2a16d86b2e3f24b19a1b99b4c0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8caebd2f4dfcbb89c597029a07c335171889761d2d200206987537fb0a228db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34e91f615b76445ff3241a304bb2964fde59de7e8a8dfff6841b2e628a31e86d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2090cab02ad74ddfd6bb57ff50eba35d18f56b0cde1c230fe011c01994b693bc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 08:11:21.171011 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 08:11:21.177674 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-644352688/tls.crt::/tmp/serving-cert-644352688/tls.key\\\\\\\"\\\\nI1211 08:11:26.508606 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 08:11:26.512741 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 08:11:26.512778 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 08:11:26.512839 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 08:11:26.512858 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 08:11:26.521684 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1211 08:11:26.521700 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1211 08:11:26.521722 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:11:26.521732 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:11:26.521738 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 08:11:26.521746 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 08:11:26.521749 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 08:11:26.521753 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1211 08:11:26.524999 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2274fb298ca7d2b37b060b4dbf9b166d39ba77a108b08edeeb4d67196e72cb3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf0aff8a6ab816b6617b1feeca9146d41bb5e8808bf1db96f8d493500c135d85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf0aff8a6ab816b6617b1feeca9146d41bb5e8808bf1db96f8d493500c135d85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:30Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:30 crc kubenswrapper[4860]: I1211 08:12:30.374969 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:30 crc kubenswrapper[4860]: I1211 08:12:30.375024 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:30 crc kubenswrapper[4860]: I1211 08:12:30.375041 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:30 crc kubenswrapper[4860]: I1211 08:12:30.375063 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:30 crc kubenswrapper[4860]: I1211 08:12:30.375085 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:30Z","lastTransitionTime":"2025-12-11T08:12:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:30 crc kubenswrapper[4860]: I1211 08:12:30.379723 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:30Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:30 crc kubenswrapper[4860]: I1211 08:12:30.395622 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a41f35db087418d1be16953f10377042abd3f4488e5ac1bc29ca5dcfc00a37e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:30Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:30 crc kubenswrapper[4860]: I1211 08:12:30.411121 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gjhlj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5cfd640a-f3d2-4b28-87e1-5f49f6341e21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knq7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knq7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:42Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gjhlj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:30Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:30 crc kubenswrapper[4860]: I1211 08:12:30.477495 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:30 crc kubenswrapper[4860]: I1211 08:12:30.477791 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:30 crc kubenswrapper[4860]: I1211 08:12:30.477821 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:30 crc kubenswrapper[4860]: I1211 08:12:30.477852 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:30 crc kubenswrapper[4860]: I1211 08:12:30.477871 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:30Z","lastTransitionTime":"2025-12-11T08:12:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:30 crc kubenswrapper[4860]: I1211 08:12:30.520816 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:12:30 crc kubenswrapper[4860]: E1211 08:12:30.521012 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:13:34.520975462 +0000 UTC m=+147.249494557 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:12:30 crc kubenswrapper[4860]: I1211 08:12:30.521203 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:12:30 crc kubenswrapper[4860]: E1211 08:12:30.521415 4860 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 11 08:12:30 crc kubenswrapper[4860]: E1211 08:12:30.521494 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-11 08:13:34.5214794 +0000 UTC m=+147.249998485 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 11 08:12:30 crc kubenswrapper[4860]: I1211 08:12:30.578139 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:12:30 crc kubenswrapper[4860]: I1211 08:12:30.578201 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:12:30 crc kubenswrapper[4860]: I1211 08:12:30.578261 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:12:30 crc kubenswrapper[4860]: E1211 08:12:30.578308 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:12:30 crc kubenswrapper[4860]: I1211 08:12:30.578371 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gjhlj" Dec 11 08:12:30 crc kubenswrapper[4860]: E1211 08:12:30.578505 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:12:30 crc kubenswrapper[4860]: E1211 08:12:30.578727 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gjhlj" podUID="5cfd640a-f3d2-4b28-87e1-5f49f6341e21" Dec 11 08:12:30 crc kubenswrapper[4860]: E1211 08:12:30.578914 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:12:30 crc kubenswrapper[4860]: I1211 08:12:30.580761 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:30 crc kubenswrapper[4860]: I1211 08:12:30.580817 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:30 crc kubenswrapper[4860]: I1211 08:12:30.580839 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:30 crc kubenswrapper[4860]: I1211 08:12:30.580867 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:30 crc kubenswrapper[4860]: I1211 08:12:30.580888 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:30Z","lastTransitionTime":"2025-12-11T08:12:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:30 crc kubenswrapper[4860]: I1211 08:12:30.621726 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:12:30 crc kubenswrapper[4860]: I1211 08:12:30.621901 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:12:30 crc kubenswrapper[4860]: E1211 08:12:30.621945 4860 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 11 08:12:30 crc kubenswrapper[4860]: I1211 08:12:30.621968 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:12:30 crc kubenswrapper[4860]: E1211 08:12:30.621987 4860 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 11 08:12:30 crc kubenswrapper[4860]: E1211 08:12:30.622013 4860 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 08:12:30 crc kubenswrapper[4860]: E1211 08:12:30.622101 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-11 08:13:34.622071537 +0000 UTC m=+147.350590622 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 08:12:30 crc kubenswrapper[4860]: E1211 08:12:30.622100 4860 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 11 08:12:30 crc kubenswrapper[4860]: E1211 08:12:30.622180 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-11 08:13:34.6221578 +0000 UTC m=+147.350676895 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 11 08:12:30 crc kubenswrapper[4860]: E1211 08:12:30.622253 4860 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 11 08:12:30 crc kubenswrapper[4860]: E1211 08:12:30.622310 4860 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 11 08:12:30 crc kubenswrapper[4860]: E1211 08:12:30.622331 4860 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 08:12:30 crc kubenswrapper[4860]: E1211 08:12:30.622445 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-11 08:13:34.622413038 +0000 UTC m=+147.350932143 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 08:12:30 crc kubenswrapper[4860]: I1211 08:12:30.684361 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:30 crc kubenswrapper[4860]: I1211 08:12:30.684428 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:30 crc kubenswrapper[4860]: I1211 08:12:30.684448 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:30 crc kubenswrapper[4860]: I1211 08:12:30.684473 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:30 crc kubenswrapper[4860]: I1211 08:12:30.684491 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:30Z","lastTransitionTime":"2025-12-11T08:12:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:30 crc kubenswrapper[4860]: I1211 08:12:30.787678 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:30 crc kubenswrapper[4860]: I1211 08:12:30.787717 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:30 crc kubenswrapper[4860]: I1211 08:12:30.787728 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:30 crc kubenswrapper[4860]: I1211 08:12:30.788086 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:30 crc kubenswrapper[4860]: I1211 08:12:30.788110 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:30Z","lastTransitionTime":"2025-12-11T08:12:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:30 crc kubenswrapper[4860]: I1211 08:12:30.891103 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:30 crc kubenswrapper[4860]: I1211 08:12:30.891154 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:30 crc kubenswrapper[4860]: I1211 08:12:30.891171 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:30 crc kubenswrapper[4860]: I1211 08:12:30.891195 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:30 crc kubenswrapper[4860]: I1211 08:12:30.891211 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:30Z","lastTransitionTime":"2025-12-11T08:12:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:30 crc kubenswrapper[4860]: I1211 08:12:30.994013 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:30 crc kubenswrapper[4860]: I1211 08:12:30.994068 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:30 crc kubenswrapper[4860]: I1211 08:12:30.994087 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:30 crc kubenswrapper[4860]: I1211 08:12:30.994110 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:30 crc kubenswrapper[4860]: I1211 08:12:30.994127 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:30Z","lastTransitionTime":"2025-12-11T08:12:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:31 crc kubenswrapper[4860]: I1211 08:12:31.096959 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:31 crc kubenswrapper[4860]: I1211 08:12:31.097023 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:31 crc kubenswrapper[4860]: I1211 08:12:31.097042 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:31 crc kubenswrapper[4860]: I1211 08:12:31.097068 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:31 crc kubenswrapper[4860]: I1211 08:12:31.097087 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:31Z","lastTransitionTime":"2025-12-11T08:12:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:31 crc kubenswrapper[4860]: I1211 08:12:31.116388 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-xvnqp_0589a204-c98c-417d-8256-bfe3bf747660/ovnkube-controller/3.log" Dec 11 08:12:31 crc kubenswrapper[4860]: I1211 08:12:31.117375 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-xvnqp_0589a204-c98c-417d-8256-bfe3bf747660/ovnkube-controller/2.log" Dec 11 08:12:31 crc kubenswrapper[4860]: I1211 08:12:31.120372 4860 generic.go:334] "Generic (PLEG): container finished" podID="0589a204-c98c-417d-8256-bfe3bf747660" containerID="d9e83cda3e40d325604114a37fe6e8b707ac1f91a034b0d979c70b5f4f24f2e7" exitCode=1 Dec 11 08:12:31 crc kubenswrapper[4860]: I1211 08:12:31.120444 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" event={"ID":"0589a204-c98c-417d-8256-bfe3bf747660","Type":"ContainerDied","Data":"d9e83cda3e40d325604114a37fe6e8b707ac1f91a034b0d979c70b5f4f24f2e7"} Dec 11 08:12:31 crc kubenswrapper[4860]: I1211 08:12:31.120503 4860 scope.go:117] "RemoveContainer" containerID="769a74d7b125e7e981e5b0541324a777d021651182d900a57685c9db458395f1" Dec 11 08:12:31 crc kubenswrapper[4860]: I1211 08:12:31.122498 4860 scope.go:117] "RemoveContainer" containerID="d9e83cda3e40d325604114a37fe6e8b707ac1f91a034b0d979c70b5f4f24f2e7" Dec 11 08:12:31 crc kubenswrapper[4860]: E1211 08:12:31.122798 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-xvnqp_openshift-ovn-kubernetes(0589a204-c98c-417d-8256-bfe3bf747660)\"" pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" podUID="0589a204-c98c-417d-8256-bfe3bf747660" Dec 11 08:12:31 crc kubenswrapper[4860]: I1211 08:12:31.140232 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9aa0aed-62b3-4ea2-82ae-9a653c1101a5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ad4ba1b769c6cb2b34073798f8e3c25eb0177e6d3059484bf365e6981cab710\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://077aa1dd2fe5189ff60a0f2365e47a6a545cca0de7aa6379370908db4ea5838a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d7c919899b74a2b0fcd9221f3f769a9d013267354b200a282669d342d6413d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5483384f1d2ebc1fefa9706c9e67a3dc5a00ff643f803e3f2707c252e27ad94\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:31Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:31 crc kubenswrapper[4860]: I1211 08:12:31.154367 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31c30642-6e60-41b4-a477-0d802424e0aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79969402f5c33852a26fc839f37a45febf4f2377d77823284fbb01769e5585fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2d4d5de7106175087d91acb1f0c7552014f0c08bfcd65cbbf184d3f8d7826ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-99qgp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:31Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:31 crc kubenswrapper[4860]: I1211 08:12:31.171050 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-26nc6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1faa733-f408-41fd-b7aa-0cd86283976a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fce5b642b843ed30823c94e0b0c3d8bb1e5d2d93d691f7b8ef901ece03b0ac6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ac45db4f858a1c18a6b87472fc1cea2300f5845f1ad62c22260374995548a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ac45db4f858a1c18a6b87472fc1cea2300f5845f1ad62c22260374995548a0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b35296a5eb17dd2bb5e72c86c1cafdb30b7712cb330264f935d3b4fff05cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c6b35296a5eb17dd2bb5e72c86c1cafdb30b7712cb330264f935d3b4fff05cd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b9e82d6d3a247309ef5bdbf5b3a86ec083add2a138b0b237801e77b09ee96a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b9e82d6d3a247309ef5bdbf5b3a86ec083add2a138b0b237801e77b09ee96a1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://908ee695e78145ef4375c4097d9f1e772502922d02a2fcb4566172a6ad9caaf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://908ee695e78145ef4375c4097d9f1e772502922d02a2fcb4566172a6ad9caaf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afb0a097f3556a8f9ed7cf8f762cb99c082fe0a36b8d8d3038bf188b08fe9a84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afb0a097f3556a8f9ed7cf8f762cb99c082fe0a36b8d8d3038bf188b08fe9a84\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce18f4f06db091e658684fa0bbc858e19b9f03010acf13b3ffccb1a6e3f88021\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce18f4f06db091e658684fa0bbc858e19b9f03010acf13b3ffccb1a6e3f88021\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-26nc6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:31Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:31 crc kubenswrapper[4860]: I1211 08:12:31.193224 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0589a204-c98c-417d-8256-bfe3bf747660\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3333b4346f59b04017e32d61f457ba7ff457166731128f339d7987ae6b56b06d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52bb8ade314afb9494f4e026d0af2aea7b1970d64dc66102f55e257ae4678154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dfcc4370edaae36a5a32b0a9c29290c50f8e4f427e4e147dc3c779beaa8b780\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d856ae95e9077524c518d82d10fb6b93b7fbcaaf2b720ffafe6f7fae58bd7caa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1426084ac580895c42354630910dd8c8e3a48ddd25f22abce1d28dc3de332a63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9984d5d242bb258405efe91fc9caf9e7ea592834445df2e78add9553a921c76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9e83cda3e40d325604114a37fe6e8b707ac1f91a034b0d979c70b5f4f24f2e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://769a74d7b125e7e981e5b0541324a777d021651182d900a57685c9db458395f1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-11T08:11:59Z\\\",\\\"message\\\":\\\"multus-pznwl\\\\nI1211 08:11:59.565441 6486 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1211 08:11:59.565459 6486 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-pznwl in node crc\\\\nI1211 08:11:59.565418 6486 ovn.go:134] Ensuring zone local for Pod openshift-kube-controller-manager/kube-controller-manager-crc in node crc\\\\nI1211 08:11:59.565467 6486 obj_retry.go:386] Retry successful for *v1.Pod openshift-multus/multus-pznwl after 0 failed attempt(s)\\\\nI1211 08:11:59.565474 6486 obj_retry.go:386] Retry successful for *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc after 0 failed attempt(s)\\\\nI1211 08:11:59.565476 6486 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-pznwl\\\\nI1211 08:11:59.565481 6486 default_network_controller.go:776] Recording success event on pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1211 08:11:59.565201 6486 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/network-metrics-daemon-gjhlj\\\\nI1211 08:11:59.565512 6486 ovn.go:134] Ensuring zone local for Pod openshift-multus/network-metrics-daemon-gjhlj in node crc\\\\nF1211 08:11:59.565540 6486 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:58Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9e83cda3e40d325604114a37fe6e8b707ac1f91a034b0d979c70b5f4f24f2e7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-11T08:12:30Z\\\",\\\"message\\\":\\\"val\\\\nI1211 08:12:30.086873 6903 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1211 08:12:30.086964 6903 handler.go:208] Removed *v1.Node event handler 7\\\\nI1211 08:12:30.086994 6903 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1211 08:12:30.087035 6903 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1211 08:12:30.087066 6903 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1211 08:12:30.087141 6903 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1211 08:12:30.087543 6903 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1211 08:12:30.087550 6903 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1211 08:12:30.087568 6903 handler.go:208] Removed *v1.Node event handler 2\\\\nI1211 08:12:30.087967 6903 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1211 08:12:30.087990 6903 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1211 08:12:30.088004 6903 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1211 08:12:30.088018 6903 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1211 08:12:30.088032 6903 factory.go:656] Stopping watch factory\\\\nI1211 08:12:30.088038 6903 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1211 08:12:30.088050 6903 ovnkube.go:599] Stopped ovnkube\\\\nI1211 08:12:3\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:12:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb80d9928e8349fc5963227d135355ad7ba820d4fd89877af601d005ca03334b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17cbcb49c06f3a9059cd5a0f05eedb8c785da30a3e502e31bae2f57b3486cb14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17cbcb49c06f3a9059cd5a0f05eedb8c785da30a3e502e31bae2f57b3486cb14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xvnqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:31Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:31 crc kubenswrapper[4860]: I1211 08:12:31.200191 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:31 crc kubenswrapper[4860]: I1211 08:12:31.200243 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:31 crc kubenswrapper[4860]: I1211 08:12:31.200258 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:31 crc kubenswrapper[4860]: I1211 08:12:31.200296 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:31 crc kubenswrapper[4860]: I1211 08:12:31.200314 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:31Z","lastTransitionTime":"2025-12-11T08:12:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:31 crc kubenswrapper[4860]: I1211 08:12:31.203664 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gsx9l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aba5b6b9-f88d-4763-9562-6b4dec606df9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://735ef012c523ddc952d46a431f27b2582a2c301d9e9da0c697a20ff6e2fb5d86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6zw9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gsx9l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:31Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:31 crc kubenswrapper[4860]: I1211 08:12:31.216399 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8bf0e9df-4e9d-4f45-a5b9-0020fdab2ea6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77c0c831f046149002cd12564aa9b1544a813da1fcd81e61451eeb504f0450b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3498078e25d1156a89ddc17f4823182b1aeb8b283f8b46bf66df67d67c5c5e44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3498078e25d1156a89ddc17f4823182b1aeb8b283f8b46bf66df67d67c5c5e44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:31Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:31 crc kubenswrapper[4860]: I1211 08:12:31.229604 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-pznwl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5310c1fc-66c6-40aa-b1bf-5a59a2410f9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34119a6367261d7e12174f914ee8cfaa55703a55210f83c58b0e4694c54a5b03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2d6ff00ce322a28607bc3f6db05c5e4fb14acb6777ea315786ef53963ca504a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-11T08:12:17Z\\\",\\\"message\\\":\\\"2025-12-11T08:11:31+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_f37a703f-1609-4581-ab99-50337287ff33\\\\n2025-12-11T08:11:31+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_f37a703f-1609-4581-ab99-50337287ff33 to /host/opt/cni/bin/\\\\n2025-12-11T08:11:32Z [verbose] multus-daemon started\\\\n2025-12-11T08:11:32Z [verbose] Readiness Indicator file check\\\\n2025-12-11T08:12:17Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:12:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hkt48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-pznwl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:31Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:31 crc kubenswrapper[4860]: I1211 08:12:31.242423 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-854v6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e6529ba7-2341-43ac-a187-95ab530fe775\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8986538a421f0184f52f288b0e171da695591caf2b10c67ad2dce1073a6e6016\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kh22b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0d4e3f761ad15ae1424204e40b9d3f60188ab4a9047bd4b5ac6a8a8ca1fb385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kh22b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-854v6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:31Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:31 crc kubenswrapper[4860]: I1211 08:12:31.267281 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9872be1e-b995-4a00-904b-875dff96ad56\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f74a32a8fefea704cc73ffd1114083dab68aa44dc33f8d9d432d2d24216546c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbe00b46063d3968d59c393b87178e135f982abb7d0f32fa798b212d4b83df6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eedc9988e1a4a3c8cae730dc4a352103a6d665535b825e157d0f87c331b137c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2137fee4b7abc9765f9517d7b3b6ea796601a31d03151bf167d17136ced8b25a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f683e48f3577a68862eb6302383538376e41f49b20b1e0ea3f1b3b143f24872a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5de794187456b3576b969069a3b839d75c7918418a97fa35dff8bb40bb76e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5de794187456b3576b969069a3b839d75c7918418a97fa35dff8bb40bb76e39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08ce42b974fba00c409c0af971f0de351e50f18b2f06f8e722150da3239bb908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08ce42b974fba00c409c0af971f0de351e50f18b2f06f8e722150da3239bb908\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cfac6be03e3cfa0b0f3450b48b67957f2d5172b9c65e27f14c2a260bda5d423f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cfac6be03e3cfa0b0f3450b48b67957f2d5172b9c65e27f14c2a260bda5d423f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:31Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:31 crc kubenswrapper[4860]: I1211 08:12:31.281525 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:31Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:31 crc kubenswrapper[4860]: I1211 08:12:31.292765 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d8c2f28628dc2f1ffbdd98a55670e22cf6cdc8cf2c6442e9d1f8e5466b2f12e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://074ea63beac52bd196507ee7cfb1c3d7fc2d37d5ec6faa4d47dde62b922e1c43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:31Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:31 crc kubenswrapper[4860]: I1211 08:12:31.302403 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:31Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:31 crc kubenswrapper[4860]: I1211 08:12:31.303206 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:31 crc kubenswrapper[4860]: I1211 08:12:31.303265 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:31 crc kubenswrapper[4860]: I1211 08:12:31.303277 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:31 crc kubenswrapper[4860]: I1211 08:12:31.303316 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:31 crc kubenswrapper[4860]: I1211 08:12:31.303328 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:31Z","lastTransitionTime":"2025-12-11T08:12:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:31 crc kubenswrapper[4860]: I1211 08:12:31.312509 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25a6a154da2b74952951ae94efc914ba078ea4bbe5893360b3c84d46be6bf973\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:31Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:31 crc kubenswrapper[4860]: I1211 08:12:31.323587 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6xq7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a75ce3c-5e1e-4b9f-bbde-7a3b0cce5eae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe8d4d675567b41728e476351811a93c47b19868ccbadab8c613a4eb8eca8562\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5c6nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6xq7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:31Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:31 crc kubenswrapper[4860]: I1211 08:12:31.336606 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"910ec852-e5d2-4370-806d-77f16a36e630\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://122eb7e14a042027de3aafaa8db89562d80565e7df20f302f872e48d412a26b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1348b83ed7b1e8c25f9dfb1b2d2d035b5adc96dc23854ee90236f06fb155f155\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://867cdac386b0576ee2958a520afd522a2558a8878ea6b5af977b136dc7bba4a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4602fc6635e6188bf4cb75c5c7cf1e71a64fc0c5420626bfafd87beae150722d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4602fc6635e6188bf4cb75c5c7cf1e71a64fc0c5420626bfafd87beae150722d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:31Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:31 crc kubenswrapper[4860]: I1211 08:12:31.349872 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:31Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:31 crc kubenswrapper[4860]: I1211 08:12:31.362370 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a41f35db087418d1be16953f10377042abd3f4488e5ac1bc29ca5dcfc00a37e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:31Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:31 crc kubenswrapper[4860]: I1211 08:12:31.374858 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gjhlj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5cfd640a-f3d2-4b28-87e1-5f49f6341e21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knq7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knq7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:42Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gjhlj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:31Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:31 crc kubenswrapper[4860]: I1211 08:12:31.390056 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ec3758-85f1-4a22-a4ec-c805472607a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e831db1a15976d3967d56479689572c26f2599a766baa0402897a3008428c56f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a498aa90f914e1f0370df01dc5ef174df113a2a16d86b2e3f24b19a1b99b4c0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8caebd2f4dfcbb89c597029a07c335171889761d2d200206987537fb0a228db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34e91f615b76445ff3241a304bb2964fde59de7e8a8dfff6841b2e628a31e86d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2090cab02ad74ddfd6bb57ff50eba35d18f56b0cde1c230fe011c01994b693bc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 08:11:21.171011 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 08:11:21.177674 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-644352688/tls.crt::/tmp/serving-cert-644352688/tls.key\\\\\\\"\\\\nI1211 08:11:26.508606 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 08:11:26.512741 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 08:11:26.512778 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 08:11:26.512839 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 08:11:26.512858 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 08:11:26.521684 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1211 08:11:26.521700 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1211 08:11:26.521722 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:11:26.521732 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:11:26.521738 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 08:11:26.521746 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 08:11:26.521749 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 08:11:26.521753 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1211 08:11:26.524999 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2274fb298ca7d2b37b060b4dbf9b166d39ba77a108b08edeeb4d67196e72cb3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf0aff8a6ab816b6617b1feeca9146d41bb5e8808bf1db96f8d493500c135d85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf0aff8a6ab816b6617b1feeca9146d41bb5e8808bf1db96f8d493500c135d85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:31Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:31 crc kubenswrapper[4860]: I1211 08:12:31.405635 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:31 crc kubenswrapper[4860]: I1211 08:12:31.405697 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:31 crc kubenswrapper[4860]: I1211 08:12:31.405708 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:31 crc kubenswrapper[4860]: I1211 08:12:31.405722 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:31 crc kubenswrapper[4860]: I1211 08:12:31.405732 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:31Z","lastTransitionTime":"2025-12-11T08:12:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:31 crc kubenswrapper[4860]: I1211 08:12:31.508118 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:31 crc kubenswrapper[4860]: I1211 08:12:31.508196 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:31 crc kubenswrapper[4860]: I1211 08:12:31.508216 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:31 crc kubenswrapper[4860]: I1211 08:12:31.508243 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:31 crc kubenswrapper[4860]: I1211 08:12:31.508262 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:31Z","lastTransitionTime":"2025-12-11T08:12:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:31 crc kubenswrapper[4860]: I1211 08:12:31.610821 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:31 crc kubenswrapper[4860]: I1211 08:12:31.610874 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:31 crc kubenswrapper[4860]: I1211 08:12:31.610888 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:31 crc kubenswrapper[4860]: I1211 08:12:31.610912 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:31 crc kubenswrapper[4860]: I1211 08:12:31.610934 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:31Z","lastTransitionTime":"2025-12-11T08:12:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:31 crc kubenswrapper[4860]: I1211 08:12:31.714622 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:31 crc kubenswrapper[4860]: I1211 08:12:31.714753 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:31 crc kubenswrapper[4860]: I1211 08:12:31.714779 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:31 crc kubenswrapper[4860]: I1211 08:12:31.714813 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:31 crc kubenswrapper[4860]: I1211 08:12:31.714837 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:31Z","lastTransitionTime":"2025-12-11T08:12:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:31 crc kubenswrapper[4860]: I1211 08:12:31.817729 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:31 crc kubenswrapper[4860]: I1211 08:12:31.817798 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:31 crc kubenswrapper[4860]: I1211 08:12:31.817822 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:31 crc kubenswrapper[4860]: I1211 08:12:31.817845 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:31 crc kubenswrapper[4860]: I1211 08:12:31.817869 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:31Z","lastTransitionTime":"2025-12-11T08:12:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:31 crc kubenswrapper[4860]: I1211 08:12:31.920225 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:31 crc kubenswrapper[4860]: I1211 08:12:31.920265 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:31 crc kubenswrapper[4860]: I1211 08:12:31.920277 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:31 crc kubenswrapper[4860]: I1211 08:12:31.920295 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:31 crc kubenswrapper[4860]: I1211 08:12:31.920306 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:31Z","lastTransitionTime":"2025-12-11T08:12:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:32 crc kubenswrapper[4860]: I1211 08:12:32.023577 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:32 crc kubenswrapper[4860]: I1211 08:12:32.023701 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:32 crc kubenswrapper[4860]: I1211 08:12:32.023721 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:32 crc kubenswrapper[4860]: I1211 08:12:32.023743 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:32 crc kubenswrapper[4860]: I1211 08:12:32.023756 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:32Z","lastTransitionTime":"2025-12-11T08:12:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:32 crc kubenswrapper[4860]: I1211 08:12:32.125420 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:32 crc kubenswrapper[4860]: I1211 08:12:32.125467 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:32 crc kubenswrapper[4860]: I1211 08:12:32.125481 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:32 crc kubenswrapper[4860]: I1211 08:12:32.125498 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:32 crc kubenswrapper[4860]: I1211 08:12:32.125508 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-xvnqp_0589a204-c98c-417d-8256-bfe3bf747660/ovnkube-controller/3.log" Dec 11 08:12:32 crc kubenswrapper[4860]: I1211 08:12:32.125512 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:32Z","lastTransitionTime":"2025-12-11T08:12:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:32 crc kubenswrapper[4860]: I1211 08:12:32.130177 4860 scope.go:117] "RemoveContainer" containerID="d9e83cda3e40d325604114a37fe6e8b707ac1f91a034b0d979c70b5f4f24f2e7" Dec 11 08:12:32 crc kubenswrapper[4860]: E1211 08:12:32.130461 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-xvnqp_openshift-ovn-kubernetes(0589a204-c98c-417d-8256-bfe3bf747660)\"" pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" podUID="0589a204-c98c-417d-8256-bfe3bf747660" Dec 11 08:12:32 crc kubenswrapper[4860]: I1211 08:12:32.154053 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9872be1e-b995-4a00-904b-875dff96ad56\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f74a32a8fefea704cc73ffd1114083dab68aa44dc33f8d9d432d2d24216546c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbe00b46063d3968d59c393b87178e135f982abb7d0f32fa798b212d4b83df6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eedc9988e1a4a3c8cae730dc4a352103a6d665535b825e157d0f87c331b137c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2137fee4b7abc9765f9517d7b3b6ea796601a31d03151bf167d17136ced8b25a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f683e48f3577a68862eb6302383538376e41f49b20b1e0ea3f1b3b143f24872a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5de794187456b3576b969069a3b839d75c7918418a97fa35dff8bb40bb76e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5de794187456b3576b969069a3b839d75c7918418a97fa35dff8bb40bb76e39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08ce42b974fba00c409c0af971f0de351e50f18b2f06f8e722150da3239bb908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08ce42b974fba00c409c0af971f0de351e50f18b2f06f8e722150da3239bb908\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cfac6be03e3cfa0b0f3450b48b67957f2d5172b9c65e27f14c2a260bda5d423f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cfac6be03e3cfa0b0f3450b48b67957f2d5172b9c65e27f14c2a260bda5d423f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:32Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:32 crc kubenswrapper[4860]: I1211 08:12:32.168986 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-pznwl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5310c1fc-66c6-40aa-b1bf-5a59a2410f9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34119a6367261d7e12174f914ee8cfaa55703a55210f83c58b0e4694c54a5b03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2d6ff00ce322a28607bc3f6db05c5e4fb14acb6777ea315786ef53963ca504a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-11T08:12:17Z\\\",\\\"message\\\":\\\"2025-12-11T08:11:31+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_f37a703f-1609-4581-ab99-50337287ff33\\\\n2025-12-11T08:11:31+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_f37a703f-1609-4581-ab99-50337287ff33 to /host/opt/cni/bin/\\\\n2025-12-11T08:11:32Z [verbose] multus-daemon started\\\\n2025-12-11T08:11:32Z [verbose] Readiness Indicator file check\\\\n2025-12-11T08:12:17Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:12:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hkt48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-pznwl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:32Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:32 crc kubenswrapper[4860]: I1211 08:12:32.186051 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-854v6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e6529ba7-2341-43ac-a187-95ab530fe775\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8986538a421f0184f52f288b0e171da695591caf2b10c67ad2dce1073a6e6016\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kh22b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0d4e3f761ad15ae1424204e40b9d3f60188ab4a9047bd4b5ac6a8a8ca1fb385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kh22b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-854v6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:32Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:32 crc kubenswrapper[4860]: I1211 08:12:32.202134 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"910ec852-e5d2-4370-806d-77f16a36e630\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://122eb7e14a042027de3aafaa8db89562d80565e7df20f302f872e48d412a26b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1348b83ed7b1e8c25f9dfb1b2d2d035b5adc96dc23854ee90236f06fb155f155\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://867cdac386b0576ee2958a520afd522a2558a8878ea6b5af977b136dc7bba4a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4602fc6635e6188bf4cb75c5c7cf1e71a64fc0c5420626bfafd87beae150722d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4602fc6635e6188bf4cb75c5c7cf1e71a64fc0c5420626bfafd87beae150722d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:32Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:32 crc kubenswrapper[4860]: I1211 08:12:32.227499 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:32Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:32 crc kubenswrapper[4860]: I1211 08:12:32.229167 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:32 crc kubenswrapper[4860]: I1211 08:12:32.229237 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:32 crc kubenswrapper[4860]: I1211 08:12:32.229263 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:32 crc kubenswrapper[4860]: I1211 08:12:32.229298 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:32 crc kubenswrapper[4860]: I1211 08:12:32.229399 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:32Z","lastTransitionTime":"2025-12-11T08:12:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:32 crc kubenswrapper[4860]: I1211 08:12:32.243570 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d8c2f28628dc2f1ffbdd98a55670e22cf6cdc8cf2c6442e9d1f8e5466b2f12e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://074ea63beac52bd196507ee7cfb1c3d7fc2d37d5ec6faa4d47dde62b922e1c43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:32Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:32 crc kubenswrapper[4860]: I1211 08:12:32.248988 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:32 crc kubenswrapper[4860]: I1211 08:12:32.249046 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:32 crc kubenswrapper[4860]: I1211 08:12:32.249064 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:32 crc kubenswrapper[4860]: I1211 08:12:32.249087 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:32 crc kubenswrapper[4860]: I1211 08:12:32.249103 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:32Z","lastTransitionTime":"2025-12-11T08:12:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:32 crc kubenswrapper[4860]: I1211 08:12:32.262581 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:32Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:32 crc kubenswrapper[4860]: E1211 08:12:32.265398 4860 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:12:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:12:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:12:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:12:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"52631a1a-bc27-4dd5-8309-5dc39b5dcc41\\\",\\\"systemUUID\\\":\\\"299d3cd2-e1ad-4f5c-8af4-f12c0a5d76e7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:32Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:32 crc kubenswrapper[4860]: I1211 08:12:32.271182 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:32 crc kubenswrapper[4860]: I1211 08:12:32.271217 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:32 crc kubenswrapper[4860]: I1211 08:12:32.271226 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:32 crc kubenswrapper[4860]: I1211 08:12:32.271240 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:32 crc kubenswrapper[4860]: I1211 08:12:32.271249 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:32Z","lastTransitionTime":"2025-12-11T08:12:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:32 crc kubenswrapper[4860]: I1211 08:12:32.280153 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25a6a154da2b74952951ae94efc914ba078ea4bbe5893360b3c84d46be6bf973\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:32Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:32 crc kubenswrapper[4860]: E1211 08:12:32.285989 4860 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:12:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:12:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:12:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:12:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"52631a1a-bc27-4dd5-8309-5dc39b5dcc41\\\",\\\"systemUUID\\\":\\\"299d3cd2-e1ad-4f5c-8af4-f12c0a5d76e7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:32Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:32 crc kubenswrapper[4860]: I1211 08:12:32.289685 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:32 crc kubenswrapper[4860]: I1211 08:12:32.289734 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:32 crc kubenswrapper[4860]: I1211 08:12:32.289749 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:32 crc kubenswrapper[4860]: I1211 08:12:32.289768 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:32 crc kubenswrapper[4860]: I1211 08:12:32.289782 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:32Z","lastTransitionTime":"2025-12-11T08:12:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:32 crc kubenswrapper[4860]: I1211 08:12:32.292978 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6xq7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a75ce3c-5e1e-4b9f-bbde-7a3b0cce5eae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe8d4d675567b41728e476351811a93c47b19868ccbadab8c613a4eb8eca8562\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5c6nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6xq7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:32Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:32 crc kubenswrapper[4860]: E1211 08:12:32.303813 4860 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:12:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:12:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:12:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:12:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"52631a1a-bc27-4dd5-8309-5dc39b5dcc41\\\",\\\"systemUUID\\\":\\\"299d3cd2-e1ad-4f5c-8af4-f12c0a5d76e7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:32Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:32 crc kubenswrapper[4860]: I1211 08:12:32.306136 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ec3758-85f1-4a22-a4ec-c805472607a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e831db1a15976d3967d56479689572c26f2599a766baa0402897a3008428c56f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a498aa90f914e1f0370df01dc5ef174df113a2a16d86b2e3f24b19a1b99b4c0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8caebd2f4dfcbb89c597029a07c335171889761d2d200206987537fb0a228db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34e91f615b76445ff3241a304bb2964fde59de7e8a8dfff6841b2e628a31e86d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2090cab02ad74ddfd6bb57ff50eba35d18f56b0cde1c230fe011c01994b693bc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 08:11:21.171011 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 08:11:21.177674 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-644352688/tls.crt::/tmp/serving-cert-644352688/tls.key\\\\\\\"\\\\nI1211 08:11:26.508606 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 08:11:26.512741 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 08:11:26.512778 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 08:11:26.512839 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 08:11:26.512858 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 08:11:26.521684 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1211 08:11:26.521700 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1211 08:11:26.521722 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:11:26.521732 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:11:26.521738 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 08:11:26.521746 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 08:11:26.521749 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 08:11:26.521753 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1211 08:11:26.524999 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2274fb298ca7d2b37b060b4dbf9b166d39ba77a108b08edeeb4d67196e72cb3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf0aff8a6ab816b6617b1feeca9146d41bb5e8808bf1db96f8d493500c135d85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf0aff8a6ab816b6617b1feeca9146d41bb5e8808bf1db96f8d493500c135d85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:32Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:32 crc kubenswrapper[4860]: I1211 08:12:32.308550 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:32 crc kubenswrapper[4860]: I1211 08:12:32.308579 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:32 crc kubenswrapper[4860]: I1211 08:12:32.308590 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:32 crc kubenswrapper[4860]: I1211 08:12:32.308606 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:32 crc kubenswrapper[4860]: I1211 08:12:32.308616 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:32Z","lastTransitionTime":"2025-12-11T08:12:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:32 crc kubenswrapper[4860]: I1211 08:12:32.318863 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:32Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:32 crc kubenswrapper[4860]: E1211 08:12:32.320571 4860 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:12:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:12:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:12:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:12:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"52631a1a-bc27-4dd5-8309-5dc39b5dcc41\\\",\\\"systemUUID\\\":\\\"299d3cd2-e1ad-4f5c-8af4-f12c0a5d76e7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:32Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:32 crc kubenswrapper[4860]: I1211 08:12:32.323801 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:32 crc kubenswrapper[4860]: I1211 08:12:32.323837 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:32 crc kubenswrapper[4860]: I1211 08:12:32.323847 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:32 crc kubenswrapper[4860]: I1211 08:12:32.323861 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:32 crc kubenswrapper[4860]: I1211 08:12:32.323870 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:32Z","lastTransitionTime":"2025-12-11T08:12:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:32 crc kubenswrapper[4860]: I1211 08:12:32.333628 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a41f35db087418d1be16953f10377042abd3f4488e5ac1bc29ca5dcfc00a37e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:32Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:32 crc kubenswrapper[4860]: E1211 08:12:32.335305 4860 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:12:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:12:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:12:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:12:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"52631a1a-bc27-4dd5-8309-5dc39b5dcc41\\\",\\\"systemUUID\\\":\\\"299d3cd2-e1ad-4f5c-8af4-f12c0a5d76e7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:32Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:32 crc kubenswrapper[4860]: E1211 08:12:32.335434 4860 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 11 08:12:32 crc kubenswrapper[4860]: I1211 08:12:32.337028 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:32 crc kubenswrapper[4860]: I1211 08:12:32.337056 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:32 crc kubenswrapper[4860]: I1211 08:12:32.337067 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:32 crc kubenswrapper[4860]: I1211 08:12:32.337083 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:32 crc kubenswrapper[4860]: I1211 08:12:32.337095 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:32Z","lastTransitionTime":"2025-12-11T08:12:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:32 crc kubenswrapper[4860]: I1211 08:12:32.344361 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gjhlj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5cfd640a-f3d2-4b28-87e1-5f49f6341e21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knq7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knq7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:42Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gjhlj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:32Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:32 crc kubenswrapper[4860]: I1211 08:12:32.355687 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8bf0e9df-4e9d-4f45-a5b9-0020fdab2ea6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77c0c831f046149002cd12564aa9b1544a813da1fcd81e61451eeb504f0450b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3498078e25d1156a89ddc17f4823182b1aeb8b283f8b46bf66df67d67c5c5e44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3498078e25d1156a89ddc17f4823182b1aeb8b283f8b46bf66df67d67c5c5e44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:32Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:32 crc kubenswrapper[4860]: I1211 08:12:32.368977 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9aa0aed-62b3-4ea2-82ae-9a653c1101a5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ad4ba1b769c6cb2b34073798f8e3c25eb0177e6d3059484bf365e6981cab710\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://077aa1dd2fe5189ff60a0f2365e47a6a545cca0de7aa6379370908db4ea5838a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d7c919899b74a2b0fcd9221f3f769a9d013267354b200a282669d342d6413d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5483384f1d2ebc1fefa9706c9e67a3dc5a00ff643f803e3f2707c252e27ad94\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:32Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:32 crc kubenswrapper[4860]: I1211 08:12:32.380679 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31c30642-6e60-41b4-a477-0d802424e0aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79969402f5c33852a26fc839f37a45febf4f2377d77823284fbb01769e5585fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2d4d5de7106175087d91acb1f0c7552014f0c08bfcd65cbbf184d3f8d7826ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-99qgp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:32Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:32 crc kubenswrapper[4860]: I1211 08:12:32.394730 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-26nc6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1faa733-f408-41fd-b7aa-0cd86283976a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fce5b642b843ed30823c94e0b0c3d8bb1e5d2d93d691f7b8ef901ece03b0ac6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ac45db4f858a1c18a6b87472fc1cea2300f5845f1ad62c22260374995548a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ac45db4f858a1c18a6b87472fc1cea2300f5845f1ad62c22260374995548a0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b35296a5eb17dd2bb5e72c86c1cafdb30b7712cb330264f935d3b4fff05cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c6b35296a5eb17dd2bb5e72c86c1cafdb30b7712cb330264f935d3b4fff05cd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b9e82d6d3a247309ef5bdbf5b3a86ec083add2a138b0b237801e77b09ee96a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b9e82d6d3a247309ef5bdbf5b3a86ec083add2a138b0b237801e77b09ee96a1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://908ee695e78145ef4375c4097d9f1e772502922d02a2fcb4566172a6ad9caaf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://908ee695e78145ef4375c4097d9f1e772502922d02a2fcb4566172a6ad9caaf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afb0a097f3556a8f9ed7cf8f762cb99c082fe0a36b8d8d3038bf188b08fe9a84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afb0a097f3556a8f9ed7cf8f762cb99c082fe0a36b8d8d3038bf188b08fe9a84\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce18f4f06db091e658684fa0bbc858e19b9f03010acf13b3ffccb1a6e3f88021\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce18f4f06db091e658684fa0bbc858e19b9f03010acf13b3ffccb1a6e3f88021\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-26nc6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:32Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:32 crc kubenswrapper[4860]: I1211 08:12:32.411866 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0589a204-c98c-417d-8256-bfe3bf747660\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3333b4346f59b04017e32d61f457ba7ff457166731128f339d7987ae6b56b06d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52bb8ade314afb9494f4e026d0af2aea7b1970d64dc66102f55e257ae4678154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dfcc4370edaae36a5a32b0a9c29290c50f8e4f427e4e147dc3c779beaa8b780\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d856ae95e9077524c518d82d10fb6b93b7fbcaaf2b720ffafe6f7fae58bd7caa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1426084ac580895c42354630910dd8c8e3a48ddd25f22abce1d28dc3de332a63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9984d5d242bb258405efe91fc9caf9e7ea592834445df2e78add9553a921c76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9e83cda3e40d325604114a37fe6e8b707ac1f91a034b0d979c70b5f4f24f2e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9e83cda3e40d325604114a37fe6e8b707ac1f91a034b0d979c70b5f4f24f2e7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-11T08:12:30Z\\\",\\\"message\\\":\\\"val\\\\nI1211 08:12:30.086873 6903 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1211 08:12:30.086964 6903 handler.go:208] Removed *v1.Node event handler 7\\\\nI1211 08:12:30.086994 6903 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1211 08:12:30.087035 6903 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1211 08:12:30.087066 6903 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1211 08:12:30.087141 6903 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1211 08:12:30.087543 6903 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1211 08:12:30.087550 6903 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1211 08:12:30.087568 6903 handler.go:208] Removed *v1.Node event handler 2\\\\nI1211 08:12:30.087967 6903 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1211 08:12:30.087990 6903 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1211 08:12:30.088004 6903 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1211 08:12:30.088018 6903 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1211 08:12:30.088032 6903 factory.go:656] Stopping watch factory\\\\nI1211 08:12:30.088038 6903 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1211 08:12:30.088050 6903 ovnkube.go:599] Stopped ovnkube\\\\nI1211 08:12:3\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:12:29Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-xvnqp_openshift-ovn-kubernetes(0589a204-c98c-417d-8256-bfe3bf747660)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb80d9928e8349fc5963227d135355ad7ba820d4fd89877af601d005ca03334b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17cbcb49c06f3a9059cd5a0f05eedb8c785da30a3e502e31bae2f57b3486cb14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17cbcb49c06f3a9059cd5a0f05eedb8c785da30a3e502e31bae2f57b3486cb14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xvnqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:32Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:32 crc kubenswrapper[4860]: I1211 08:12:32.421396 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gsx9l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aba5b6b9-f88d-4763-9562-6b4dec606df9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://735ef012c523ddc952d46a431f27b2582a2c301d9e9da0c697a20ff6e2fb5d86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6zw9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gsx9l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:32Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:32 crc kubenswrapper[4860]: I1211 08:12:32.440068 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:32 crc kubenswrapper[4860]: I1211 08:12:32.440102 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:32 crc kubenswrapper[4860]: I1211 08:12:32.440113 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:32 crc kubenswrapper[4860]: I1211 08:12:32.440127 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:32 crc kubenswrapper[4860]: I1211 08:12:32.440137 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:32Z","lastTransitionTime":"2025-12-11T08:12:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:32 crc kubenswrapper[4860]: I1211 08:12:32.543299 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:32 crc kubenswrapper[4860]: I1211 08:12:32.543347 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:32 crc kubenswrapper[4860]: I1211 08:12:32.543355 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:32 crc kubenswrapper[4860]: I1211 08:12:32.543369 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:32 crc kubenswrapper[4860]: I1211 08:12:32.543378 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:32Z","lastTransitionTime":"2025-12-11T08:12:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:32 crc kubenswrapper[4860]: I1211 08:12:32.578175 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gjhlj" Dec 11 08:12:32 crc kubenswrapper[4860]: I1211 08:12:32.578227 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:12:32 crc kubenswrapper[4860]: I1211 08:12:32.578217 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:12:32 crc kubenswrapper[4860]: I1211 08:12:32.578304 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:12:32 crc kubenswrapper[4860]: E1211 08:12:32.578452 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:12:32 crc kubenswrapper[4860]: E1211 08:12:32.578546 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:12:32 crc kubenswrapper[4860]: E1211 08:12:32.578855 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:12:32 crc kubenswrapper[4860]: E1211 08:12:32.578946 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gjhlj" podUID="5cfd640a-f3d2-4b28-87e1-5f49f6341e21" Dec 11 08:12:32 crc kubenswrapper[4860]: I1211 08:12:32.646039 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:32 crc kubenswrapper[4860]: I1211 08:12:32.646103 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:32 crc kubenswrapper[4860]: I1211 08:12:32.646123 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:32 crc kubenswrapper[4860]: I1211 08:12:32.646190 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:32 crc kubenswrapper[4860]: I1211 08:12:32.646210 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:32Z","lastTransitionTime":"2025-12-11T08:12:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:32 crc kubenswrapper[4860]: I1211 08:12:32.749598 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:32 crc kubenswrapper[4860]: I1211 08:12:32.749706 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:32 crc kubenswrapper[4860]: I1211 08:12:32.749744 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:32 crc kubenswrapper[4860]: I1211 08:12:32.749785 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:32 crc kubenswrapper[4860]: I1211 08:12:32.749808 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:32Z","lastTransitionTime":"2025-12-11T08:12:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:32 crc kubenswrapper[4860]: I1211 08:12:32.858749 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:32 crc kubenswrapper[4860]: I1211 08:12:32.859184 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:32 crc kubenswrapper[4860]: I1211 08:12:32.859639 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:32 crc kubenswrapper[4860]: I1211 08:12:32.860075 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:32 crc kubenswrapper[4860]: I1211 08:12:32.860485 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:32Z","lastTransitionTime":"2025-12-11T08:12:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:32 crc kubenswrapper[4860]: I1211 08:12:32.963833 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:32 crc kubenswrapper[4860]: I1211 08:12:32.963881 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:32 crc kubenswrapper[4860]: I1211 08:12:32.963896 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:32 crc kubenswrapper[4860]: I1211 08:12:32.963915 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:32 crc kubenswrapper[4860]: I1211 08:12:32.963927 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:32Z","lastTransitionTime":"2025-12-11T08:12:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:33 crc kubenswrapper[4860]: I1211 08:12:33.066527 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:33 crc kubenswrapper[4860]: I1211 08:12:33.066588 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:33 crc kubenswrapper[4860]: I1211 08:12:33.066605 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:33 crc kubenswrapper[4860]: I1211 08:12:33.066628 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:33 crc kubenswrapper[4860]: I1211 08:12:33.066670 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:33Z","lastTransitionTime":"2025-12-11T08:12:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:33 crc kubenswrapper[4860]: I1211 08:12:33.170546 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:33 crc kubenswrapper[4860]: I1211 08:12:33.170726 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:33 crc kubenswrapper[4860]: I1211 08:12:33.170806 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:33 crc kubenswrapper[4860]: I1211 08:12:33.170841 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:33 crc kubenswrapper[4860]: I1211 08:12:33.170913 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:33Z","lastTransitionTime":"2025-12-11T08:12:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:33 crc kubenswrapper[4860]: I1211 08:12:33.274371 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:33 crc kubenswrapper[4860]: I1211 08:12:33.274721 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:33 crc kubenswrapper[4860]: I1211 08:12:33.274970 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:33 crc kubenswrapper[4860]: I1211 08:12:33.275193 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:33 crc kubenswrapper[4860]: I1211 08:12:33.275374 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:33Z","lastTransitionTime":"2025-12-11T08:12:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:33 crc kubenswrapper[4860]: I1211 08:12:33.378604 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:33 crc kubenswrapper[4860]: I1211 08:12:33.379164 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:33 crc kubenswrapper[4860]: I1211 08:12:33.379332 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:33 crc kubenswrapper[4860]: I1211 08:12:33.379415 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:33 crc kubenswrapper[4860]: I1211 08:12:33.379428 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:33Z","lastTransitionTime":"2025-12-11T08:12:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:33 crc kubenswrapper[4860]: I1211 08:12:33.482754 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:33 crc kubenswrapper[4860]: I1211 08:12:33.482805 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:33 crc kubenswrapper[4860]: I1211 08:12:33.482821 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:33 crc kubenswrapper[4860]: I1211 08:12:33.482858 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:33 crc kubenswrapper[4860]: I1211 08:12:33.482876 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:33Z","lastTransitionTime":"2025-12-11T08:12:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:33 crc kubenswrapper[4860]: I1211 08:12:33.585377 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:33 crc kubenswrapper[4860]: I1211 08:12:33.585767 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:33 crc kubenswrapper[4860]: I1211 08:12:33.585784 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:33 crc kubenswrapper[4860]: I1211 08:12:33.585805 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:33 crc kubenswrapper[4860]: I1211 08:12:33.585818 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:33Z","lastTransitionTime":"2025-12-11T08:12:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:33 crc kubenswrapper[4860]: I1211 08:12:33.688378 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:33 crc kubenswrapper[4860]: I1211 08:12:33.688709 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:33 crc kubenswrapper[4860]: I1211 08:12:33.688821 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:33 crc kubenswrapper[4860]: I1211 08:12:33.688941 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:33 crc kubenswrapper[4860]: I1211 08:12:33.689061 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:33Z","lastTransitionTime":"2025-12-11T08:12:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:33 crc kubenswrapper[4860]: I1211 08:12:33.792603 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:33 crc kubenswrapper[4860]: I1211 08:12:33.792700 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:33 crc kubenswrapper[4860]: I1211 08:12:33.792719 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:33 crc kubenswrapper[4860]: I1211 08:12:33.792743 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:33 crc kubenswrapper[4860]: I1211 08:12:33.792760 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:33Z","lastTransitionTime":"2025-12-11T08:12:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:33 crc kubenswrapper[4860]: I1211 08:12:33.896337 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:33 crc kubenswrapper[4860]: I1211 08:12:33.896409 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:33 crc kubenswrapper[4860]: I1211 08:12:33.896428 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:33 crc kubenswrapper[4860]: I1211 08:12:33.896454 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:33 crc kubenswrapper[4860]: I1211 08:12:33.896473 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:33Z","lastTransitionTime":"2025-12-11T08:12:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:34 crc kubenswrapper[4860]: I1211 08:12:34.000565 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:34 crc kubenswrapper[4860]: I1211 08:12:34.000723 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:34 crc kubenswrapper[4860]: I1211 08:12:34.000757 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:34 crc kubenswrapper[4860]: I1211 08:12:34.000785 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:34 crc kubenswrapper[4860]: I1211 08:12:34.000803 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:34Z","lastTransitionTime":"2025-12-11T08:12:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:34 crc kubenswrapper[4860]: I1211 08:12:34.104971 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:34 crc kubenswrapper[4860]: I1211 08:12:34.105038 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:34 crc kubenswrapper[4860]: I1211 08:12:34.105054 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:34 crc kubenswrapper[4860]: I1211 08:12:34.105079 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:34 crc kubenswrapper[4860]: I1211 08:12:34.105096 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:34Z","lastTransitionTime":"2025-12-11T08:12:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:34 crc kubenswrapper[4860]: I1211 08:12:34.208624 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:34 crc kubenswrapper[4860]: I1211 08:12:34.208733 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:34 crc kubenswrapper[4860]: I1211 08:12:34.208758 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:34 crc kubenswrapper[4860]: I1211 08:12:34.208791 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:34 crc kubenswrapper[4860]: I1211 08:12:34.208853 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:34Z","lastTransitionTime":"2025-12-11T08:12:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:34 crc kubenswrapper[4860]: I1211 08:12:34.312776 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:34 crc kubenswrapper[4860]: I1211 08:12:34.312866 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:34 crc kubenswrapper[4860]: I1211 08:12:34.312891 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:34 crc kubenswrapper[4860]: I1211 08:12:34.312972 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:34 crc kubenswrapper[4860]: I1211 08:12:34.312996 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:34Z","lastTransitionTime":"2025-12-11T08:12:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:34 crc kubenswrapper[4860]: I1211 08:12:34.415992 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:34 crc kubenswrapper[4860]: I1211 08:12:34.416062 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:34 crc kubenswrapper[4860]: I1211 08:12:34.416091 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:34 crc kubenswrapper[4860]: I1211 08:12:34.416123 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:34 crc kubenswrapper[4860]: I1211 08:12:34.416141 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:34Z","lastTransitionTime":"2025-12-11T08:12:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:34 crc kubenswrapper[4860]: I1211 08:12:34.519087 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:34 crc kubenswrapper[4860]: I1211 08:12:34.519151 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:34 crc kubenswrapper[4860]: I1211 08:12:34.519177 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:34 crc kubenswrapper[4860]: I1211 08:12:34.519213 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:34 crc kubenswrapper[4860]: I1211 08:12:34.519236 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:34Z","lastTransitionTime":"2025-12-11T08:12:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:34 crc kubenswrapper[4860]: I1211 08:12:34.578608 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:12:34 crc kubenswrapper[4860]: E1211 08:12:34.578865 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:12:34 crc kubenswrapper[4860]: I1211 08:12:34.578924 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gjhlj" Dec 11 08:12:34 crc kubenswrapper[4860]: E1211 08:12:34.579062 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gjhlj" podUID="5cfd640a-f3d2-4b28-87e1-5f49f6341e21" Dec 11 08:12:34 crc kubenswrapper[4860]: I1211 08:12:34.579147 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:12:34 crc kubenswrapper[4860]: E1211 08:12:34.579273 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:12:34 crc kubenswrapper[4860]: I1211 08:12:34.579351 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:12:34 crc kubenswrapper[4860]: E1211 08:12:34.579432 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:12:34 crc kubenswrapper[4860]: I1211 08:12:34.621401 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:34 crc kubenswrapper[4860]: I1211 08:12:34.621487 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:34 crc kubenswrapper[4860]: I1211 08:12:34.621499 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:34 crc kubenswrapper[4860]: I1211 08:12:34.621515 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:34 crc kubenswrapper[4860]: I1211 08:12:34.621527 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:34Z","lastTransitionTime":"2025-12-11T08:12:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:34 crc kubenswrapper[4860]: I1211 08:12:34.724871 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:34 crc kubenswrapper[4860]: I1211 08:12:34.724943 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:34 crc kubenswrapper[4860]: I1211 08:12:34.724967 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:34 crc kubenswrapper[4860]: I1211 08:12:34.724996 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:34 crc kubenswrapper[4860]: I1211 08:12:34.725021 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:34Z","lastTransitionTime":"2025-12-11T08:12:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:34 crc kubenswrapper[4860]: I1211 08:12:34.827850 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:34 crc kubenswrapper[4860]: I1211 08:12:34.827914 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:34 crc kubenswrapper[4860]: I1211 08:12:34.827939 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:34 crc kubenswrapper[4860]: I1211 08:12:34.827966 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:34 crc kubenswrapper[4860]: I1211 08:12:34.827987 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:34Z","lastTransitionTime":"2025-12-11T08:12:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:34 crc kubenswrapper[4860]: I1211 08:12:34.931116 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:34 crc kubenswrapper[4860]: I1211 08:12:34.931188 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:34 crc kubenswrapper[4860]: I1211 08:12:34.931207 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:34 crc kubenswrapper[4860]: I1211 08:12:34.931230 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:34 crc kubenswrapper[4860]: I1211 08:12:34.931249 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:34Z","lastTransitionTime":"2025-12-11T08:12:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:35 crc kubenswrapper[4860]: I1211 08:12:35.034120 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:35 crc kubenswrapper[4860]: I1211 08:12:35.034183 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:35 crc kubenswrapper[4860]: I1211 08:12:35.034201 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:35 crc kubenswrapper[4860]: I1211 08:12:35.034225 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:35 crc kubenswrapper[4860]: I1211 08:12:35.034242 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:35Z","lastTransitionTime":"2025-12-11T08:12:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:35 crc kubenswrapper[4860]: I1211 08:12:35.137776 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:35 crc kubenswrapper[4860]: I1211 08:12:35.137839 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:35 crc kubenswrapper[4860]: I1211 08:12:35.137861 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:35 crc kubenswrapper[4860]: I1211 08:12:35.137891 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:35 crc kubenswrapper[4860]: I1211 08:12:35.137915 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:35Z","lastTransitionTime":"2025-12-11T08:12:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:35 crc kubenswrapper[4860]: I1211 08:12:35.240630 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:35 crc kubenswrapper[4860]: I1211 08:12:35.240740 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:35 crc kubenswrapper[4860]: I1211 08:12:35.240767 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:35 crc kubenswrapper[4860]: I1211 08:12:35.240802 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:35 crc kubenswrapper[4860]: I1211 08:12:35.240837 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:35Z","lastTransitionTime":"2025-12-11T08:12:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:35 crc kubenswrapper[4860]: I1211 08:12:35.344460 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:35 crc kubenswrapper[4860]: I1211 08:12:35.344513 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:35 crc kubenswrapper[4860]: I1211 08:12:35.344530 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:35 crc kubenswrapper[4860]: I1211 08:12:35.344556 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:35 crc kubenswrapper[4860]: I1211 08:12:35.344574 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:35Z","lastTransitionTime":"2025-12-11T08:12:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:35 crc kubenswrapper[4860]: I1211 08:12:35.447854 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:35 crc kubenswrapper[4860]: I1211 08:12:35.447909 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:35 crc kubenswrapper[4860]: I1211 08:12:35.447922 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:35 crc kubenswrapper[4860]: I1211 08:12:35.447943 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:35 crc kubenswrapper[4860]: I1211 08:12:35.447957 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:35Z","lastTransitionTime":"2025-12-11T08:12:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:35 crc kubenswrapper[4860]: I1211 08:12:35.551040 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:35 crc kubenswrapper[4860]: I1211 08:12:35.551128 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:35 crc kubenswrapper[4860]: I1211 08:12:35.551151 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:35 crc kubenswrapper[4860]: I1211 08:12:35.551175 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:35 crc kubenswrapper[4860]: I1211 08:12:35.551192 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:35Z","lastTransitionTime":"2025-12-11T08:12:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:35 crc kubenswrapper[4860]: I1211 08:12:35.655264 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:35 crc kubenswrapper[4860]: I1211 08:12:35.655346 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:35 crc kubenswrapper[4860]: I1211 08:12:35.655365 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:35 crc kubenswrapper[4860]: I1211 08:12:35.655400 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:35 crc kubenswrapper[4860]: I1211 08:12:35.655421 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:35Z","lastTransitionTime":"2025-12-11T08:12:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:35 crc kubenswrapper[4860]: I1211 08:12:35.757903 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:35 crc kubenswrapper[4860]: I1211 08:12:35.757973 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:35 crc kubenswrapper[4860]: I1211 08:12:35.757989 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:35 crc kubenswrapper[4860]: I1211 08:12:35.758041 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:35 crc kubenswrapper[4860]: I1211 08:12:35.758055 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:35Z","lastTransitionTime":"2025-12-11T08:12:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:35 crc kubenswrapper[4860]: I1211 08:12:35.860782 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:35 crc kubenswrapper[4860]: I1211 08:12:35.860898 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:35 crc kubenswrapper[4860]: I1211 08:12:35.860918 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:35 crc kubenswrapper[4860]: I1211 08:12:35.860943 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:35 crc kubenswrapper[4860]: I1211 08:12:35.860960 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:35Z","lastTransitionTime":"2025-12-11T08:12:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:35 crc kubenswrapper[4860]: I1211 08:12:35.964149 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:35 crc kubenswrapper[4860]: I1211 08:12:35.964210 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:35 crc kubenswrapper[4860]: I1211 08:12:35.964229 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:35 crc kubenswrapper[4860]: I1211 08:12:35.964252 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:35 crc kubenswrapper[4860]: I1211 08:12:35.964286 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:35Z","lastTransitionTime":"2025-12-11T08:12:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:36 crc kubenswrapper[4860]: I1211 08:12:36.067542 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:36 crc kubenswrapper[4860]: I1211 08:12:36.067610 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:36 crc kubenswrapper[4860]: I1211 08:12:36.067628 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:36 crc kubenswrapper[4860]: I1211 08:12:36.067677 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:36 crc kubenswrapper[4860]: I1211 08:12:36.067696 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:36Z","lastTransitionTime":"2025-12-11T08:12:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:36 crc kubenswrapper[4860]: I1211 08:12:36.170884 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:36 crc kubenswrapper[4860]: I1211 08:12:36.170942 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:36 crc kubenswrapper[4860]: I1211 08:12:36.170964 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:36 crc kubenswrapper[4860]: I1211 08:12:36.170992 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:36 crc kubenswrapper[4860]: I1211 08:12:36.171015 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:36Z","lastTransitionTime":"2025-12-11T08:12:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:36 crc kubenswrapper[4860]: I1211 08:12:36.273826 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:36 crc kubenswrapper[4860]: I1211 08:12:36.273856 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:36 crc kubenswrapper[4860]: I1211 08:12:36.273864 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:36 crc kubenswrapper[4860]: I1211 08:12:36.273877 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:36 crc kubenswrapper[4860]: I1211 08:12:36.273885 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:36Z","lastTransitionTime":"2025-12-11T08:12:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:36 crc kubenswrapper[4860]: I1211 08:12:36.377184 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:36 crc kubenswrapper[4860]: I1211 08:12:36.377267 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:36 crc kubenswrapper[4860]: I1211 08:12:36.377285 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:36 crc kubenswrapper[4860]: I1211 08:12:36.377312 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:36 crc kubenswrapper[4860]: I1211 08:12:36.377330 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:36Z","lastTransitionTime":"2025-12-11T08:12:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:36 crc kubenswrapper[4860]: I1211 08:12:36.481082 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:36 crc kubenswrapper[4860]: I1211 08:12:36.481169 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:36 crc kubenswrapper[4860]: I1211 08:12:36.481187 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:36 crc kubenswrapper[4860]: I1211 08:12:36.481212 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:36 crc kubenswrapper[4860]: I1211 08:12:36.481232 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:36Z","lastTransitionTime":"2025-12-11T08:12:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:36 crc kubenswrapper[4860]: I1211 08:12:36.578866 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:12:36 crc kubenswrapper[4860]: I1211 08:12:36.578989 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:12:36 crc kubenswrapper[4860]: I1211 08:12:36.579192 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:12:36 crc kubenswrapper[4860]: I1211 08:12:36.579247 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gjhlj" Dec 11 08:12:36 crc kubenswrapper[4860]: E1211 08:12:36.579369 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:12:36 crc kubenswrapper[4860]: E1211 08:12:36.579493 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gjhlj" podUID="5cfd640a-f3d2-4b28-87e1-5f49f6341e21" Dec 11 08:12:36 crc kubenswrapper[4860]: E1211 08:12:36.579699 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:12:36 crc kubenswrapper[4860]: E1211 08:12:36.579811 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:12:36 crc kubenswrapper[4860]: I1211 08:12:36.583524 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:36 crc kubenswrapper[4860]: I1211 08:12:36.583579 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:36 crc kubenswrapper[4860]: I1211 08:12:36.583598 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:36 crc kubenswrapper[4860]: I1211 08:12:36.583622 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:36 crc kubenswrapper[4860]: I1211 08:12:36.583639 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:36Z","lastTransitionTime":"2025-12-11T08:12:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:36 crc kubenswrapper[4860]: I1211 08:12:36.687169 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:36 crc kubenswrapper[4860]: I1211 08:12:36.687251 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:36 crc kubenswrapper[4860]: I1211 08:12:36.687270 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:36 crc kubenswrapper[4860]: I1211 08:12:36.687306 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:36 crc kubenswrapper[4860]: I1211 08:12:36.687330 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:36Z","lastTransitionTime":"2025-12-11T08:12:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:36 crc kubenswrapper[4860]: I1211 08:12:36.790211 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:36 crc kubenswrapper[4860]: I1211 08:12:36.790289 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:36 crc kubenswrapper[4860]: I1211 08:12:36.790309 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:36 crc kubenswrapper[4860]: I1211 08:12:36.790796 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:36 crc kubenswrapper[4860]: I1211 08:12:36.790836 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:36Z","lastTransitionTime":"2025-12-11T08:12:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:36 crc kubenswrapper[4860]: I1211 08:12:36.893345 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:36 crc kubenswrapper[4860]: I1211 08:12:36.893378 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:36 crc kubenswrapper[4860]: I1211 08:12:36.893389 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:36 crc kubenswrapper[4860]: I1211 08:12:36.893406 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:36 crc kubenswrapper[4860]: I1211 08:12:36.893416 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:36Z","lastTransitionTime":"2025-12-11T08:12:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:36 crc kubenswrapper[4860]: I1211 08:12:36.995775 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:36 crc kubenswrapper[4860]: I1211 08:12:36.995847 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:36 crc kubenswrapper[4860]: I1211 08:12:36.995866 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:36 crc kubenswrapper[4860]: I1211 08:12:36.995892 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:36 crc kubenswrapper[4860]: I1211 08:12:36.995910 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:36Z","lastTransitionTime":"2025-12-11T08:12:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:37 crc kubenswrapper[4860]: I1211 08:12:37.098693 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:37 crc kubenswrapper[4860]: I1211 08:12:37.098764 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:37 crc kubenswrapper[4860]: I1211 08:12:37.098782 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:37 crc kubenswrapper[4860]: I1211 08:12:37.098806 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:37 crc kubenswrapper[4860]: I1211 08:12:37.098822 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:37Z","lastTransitionTime":"2025-12-11T08:12:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:37 crc kubenswrapper[4860]: I1211 08:12:37.202355 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:37 crc kubenswrapper[4860]: I1211 08:12:37.202424 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:37 crc kubenswrapper[4860]: I1211 08:12:37.202450 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:37 crc kubenswrapper[4860]: I1211 08:12:37.202480 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:37 crc kubenswrapper[4860]: I1211 08:12:37.202502 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:37Z","lastTransitionTime":"2025-12-11T08:12:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:37 crc kubenswrapper[4860]: I1211 08:12:37.304776 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:37 crc kubenswrapper[4860]: I1211 08:12:37.304824 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:37 crc kubenswrapper[4860]: I1211 08:12:37.304836 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:37 crc kubenswrapper[4860]: I1211 08:12:37.304856 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:37 crc kubenswrapper[4860]: I1211 08:12:37.304869 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:37Z","lastTransitionTime":"2025-12-11T08:12:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:37 crc kubenswrapper[4860]: I1211 08:12:37.407877 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:37 crc kubenswrapper[4860]: I1211 08:12:37.407912 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:37 crc kubenswrapper[4860]: I1211 08:12:37.407923 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:37 crc kubenswrapper[4860]: I1211 08:12:37.407938 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:37 crc kubenswrapper[4860]: I1211 08:12:37.407948 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:37Z","lastTransitionTime":"2025-12-11T08:12:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:37 crc kubenswrapper[4860]: I1211 08:12:37.511357 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:37 crc kubenswrapper[4860]: I1211 08:12:37.511449 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:37 crc kubenswrapper[4860]: I1211 08:12:37.511474 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:37 crc kubenswrapper[4860]: I1211 08:12:37.511505 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:37 crc kubenswrapper[4860]: I1211 08:12:37.511570 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:37Z","lastTransitionTime":"2025-12-11T08:12:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:37 crc kubenswrapper[4860]: I1211 08:12:37.601165 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ec3758-85f1-4a22-a4ec-c805472607a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e831db1a15976d3967d56479689572c26f2599a766baa0402897a3008428c56f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a498aa90f914e1f0370df01dc5ef174df113a2a16d86b2e3f24b19a1b99b4c0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8caebd2f4dfcbb89c597029a07c335171889761d2d200206987537fb0a228db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34e91f615b76445ff3241a304bb2964fde59de7e8a8dfff6841b2e628a31e86d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2090cab02ad74ddfd6bb57ff50eba35d18f56b0cde1c230fe011c01994b693bc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 08:11:21.171011 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 08:11:21.177674 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-644352688/tls.crt::/tmp/serving-cert-644352688/tls.key\\\\\\\"\\\\nI1211 08:11:26.508606 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 08:11:26.512741 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 08:11:26.512778 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 08:11:26.512839 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 08:11:26.512858 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 08:11:26.521684 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1211 08:11:26.521700 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1211 08:11:26.521722 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:11:26.521732 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:11:26.521738 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 08:11:26.521746 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 08:11:26.521749 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 08:11:26.521753 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1211 08:11:26.524999 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2274fb298ca7d2b37b060b4dbf9b166d39ba77a108b08edeeb4d67196e72cb3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf0aff8a6ab816b6617b1feeca9146d41bb5e8808bf1db96f8d493500c135d85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf0aff8a6ab816b6617b1feeca9146d41bb5e8808bf1db96f8d493500c135d85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:37Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:37 crc kubenswrapper[4860]: I1211 08:12:37.614718 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:37 crc kubenswrapper[4860]: I1211 08:12:37.614761 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:37 crc kubenswrapper[4860]: I1211 08:12:37.614775 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:37 crc kubenswrapper[4860]: I1211 08:12:37.614794 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:37 crc kubenswrapper[4860]: I1211 08:12:37.614807 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:37Z","lastTransitionTime":"2025-12-11T08:12:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:37 crc kubenswrapper[4860]: I1211 08:12:37.620770 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:37Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:37 crc kubenswrapper[4860]: I1211 08:12:37.642281 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a41f35db087418d1be16953f10377042abd3f4488e5ac1bc29ca5dcfc00a37e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:37Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:37 crc kubenswrapper[4860]: I1211 08:12:37.661507 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gjhlj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5cfd640a-f3d2-4b28-87e1-5f49f6341e21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knq7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knq7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:42Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gjhlj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:37Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:37 crc kubenswrapper[4860]: I1211 08:12:37.673493 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gsx9l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aba5b6b9-f88d-4763-9562-6b4dec606df9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://735ef012c523ddc952d46a431f27b2582a2c301d9e9da0c697a20ff6e2fb5d86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6zw9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gsx9l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:37Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:37 crc kubenswrapper[4860]: I1211 08:12:37.685926 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8bf0e9df-4e9d-4f45-a5b9-0020fdab2ea6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77c0c831f046149002cd12564aa9b1544a813da1fcd81e61451eeb504f0450b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3498078e25d1156a89ddc17f4823182b1aeb8b283f8b46bf66df67d67c5c5e44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3498078e25d1156a89ddc17f4823182b1aeb8b283f8b46bf66df67d67c5c5e44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:37Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:37 crc kubenswrapper[4860]: I1211 08:12:37.699456 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9aa0aed-62b3-4ea2-82ae-9a653c1101a5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ad4ba1b769c6cb2b34073798f8e3c25eb0177e6d3059484bf365e6981cab710\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://077aa1dd2fe5189ff60a0f2365e47a6a545cca0de7aa6379370908db4ea5838a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d7c919899b74a2b0fcd9221f3f769a9d013267354b200a282669d342d6413d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5483384f1d2ebc1fefa9706c9e67a3dc5a00ff643f803e3f2707c252e27ad94\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:37Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:37 crc kubenswrapper[4860]: I1211 08:12:37.717759 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:37 crc kubenswrapper[4860]: I1211 08:12:37.717803 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:37 crc kubenswrapper[4860]: I1211 08:12:37.717817 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:37 crc kubenswrapper[4860]: I1211 08:12:37.717833 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:37 crc kubenswrapper[4860]: I1211 08:12:37.717844 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:37Z","lastTransitionTime":"2025-12-11T08:12:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:37 crc kubenswrapper[4860]: I1211 08:12:37.718709 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31c30642-6e60-41b4-a477-0d802424e0aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79969402f5c33852a26fc839f37a45febf4f2377d77823284fbb01769e5585fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2d4d5de7106175087d91acb1f0c7552014f0c08bfcd65cbbf184d3f8d7826ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-99qgp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:37Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:37 crc kubenswrapper[4860]: I1211 08:12:37.736596 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-26nc6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1faa733-f408-41fd-b7aa-0cd86283976a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fce5b642b843ed30823c94e0b0c3d8bb1e5d2d93d691f7b8ef901ece03b0ac6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ac45db4f858a1c18a6b87472fc1cea2300f5845f1ad62c22260374995548a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ac45db4f858a1c18a6b87472fc1cea2300f5845f1ad62c22260374995548a0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b35296a5eb17dd2bb5e72c86c1cafdb30b7712cb330264f935d3b4fff05cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c6b35296a5eb17dd2bb5e72c86c1cafdb30b7712cb330264f935d3b4fff05cd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b9e82d6d3a247309ef5bdbf5b3a86ec083add2a138b0b237801e77b09ee96a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b9e82d6d3a247309ef5bdbf5b3a86ec083add2a138b0b237801e77b09ee96a1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://908ee695e78145ef4375c4097d9f1e772502922d02a2fcb4566172a6ad9caaf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://908ee695e78145ef4375c4097d9f1e772502922d02a2fcb4566172a6ad9caaf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afb0a097f3556a8f9ed7cf8f762cb99c082fe0a36b8d8d3038bf188b08fe9a84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afb0a097f3556a8f9ed7cf8f762cb99c082fe0a36b8d8d3038bf188b08fe9a84\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce18f4f06db091e658684fa0bbc858e19b9f03010acf13b3ffccb1a6e3f88021\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce18f4f06db091e658684fa0bbc858e19b9f03010acf13b3ffccb1a6e3f88021\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-26nc6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:37Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:37 crc kubenswrapper[4860]: I1211 08:12:37.755277 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0589a204-c98c-417d-8256-bfe3bf747660\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3333b4346f59b04017e32d61f457ba7ff457166731128f339d7987ae6b56b06d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52bb8ade314afb9494f4e026d0af2aea7b1970d64dc66102f55e257ae4678154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dfcc4370edaae36a5a32b0a9c29290c50f8e4f427e4e147dc3c779beaa8b780\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d856ae95e9077524c518d82d10fb6b93b7fbcaaf2b720ffafe6f7fae58bd7caa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1426084ac580895c42354630910dd8c8e3a48ddd25f22abce1d28dc3de332a63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9984d5d242bb258405efe91fc9caf9e7ea592834445df2e78add9553a921c76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9e83cda3e40d325604114a37fe6e8b707ac1f91a034b0d979c70b5f4f24f2e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9e83cda3e40d325604114a37fe6e8b707ac1f91a034b0d979c70b5f4f24f2e7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-11T08:12:30Z\\\",\\\"message\\\":\\\"val\\\\nI1211 08:12:30.086873 6903 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1211 08:12:30.086964 6903 handler.go:208] Removed *v1.Node event handler 7\\\\nI1211 08:12:30.086994 6903 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1211 08:12:30.087035 6903 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1211 08:12:30.087066 6903 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1211 08:12:30.087141 6903 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1211 08:12:30.087543 6903 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1211 08:12:30.087550 6903 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1211 08:12:30.087568 6903 handler.go:208] Removed *v1.Node event handler 2\\\\nI1211 08:12:30.087967 6903 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1211 08:12:30.087990 6903 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1211 08:12:30.088004 6903 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1211 08:12:30.088018 6903 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1211 08:12:30.088032 6903 factory.go:656] Stopping watch factory\\\\nI1211 08:12:30.088038 6903 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1211 08:12:30.088050 6903 ovnkube.go:599] Stopped ovnkube\\\\nI1211 08:12:3\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:12:29Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-xvnqp_openshift-ovn-kubernetes(0589a204-c98c-417d-8256-bfe3bf747660)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb80d9928e8349fc5963227d135355ad7ba820d4fd89877af601d005ca03334b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17cbcb49c06f3a9059cd5a0f05eedb8c785da30a3e502e31bae2f57b3486cb14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17cbcb49c06f3a9059cd5a0f05eedb8c785da30a3e502e31bae2f57b3486cb14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xvnqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:37Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:37 crc kubenswrapper[4860]: I1211 08:12:37.777082 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9872be1e-b995-4a00-904b-875dff96ad56\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f74a32a8fefea704cc73ffd1114083dab68aa44dc33f8d9d432d2d24216546c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbe00b46063d3968d59c393b87178e135f982abb7d0f32fa798b212d4b83df6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eedc9988e1a4a3c8cae730dc4a352103a6d665535b825e157d0f87c331b137c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2137fee4b7abc9765f9517d7b3b6ea796601a31d03151bf167d17136ced8b25a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f683e48f3577a68862eb6302383538376e41f49b20b1e0ea3f1b3b143f24872a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5de794187456b3576b969069a3b839d75c7918418a97fa35dff8bb40bb76e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5de794187456b3576b969069a3b839d75c7918418a97fa35dff8bb40bb76e39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08ce42b974fba00c409c0af971f0de351e50f18b2f06f8e722150da3239bb908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08ce42b974fba00c409c0af971f0de351e50f18b2f06f8e722150da3239bb908\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cfac6be03e3cfa0b0f3450b48b67957f2d5172b9c65e27f14c2a260bda5d423f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cfac6be03e3cfa0b0f3450b48b67957f2d5172b9c65e27f14c2a260bda5d423f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:37Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:37 crc kubenswrapper[4860]: I1211 08:12:37.794825 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-pznwl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5310c1fc-66c6-40aa-b1bf-5a59a2410f9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34119a6367261d7e12174f914ee8cfaa55703a55210f83c58b0e4694c54a5b03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2d6ff00ce322a28607bc3f6db05c5e4fb14acb6777ea315786ef53963ca504a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-11T08:12:17Z\\\",\\\"message\\\":\\\"2025-12-11T08:11:31+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_f37a703f-1609-4581-ab99-50337287ff33\\\\n2025-12-11T08:11:31+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_f37a703f-1609-4581-ab99-50337287ff33 to /host/opt/cni/bin/\\\\n2025-12-11T08:11:32Z [verbose] multus-daemon started\\\\n2025-12-11T08:11:32Z [verbose] Readiness Indicator file check\\\\n2025-12-11T08:12:17Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:12:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hkt48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-pznwl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:37Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:37 crc kubenswrapper[4860]: I1211 08:12:37.808027 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-854v6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e6529ba7-2341-43ac-a187-95ab530fe775\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8986538a421f0184f52f288b0e171da695591caf2b10c67ad2dce1073a6e6016\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kh22b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0d4e3f761ad15ae1424204e40b9d3f60188ab4a9047bd4b5ac6a8a8ca1fb385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kh22b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-854v6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:37Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:37 crc kubenswrapper[4860]: I1211 08:12:37.820050 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:37 crc kubenswrapper[4860]: I1211 08:12:37.820088 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:37 crc kubenswrapper[4860]: I1211 08:12:37.820098 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:37 crc kubenswrapper[4860]: I1211 08:12:37.820114 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:37 crc kubenswrapper[4860]: I1211 08:12:37.820126 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:37Z","lastTransitionTime":"2025-12-11T08:12:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:37 crc kubenswrapper[4860]: I1211 08:12:37.820544 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6xq7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a75ce3c-5e1e-4b9f-bbde-7a3b0cce5eae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe8d4d675567b41728e476351811a93c47b19868ccbadab8c613a4eb8eca8562\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5c6nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6xq7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:37Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:37 crc kubenswrapper[4860]: I1211 08:12:37.836590 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"910ec852-e5d2-4370-806d-77f16a36e630\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://122eb7e14a042027de3aafaa8db89562d80565e7df20f302f872e48d412a26b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1348b83ed7b1e8c25f9dfb1b2d2d035b5adc96dc23854ee90236f06fb155f155\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://867cdac386b0576ee2958a520afd522a2558a8878ea6b5af977b136dc7bba4a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4602fc6635e6188bf4cb75c5c7cf1e71a64fc0c5420626bfafd87beae150722d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4602fc6635e6188bf4cb75c5c7cf1e71a64fc0c5420626bfafd87beae150722d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:37Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:37 crc kubenswrapper[4860]: I1211 08:12:37.846586 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:37Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:37 crc kubenswrapper[4860]: I1211 08:12:37.863485 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d8c2f28628dc2f1ffbdd98a55670e22cf6cdc8cf2c6442e9d1f8e5466b2f12e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://074ea63beac52bd196507ee7cfb1c3d7fc2d37d5ec6faa4d47dde62b922e1c43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:37Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:37 crc kubenswrapper[4860]: I1211 08:12:37.882004 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:37Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:37 crc kubenswrapper[4860]: I1211 08:12:37.894550 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25a6a154da2b74952951ae94efc914ba078ea4bbe5893360b3c84d46be6bf973\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:37Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:37 crc kubenswrapper[4860]: I1211 08:12:37.922626 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:37 crc kubenswrapper[4860]: I1211 08:12:37.922741 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:37 crc kubenswrapper[4860]: I1211 08:12:37.922762 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:37 crc kubenswrapper[4860]: I1211 08:12:37.922786 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:37 crc kubenswrapper[4860]: I1211 08:12:37.922805 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:37Z","lastTransitionTime":"2025-12-11T08:12:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:38 crc kubenswrapper[4860]: I1211 08:12:38.025772 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:38 crc kubenswrapper[4860]: I1211 08:12:38.025806 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:38 crc kubenswrapper[4860]: I1211 08:12:38.025815 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:38 crc kubenswrapper[4860]: I1211 08:12:38.025827 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:38 crc kubenswrapper[4860]: I1211 08:12:38.025837 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:38Z","lastTransitionTime":"2025-12-11T08:12:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:38 crc kubenswrapper[4860]: I1211 08:12:38.128612 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:38 crc kubenswrapper[4860]: I1211 08:12:38.128756 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:38 crc kubenswrapper[4860]: I1211 08:12:38.128779 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:38 crc kubenswrapper[4860]: I1211 08:12:38.128808 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:38 crc kubenswrapper[4860]: I1211 08:12:38.128829 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:38Z","lastTransitionTime":"2025-12-11T08:12:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:38 crc kubenswrapper[4860]: I1211 08:12:38.231595 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:38 crc kubenswrapper[4860]: I1211 08:12:38.231661 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:38 crc kubenswrapper[4860]: I1211 08:12:38.231673 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:38 crc kubenswrapper[4860]: I1211 08:12:38.231690 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:38 crc kubenswrapper[4860]: I1211 08:12:38.231701 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:38Z","lastTransitionTime":"2025-12-11T08:12:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:38 crc kubenswrapper[4860]: I1211 08:12:38.335793 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:38 crc kubenswrapper[4860]: I1211 08:12:38.335837 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:38 crc kubenswrapper[4860]: I1211 08:12:38.335849 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:38 crc kubenswrapper[4860]: I1211 08:12:38.335868 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:38 crc kubenswrapper[4860]: I1211 08:12:38.335879 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:38Z","lastTransitionTime":"2025-12-11T08:12:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:38 crc kubenswrapper[4860]: I1211 08:12:38.438325 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:38 crc kubenswrapper[4860]: I1211 08:12:38.438393 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:38 crc kubenswrapper[4860]: I1211 08:12:38.438417 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:38 crc kubenswrapper[4860]: I1211 08:12:38.438446 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:38 crc kubenswrapper[4860]: I1211 08:12:38.438467 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:38Z","lastTransitionTime":"2025-12-11T08:12:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:38 crc kubenswrapper[4860]: I1211 08:12:38.542360 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:38 crc kubenswrapper[4860]: I1211 08:12:38.542435 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:38 crc kubenswrapper[4860]: I1211 08:12:38.542461 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:38 crc kubenswrapper[4860]: I1211 08:12:38.542493 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:38 crc kubenswrapper[4860]: I1211 08:12:38.542515 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:38Z","lastTransitionTime":"2025-12-11T08:12:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:38 crc kubenswrapper[4860]: I1211 08:12:38.578028 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:12:38 crc kubenswrapper[4860]: I1211 08:12:38.578113 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:12:38 crc kubenswrapper[4860]: I1211 08:12:38.578268 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gjhlj" Dec 11 08:12:38 crc kubenswrapper[4860]: E1211 08:12:38.578272 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:12:38 crc kubenswrapper[4860]: I1211 08:12:38.578325 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:12:38 crc kubenswrapper[4860]: E1211 08:12:38.578460 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gjhlj" podUID="5cfd640a-f3d2-4b28-87e1-5f49f6341e21" Dec 11 08:12:38 crc kubenswrapper[4860]: E1211 08:12:38.578689 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:12:38 crc kubenswrapper[4860]: E1211 08:12:38.578753 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:12:38 crc kubenswrapper[4860]: I1211 08:12:38.646434 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:38 crc kubenswrapper[4860]: I1211 08:12:38.646486 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:38 crc kubenswrapper[4860]: I1211 08:12:38.646498 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:38 crc kubenswrapper[4860]: I1211 08:12:38.646515 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:38 crc kubenswrapper[4860]: I1211 08:12:38.646534 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:38Z","lastTransitionTime":"2025-12-11T08:12:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:38 crc kubenswrapper[4860]: I1211 08:12:38.749514 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:38 crc kubenswrapper[4860]: I1211 08:12:38.749608 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:38 crc kubenswrapper[4860]: I1211 08:12:38.749627 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:38 crc kubenswrapper[4860]: I1211 08:12:38.749674 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:38 crc kubenswrapper[4860]: I1211 08:12:38.749692 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:38Z","lastTransitionTime":"2025-12-11T08:12:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:38 crc kubenswrapper[4860]: I1211 08:12:38.852939 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:38 crc kubenswrapper[4860]: I1211 08:12:38.852978 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:38 crc kubenswrapper[4860]: I1211 08:12:38.852987 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:38 crc kubenswrapper[4860]: I1211 08:12:38.853001 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:38 crc kubenswrapper[4860]: I1211 08:12:38.853011 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:38Z","lastTransitionTime":"2025-12-11T08:12:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:38 crc kubenswrapper[4860]: I1211 08:12:38.956444 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:38 crc kubenswrapper[4860]: I1211 08:12:38.956489 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:38 crc kubenswrapper[4860]: I1211 08:12:38.956501 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:38 crc kubenswrapper[4860]: I1211 08:12:38.956518 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:38 crc kubenswrapper[4860]: I1211 08:12:38.956529 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:38Z","lastTransitionTime":"2025-12-11T08:12:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:39 crc kubenswrapper[4860]: I1211 08:12:39.059330 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:39 crc kubenswrapper[4860]: I1211 08:12:39.059406 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:39 crc kubenswrapper[4860]: I1211 08:12:39.059422 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:39 crc kubenswrapper[4860]: I1211 08:12:39.059442 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:39 crc kubenswrapper[4860]: I1211 08:12:39.059455 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:39Z","lastTransitionTime":"2025-12-11T08:12:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:39 crc kubenswrapper[4860]: I1211 08:12:39.162520 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:39 crc kubenswrapper[4860]: I1211 08:12:39.162571 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:39 crc kubenswrapper[4860]: I1211 08:12:39.162582 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:39 crc kubenswrapper[4860]: I1211 08:12:39.162600 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:39 crc kubenswrapper[4860]: I1211 08:12:39.162613 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:39Z","lastTransitionTime":"2025-12-11T08:12:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:39 crc kubenswrapper[4860]: I1211 08:12:39.264750 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:39 crc kubenswrapper[4860]: I1211 08:12:39.264829 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:39 crc kubenswrapper[4860]: I1211 08:12:39.264858 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:39 crc kubenswrapper[4860]: I1211 08:12:39.264887 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:39 crc kubenswrapper[4860]: I1211 08:12:39.264908 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:39Z","lastTransitionTime":"2025-12-11T08:12:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:39 crc kubenswrapper[4860]: I1211 08:12:39.367052 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:39 crc kubenswrapper[4860]: I1211 08:12:39.367096 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:39 crc kubenswrapper[4860]: I1211 08:12:39.367108 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:39 crc kubenswrapper[4860]: I1211 08:12:39.367125 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:39 crc kubenswrapper[4860]: I1211 08:12:39.367139 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:39Z","lastTransitionTime":"2025-12-11T08:12:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:39 crc kubenswrapper[4860]: I1211 08:12:39.469531 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:39 crc kubenswrapper[4860]: I1211 08:12:39.469564 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:39 crc kubenswrapper[4860]: I1211 08:12:39.469572 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:39 crc kubenswrapper[4860]: I1211 08:12:39.469606 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:39 crc kubenswrapper[4860]: I1211 08:12:39.469616 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:39Z","lastTransitionTime":"2025-12-11T08:12:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:39 crc kubenswrapper[4860]: I1211 08:12:39.572208 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:39 crc kubenswrapper[4860]: I1211 08:12:39.572239 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:39 crc kubenswrapper[4860]: I1211 08:12:39.572247 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:39 crc kubenswrapper[4860]: I1211 08:12:39.572259 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:39 crc kubenswrapper[4860]: I1211 08:12:39.572268 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:39Z","lastTransitionTime":"2025-12-11T08:12:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:39 crc kubenswrapper[4860]: I1211 08:12:39.675564 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:39 crc kubenswrapper[4860]: I1211 08:12:39.675628 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:39 crc kubenswrapper[4860]: I1211 08:12:39.675677 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:39 crc kubenswrapper[4860]: I1211 08:12:39.675704 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:39 crc kubenswrapper[4860]: I1211 08:12:39.675728 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:39Z","lastTransitionTime":"2025-12-11T08:12:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:39 crc kubenswrapper[4860]: I1211 08:12:39.779813 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:39 crc kubenswrapper[4860]: I1211 08:12:39.779886 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:39 crc kubenswrapper[4860]: I1211 08:12:39.780009 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:39 crc kubenswrapper[4860]: I1211 08:12:39.780085 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:39 crc kubenswrapper[4860]: I1211 08:12:39.780102 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:39Z","lastTransitionTime":"2025-12-11T08:12:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:39 crc kubenswrapper[4860]: I1211 08:12:39.883938 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:39 crc kubenswrapper[4860]: I1211 08:12:39.884002 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:39 crc kubenswrapper[4860]: I1211 08:12:39.884016 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:39 crc kubenswrapper[4860]: I1211 08:12:39.884036 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:39 crc kubenswrapper[4860]: I1211 08:12:39.884054 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:39Z","lastTransitionTime":"2025-12-11T08:12:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:39 crc kubenswrapper[4860]: I1211 08:12:39.987512 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:39 crc kubenswrapper[4860]: I1211 08:12:39.987580 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:39 crc kubenswrapper[4860]: I1211 08:12:39.987599 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:39 crc kubenswrapper[4860]: I1211 08:12:39.987624 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:39 crc kubenswrapper[4860]: I1211 08:12:39.987648 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:39Z","lastTransitionTime":"2025-12-11T08:12:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:40 crc kubenswrapper[4860]: I1211 08:12:40.090205 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:40 crc kubenswrapper[4860]: I1211 08:12:40.090284 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:40 crc kubenswrapper[4860]: I1211 08:12:40.090312 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:40 crc kubenswrapper[4860]: I1211 08:12:40.090344 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:40 crc kubenswrapper[4860]: I1211 08:12:40.090368 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:40Z","lastTransitionTime":"2025-12-11T08:12:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:40 crc kubenswrapper[4860]: I1211 08:12:40.193193 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:40 crc kubenswrapper[4860]: I1211 08:12:40.193249 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:40 crc kubenswrapper[4860]: I1211 08:12:40.193270 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:40 crc kubenswrapper[4860]: I1211 08:12:40.193295 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:40 crc kubenswrapper[4860]: I1211 08:12:40.193312 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:40Z","lastTransitionTime":"2025-12-11T08:12:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:40 crc kubenswrapper[4860]: I1211 08:12:40.301297 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:40 crc kubenswrapper[4860]: I1211 08:12:40.302003 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:40 crc kubenswrapper[4860]: I1211 08:12:40.302065 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:40 crc kubenswrapper[4860]: I1211 08:12:40.302104 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:40 crc kubenswrapper[4860]: I1211 08:12:40.302127 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:40Z","lastTransitionTime":"2025-12-11T08:12:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:40 crc kubenswrapper[4860]: I1211 08:12:40.405258 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:40 crc kubenswrapper[4860]: I1211 08:12:40.405332 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:40 crc kubenswrapper[4860]: I1211 08:12:40.405348 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:40 crc kubenswrapper[4860]: I1211 08:12:40.405373 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:40 crc kubenswrapper[4860]: I1211 08:12:40.405390 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:40Z","lastTransitionTime":"2025-12-11T08:12:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:40 crc kubenswrapper[4860]: I1211 08:12:40.509071 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:40 crc kubenswrapper[4860]: I1211 08:12:40.509127 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:40 crc kubenswrapper[4860]: I1211 08:12:40.509144 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:40 crc kubenswrapper[4860]: I1211 08:12:40.509168 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:40 crc kubenswrapper[4860]: I1211 08:12:40.509184 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:40Z","lastTransitionTime":"2025-12-11T08:12:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:40 crc kubenswrapper[4860]: I1211 08:12:40.578970 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gjhlj" Dec 11 08:12:40 crc kubenswrapper[4860]: E1211 08:12:40.579166 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gjhlj" podUID="5cfd640a-f3d2-4b28-87e1-5f49f6341e21" Dec 11 08:12:40 crc kubenswrapper[4860]: I1211 08:12:40.579449 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:12:40 crc kubenswrapper[4860]: E1211 08:12:40.579547 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:12:40 crc kubenswrapper[4860]: I1211 08:12:40.579823 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:12:40 crc kubenswrapper[4860]: E1211 08:12:40.579995 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:12:40 crc kubenswrapper[4860]: I1211 08:12:40.580078 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:12:40 crc kubenswrapper[4860]: E1211 08:12:40.580233 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:12:40 crc kubenswrapper[4860]: I1211 08:12:40.612297 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:40 crc kubenswrapper[4860]: I1211 08:12:40.612359 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:40 crc kubenswrapper[4860]: I1211 08:12:40.612378 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:40 crc kubenswrapper[4860]: I1211 08:12:40.612412 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:40 crc kubenswrapper[4860]: I1211 08:12:40.612429 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:40Z","lastTransitionTime":"2025-12-11T08:12:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:40 crc kubenswrapper[4860]: I1211 08:12:40.715124 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:40 crc kubenswrapper[4860]: I1211 08:12:40.715191 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:40 crc kubenswrapper[4860]: I1211 08:12:40.715212 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:40 crc kubenswrapper[4860]: I1211 08:12:40.715243 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:40 crc kubenswrapper[4860]: I1211 08:12:40.715268 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:40Z","lastTransitionTime":"2025-12-11T08:12:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:40 crc kubenswrapper[4860]: I1211 08:12:40.818985 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:40 crc kubenswrapper[4860]: I1211 08:12:40.819043 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:40 crc kubenswrapper[4860]: I1211 08:12:40.819082 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:40 crc kubenswrapper[4860]: I1211 08:12:40.819114 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:40 crc kubenswrapper[4860]: I1211 08:12:40.819137 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:40Z","lastTransitionTime":"2025-12-11T08:12:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:40 crc kubenswrapper[4860]: I1211 08:12:40.922255 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:40 crc kubenswrapper[4860]: I1211 08:12:40.922309 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:40 crc kubenswrapper[4860]: I1211 08:12:40.922326 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:40 crc kubenswrapper[4860]: I1211 08:12:40.922348 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:40 crc kubenswrapper[4860]: I1211 08:12:40.922365 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:40Z","lastTransitionTime":"2025-12-11T08:12:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:41 crc kubenswrapper[4860]: I1211 08:12:41.025999 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:41 crc kubenswrapper[4860]: I1211 08:12:41.026072 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:41 crc kubenswrapper[4860]: I1211 08:12:41.026095 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:41 crc kubenswrapper[4860]: I1211 08:12:41.026125 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:41 crc kubenswrapper[4860]: I1211 08:12:41.026147 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:41Z","lastTransitionTime":"2025-12-11T08:12:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:41 crc kubenswrapper[4860]: I1211 08:12:41.128720 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:41 crc kubenswrapper[4860]: I1211 08:12:41.128787 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:41 crc kubenswrapper[4860]: I1211 08:12:41.128811 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:41 crc kubenswrapper[4860]: I1211 08:12:41.128841 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:41 crc kubenswrapper[4860]: I1211 08:12:41.128867 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:41Z","lastTransitionTime":"2025-12-11T08:12:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:41 crc kubenswrapper[4860]: I1211 08:12:41.232220 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:41 crc kubenswrapper[4860]: I1211 08:12:41.232259 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:41 crc kubenswrapper[4860]: I1211 08:12:41.232270 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:41 crc kubenswrapper[4860]: I1211 08:12:41.232284 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:41 crc kubenswrapper[4860]: I1211 08:12:41.232295 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:41Z","lastTransitionTime":"2025-12-11T08:12:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:41 crc kubenswrapper[4860]: I1211 08:12:41.334939 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:41 crc kubenswrapper[4860]: I1211 08:12:41.335016 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:41 crc kubenswrapper[4860]: I1211 08:12:41.335052 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:41 crc kubenswrapper[4860]: I1211 08:12:41.335122 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:41 crc kubenswrapper[4860]: I1211 08:12:41.335179 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:41Z","lastTransitionTime":"2025-12-11T08:12:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:41 crc kubenswrapper[4860]: I1211 08:12:41.438561 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:41 crc kubenswrapper[4860]: I1211 08:12:41.438622 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:41 crc kubenswrapper[4860]: I1211 08:12:41.438634 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:41 crc kubenswrapper[4860]: I1211 08:12:41.438679 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:41 crc kubenswrapper[4860]: I1211 08:12:41.438692 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:41Z","lastTransitionTime":"2025-12-11T08:12:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:41 crc kubenswrapper[4860]: I1211 08:12:41.541473 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:41 crc kubenswrapper[4860]: I1211 08:12:41.541543 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:41 crc kubenswrapper[4860]: I1211 08:12:41.541566 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:41 crc kubenswrapper[4860]: I1211 08:12:41.541597 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:41 crc kubenswrapper[4860]: I1211 08:12:41.541618 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:41Z","lastTransitionTime":"2025-12-11T08:12:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:41 crc kubenswrapper[4860]: I1211 08:12:41.644790 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:41 crc kubenswrapper[4860]: I1211 08:12:41.644839 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:41 crc kubenswrapper[4860]: I1211 08:12:41.644857 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:41 crc kubenswrapper[4860]: I1211 08:12:41.644880 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:41 crc kubenswrapper[4860]: I1211 08:12:41.644899 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:41Z","lastTransitionTime":"2025-12-11T08:12:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:41 crc kubenswrapper[4860]: I1211 08:12:41.748500 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:41 crc kubenswrapper[4860]: I1211 08:12:41.748715 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:41 crc kubenswrapper[4860]: I1211 08:12:41.748750 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:41 crc kubenswrapper[4860]: I1211 08:12:41.748777 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:41 crc kubenswrapper[4860]: I1211 08:12:41.748796 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:41Z","lastTransitionTime":"2025-12-11T08:12:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:41 crc kubenswrapper[4860]: I1211 08:12:41.851793 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:41 crc kubenswrapper[4860]: I1211 08:12:41.851841 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:41 crc kubenswrapper[4860]: I1211 08:12:41.851853 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:41 crc kubenswrapper[4860]: I1211 08:12:41.851872 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:41 crc kubenswrapper[4860]: I1211 08:12:41.851886 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:41Z","lastTransitionTime":"2025-12-11T08:12:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:41 crc kubenswrapper[4860]: I1211 08:12:41.954865 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:41 crc kubenswrapper[4860]: I1211 08:12:41.954940 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:41 crc kubenswrapper[4860]: I1211 08:12:41.954960 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:41 crc kubenswrapper[4860]: I1211 08:12:41.954985 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:41 crc kubenswrapper[4860]: I1211 08:12:41.955003 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:41Z","lastTransitionTime":"2025-12-11T08:12:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:42 crc kubenswrapper[4860]: I1211 08:12:42.059215 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:42 crc kubenswrapper[4860]: I1211 08:12:42.059284 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:42 crc kubenswrapper[4860]: I1211 08:12:42.059299 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:42 crc kubenswrapper[4860]: I1211 08:12:42.059358 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:42 crc kubenswrapper[4860]: I1211 08:12:42.059377 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:42Z","lastTransitionTime":"2025-12-11T08:12:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:42 crc kubenswrapper[4860]: I1211 08:12:42.162722 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:42 crc kubenswrapper[4860]: I1211 08:12:42.162775 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:42 crc kubenswrapper[4860]: I1211 08:12:42.162790 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:42 crc kubenswrapper[4860]: I1211 08:12:42.162807 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:42 crc kubenswrapper[4860]: I1211 08:12:42.162818 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:42Z","lastTransitionTime":"2025-12-11T08:12:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:42 crc kubenswrapper[4860]: I1211 08:12:42.265456 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:42 crc kubenswrapper[4860]: I1211 08:12:42.265518 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:42 crc kubenswrapper[4860]: I1211 08:12:42.265532 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:42 crc kubenswrapper[4860]: I1211 08:12:42.265550 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:42 crc kubenswrapper[4860]: I1211 08:12:42.265564 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:42Z","lastTransitionTime":"2025-12-11T08:12:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:42 crc kubenswrapper[4860]: I1211 08:12:42.369428 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:42 crc kubenswrapper[4860]: I1211 08:12:42.369471 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:42 crc kubenswrapper[4860]: I1211 08:12:42.369482 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:42 crc kubenswrapper[4860]: I1211 08:12:42.369496 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:42 crc kubenswrapper[4860]: I1211 08:12:42.369507 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:42Z","lastTransitionTime":"2025-12-11T08:12:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:42 crc kubenswrapper[4860]: I1211 08:12:42.472717 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:42 crc kubenswrapper[4860]: I1211 08:12:42.472764 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:42 crc kubenswrapper[4860]: I1211 08:12:42.472772 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:42 crc kubenswrapper[4860]: I1211 08:12:42.472787 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:42 crc kubenswrapper[4860]: I1211 08:12:42.472796 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:42Z","lastTransitionTime":"2025-12-11T08:12:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:42 crc kubenswrapper[4860]: I1211 08:12:42.564299 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:42 crc kubenswrapper[4860]: I1211 08:12:42.564378 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:42 crc kubenswrapper[4860]: I1211 08:12:42.564395 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:42 crc kubenswrapper[4860]: I1211 08:12:42.564413 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:42 crc kubenswrapper[4860]: I1211 08:12:42.564426 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:42Z","lastTransitionTime":"2025-12-11T08:12:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:42 crc kubenswrapper[4860]: I1211 08:12:42.578798 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gjhlj" Dec 11 08:12:42 crc kubenswrapper[4860]: E1211 08:12:42.579831 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gjhlj" podUID="5cfd640a-f3d2-4b28-87e1-5f49f6341e21" Dec 11 08:12:42 crc kubenswrapper[4860]: I1211 08:12:42.578908 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:12:42 crc kubenswrapper[4860]: E1211 08:12:42.579928 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:12:42 crc kubenswrapper[4860]: I1211 08:12:42.578984 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:12:42 crc kubenswrapper[4860]: E1211 08:12:42.579984 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:12:42 crc kubenswrapper[4860]: I1211 08:12:42.578882 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:12:42 crc kubenswrapper[4860]: E1211 08:12:42.580033 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:12:42 crc kubenswrapper[4860]: E1211 08:12:42.581392 4860 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:12:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:12:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:12:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:12:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"52631a1a-bc27-4dd5-8309-5dc39b5dcc41\\\",\\\"systemUUID\\\":\\\"299d3cd2-e1ad-4f5c-8af4-f12c0a5d76e7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:42Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:42 crc kubenswrapper[4860]: I1211 08:12:42.587111 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:42 crc kubenswrapper[4860]: I1211 08:12:42.587158 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:42 crc kubenswrapper[4860]: I1211 08:12:42.587171 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:42 crc kubenswrapper[4860]: I1211 08:12:42.587191 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:42 crc kubenswrapper[4860]: I1211 08:12:42.587206 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:42Z","lastTransitionTime":"2025-12-11T08:12:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:42 crc kubenswrapper[4860]: E1211 08:12:42.600999 4860 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:12:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:12:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:12:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:12:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"52631a1a-bc27-4dd5-8309-5dc39b5dcc41\\\",\\\"systemUUID\\\":\\\"299d3cd2-e1ad-4f5c-8af4-f12c0a5d76e7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:42Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:42 crc kubenswrapper[4860]: I1211 08:12:42.604382 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:42 crc kubenswrapper[4860]: I1211 08:12:42.604411 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:42 crc kubenswrapper[4860]: I1211 08:12:42.604422 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:42 crc kubenswrapper[4860]: I1211 08:12:42.604437 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:42 crc kubenswrapper[4860]: I1211 08:12:42.604448 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:42Z","lastTransitionTime":"2025-12-11T08:12:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:42 crc kubenswrapper[4860]: E1211 08:12:42.614959 4860 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:12:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:12:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:12:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:12:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"52631a1a-bc27-4dd5-8309-5dc39b5dcc41\\\",\\\"systemUUID\\\":\\\"299d3cd2-e1ad-4f5c-8af4-f12c0a5d76e7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:42Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:42 crc kubenswrapper[4860]: I1211 08:12:42.623653 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:42 crc kubenswrapper[4860]: I1211 08:12:42.623743 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:42 crc kubenswrapper[4860]: I1211 08:12:42.623760 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:42 crc kubenswrapper[4860]: I1211 08:12:42.623784 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:42 crc kubenswrapper[4860]: I1211 08:12:42.623800 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:42Z","lastTransitionTime":"2025-12-11T08:12:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:42 crc kubenswrapper[4860]: E1211 08:12:42.636265 4860 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:12:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:12:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:12:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:12:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"52631a1a-bc27-4dd5-8309-5dc39b5dcc41\\\",\\\"systemUUID\\\":\\\"299d3cd2-e1ad-4f5c-8af4-f12c0a5d76e7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:42Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:42 crc kubenswrapper[4860]: I1211 08:12:42.640113 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:42 crc kubenswrapper[4860]: I1211 08:12:42.640159 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:42 crc kubenswrapper[4860]: I1211 08:12:42.640168 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:42 crc kubenswrapper[4860]: I1211 08:12:42.640183 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:42 crc kubenswrapper[4860]: I1211 08:12:42.640193 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:42Z","lastTransitionTime":"2025-12-11T08:12:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:42 crc kubenswrapper[4860]: E1211 08:12:42.650790 4860 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:12:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:12:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:12:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:12:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"52631a1a-bc27-4dd5-8309-5dc39b5dcc41\\\",\\\"systemUUID\\\":\\\"299d3cd2-e1ad-4f5c-8af4-f12c0a5d76e7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:42Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:42 crc kubenswrapper[4860]: E1211 08:12:42.651000 4860 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 11 08:12:42 crc kubenswrapper[4860]: I1211 08:12:42.652612 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:42 crc kubenswrapper[4860]: I1211 08:12:42.652673 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:42 crc kubenswrapper[4860]: I1211 08:12:42.652686 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:42 crc kubenswrapper[4860]: I1211 08:12:42.652702 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:42 crc kubenswrapper[4860]: I1211 08:12:42.652713 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:42Z","lastTransitionTime":"2025-12-11T08:12:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:42 crc kubenswrapper[4860]: I1211 08:12:42.755132 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:42 crc kubenswrapper[4860]: I1211 08:12:42.755201 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:42 crc kubenswrapper[4860]: I1211 08:12:42.755221 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:42 crc kubenswrapper[4860]: I1211 08:12:42.755245 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:42 crc kubenswrapper[4860]: I1211 08:12:42.755262 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:42Z","lastTransitionTime":"2025-12-11T08:12:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:42 crc kubenswrapper[4860]: I1211 08:12:42.857586 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:42 crc kubenswrapper[4860]: I1211 08:12:42.857622 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:42 crc kubenswrapper[4860]: I1211 08:12:42.857632 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:42 crc kubenswrapper[4860]: I1211 08:12:42.857649 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:42 crc kubenswrapper[4860]: I1211 08:12:42.857671 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:42Z","lastTransitionTime":"2025-12-11T08:12:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:42 crc kubenswrapper[4860]: I1211 08:12:42.960926 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:42 crc kubenswrapper[4860]: I1211 08:12:42.960959 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:42 crc kubenswrapper[4860]: I1211 08:12:42.960968 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:42 crc kubenswrapper[4860]: I1211 08:12:42.960983 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:42 crc kubenswrapper[4860]: I1211 08:12:42.960992 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:42Z","lastTransitionTime":"2025-12-11T08:12:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:43 crc kubenswrapper[4860]: I1211 08:12:43.064356 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:43 crc kubenswrapper[4860]: I1211 08:12:43.064425 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:43 crc kubenswrapper[4860]: I1211 08:12:43.064455 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:43 crc kubenswrapper[4860]: I1211 08:12:43.064524 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:43 crc kubenswrapper[4860]: I1211 08:12:43.064547 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:43Z","lastTransitionTime":"2025-12-11T08:12:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:43 crc kubenswrapper[4860]: I1211 08:12:43.167656 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:43 crc kubenswrapper[4860]: I1211 08:12:43.168309 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:43 crc kubenswrapper[4860]: I1211 08:12:43.168324 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:43 crc kubenswrapper[4860]: I1211 08:12:43.168346 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:43 crc kubenswrapper[4860]: I1211 08:12:43.168358 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:43Z","lastTransitionTime":"2025-12-11T08:12:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:43 crc kubenswrapper[4860]: I1211 08:12:43.270786 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:43 crc kubenswrapper[4860]: I1211 08:12:43.270829 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:43 crc kubenswrapper[4860]: I1211 08:12:43.270842 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:43 crc kubenswrapper[4860]: I1211 08:12:43.270860 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:43 crc kubenswrapper[4860]: I1211 08:12:43.270873 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:43Z","lastTransitionTime":"2025-12-11T08:12:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:43 crc kubenswrapper[4860]: I1211 08:12:43.373313 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:43 crc kubenswrapper[4860]: I1211 08:12:43.373348 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:43 crc kubenswrapper[4860]: I1211 08:12:43.373360 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:43 crc kubenswrapper[4860]: I1211 08:12:43.373376 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:43 crc kubenswrapper[4860]: I1211 08:12:43.373387 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:43Z","lastTransitionTime":"2025-12-11T08:12:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:43 crc kubenswrapper[4860]: I1211 08:12:43.475023 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:43 crc kubenswrapper[4860]: I1211 08:12:43.475073 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:43 crc kubenswrapper[4860]: I1211 08:12:43.475084 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:43 crc kubenswrapper[4860]: I1211 08:12:43.475101 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:43 crc kubenswrapper[4860]: I1211 08:12:43.475113 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:43Z","lastTransitionTime":"2025-12-11T08:12:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:43 crc kubenswrapper[4860]: I1211 08:12:43.578261 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:43 crc kubenswrapper[4860]: I1211 08:12:43.578343 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:43 crc kubenswrapper[4860]: I1211 08:12:43.578369 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:43 crc kubenswrapper[4860]: I1211 08:12:43.578398 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:43 crc kubenswrapper[4860]: I1211 08:12:43.578422 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:43Z","lastTransitionTime":"2025-12-11T08:12:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:43 crc kubenswrapper[4860]: I1211 08:12:43.579011 4860 scope.go:117] "RemoveContainer" containerID="d9e83cda3e40d325604114a37fe6e8b707ac1f91a034b0d979c70b5f4f24f2e7" Dec 11 08:12:43 crc kubenswrapper[4860]: E1211 08:12:43.581240 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-xvnqp_openshift-ovn-kubernetes(0589a204-c98c-417d-8256-bfe3bf747660)\"" pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" podUID="0589a204-c98c-417d-8256-bfe3bf747660" Dec 11 08:12:43 crc kubenswrapper[4860]: I1211 08:12:43.681687 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:43 crc kubenswrapper[4860]: I1211 08:12:43.681727 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:43 crc kubenswrapper[4860]: I1211 08:12:43.681735 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:43 crc kubenswrapper[4860]: I1211 08:12:43.681750 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:43 crc kubenswrapper[4860]: I1211 08:12:43.681759 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:43Z","lastTransitionTime":"2025-12-11T08:12:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:43 crc kubenswrapper[4860]: I1211 08:12:43.784368 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:43 crc kubenswrapper[4860]: I1211 08:12:43.784428 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:43 crc kubenswrapper[4860]: I1211 08:12:43.784446 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:43 crc kubenswrapper[4860]: I1211 08:12:43.784468 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:43 crc kubenswrapper[4860]: I1211 08:12:43.784487 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:43Z","lastTransitionTime":"2025-12-11T08:12:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:43 crc kubenswrapper[4860]: I1211 08:12:43.887111 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:43 crc kubenswrapper[4860]: I1211 08:12:43.887153 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:43 crc kubenswrapper[4860]: I1211 08:12:43.887161 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:43 crc kubenswrapper[4860]: I1211 08:12:43.887176 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:43 crc kubenswrapper[4860]: I1211 08:12:43.887184 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:43Z","lastTransitionTime":"2025-12-11T08:12:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:43 crc kubenswrapper[4860]: I1211 08:12:43.989036 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:43 crc kubenswrapper[4860]: I1211 08:12:43.989073 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:43 crc kubenswrapper[4860]: I1211 08:12:43.989082 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:43 crc kubenswrapper[4860]: I1211 08:12:43.989098 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:43 crc kubenswrapper[4860]: I1211 08:12:43.989107 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:43Z","lastTransitionTime":"2025-12-11T08:12:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:44 crc kubenswrapper[4860]: I1211 08:12:44.091539 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:44 crc kubenswrapper[4860]: I1211 08:12:44.091603 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:44 crc kubenswrapper[4860]: I1211 08:12:44.091618 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:44 crc kubenswrapper[4860]: I1211 08:12:44.091645 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:44 crc kubenswrapper[4860]: I1211 08:12:44.091682 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:44Z","lastTransitionTime":"2025-12-11T08:12:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:44 crc kubenswrapper[4860]: I1211 08:12:44.194585 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:44 crc kubenswrapper[4860]: I1211 08:12:44.194622 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:44 crc kubenswrapper[4860]: I1211 08:12:44.194634 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:44 crc kubenswrapper[4860]: I1211 08:12:44.194691 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:44 crc kubenswrapper[4860]: I1211 08:12:44.194703 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:44Z","lastTransitionTime":"2025-12-11T08:12:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:44 crc kubenswrapper[4860]: I1211 08:12:44.297584 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:44 crc kubenswrapper[4860]: I1211 08:12:44.297664 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:44 crc kubenswrapper[4860]: I1211 08:12:44.297679 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:44 crc kubenswrapper[4860]: I1211 08:12:44.297694 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:44 crc kubenswrapper[4860]: I1211 08:12:44.297706 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:44Z","lastTransitionTime":"2025-12-11T08:12:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:44 crc kubenswrapper[4860]: I1211 08:12:44.400809 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:44 crc kubenswrapper[4860]: I1211 08:12:44.400885 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:44 crc kubenswrapper[4860]: I1211 08:12:44.400908 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:44 crc kubenswrapper[4860]: I1211 08:12:44.400936 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:44 crc kubenswrapper[4860]: I1211 08:12:44.400956 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:44Z","lastTransitionTime":"2025-12-11T08:12:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:44 crc kubenswrapper[4860]: I1211 08:12:44.503990 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:44 crc kubenswrapper[4860]: I1211 08:12:44.504078 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:44 crc kubenswrapper[4860]: I1211 08:12:44.504101 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:44 crc kubenswrapper[4860]: I1211 08:12:44.504140 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:44 crc kubenswrapper[4860]: I1211 08:12:44.504194 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:44Z","lastTransitionTime":"2025-12-11T08:12:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:44 crc kubenswrapper[4860]: I1211 08:12:44.578632 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:12:44 crc kubenswrapper[4860]: I1211 08:12:44.578732 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gjhlj" Dec 11 08:12:44 crc kubenswrapper[4860]: I1211 08:12:44.578782 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:12:44 crc kubenswrapper[4860]: E1211 08:12:44.578854 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:12:44 crc kubenswrapper[4860]: E1211 08:12:44.579030 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:12:44 crc kubenswrapper[4860]: E1211 08:12:44.579182 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gjhlj" podUID="5cfd640a-f3d2-4b28-87e1-5f49f6341e21" Dec 11 08:12:44 crc kubenswrapper[4860]: I1211 08:12:44.579595 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:12:44 crc kubenswrapper[4860]: E1211 08:12:44.580019 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:12:44 crc kubenswrapper[4860]: I1211 08:12:44.607562 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:44 crc kubenswrapper[4860]: I1211 08:12:44.607648 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:44 crc kubenswrapper[4860]: I1211 08:12:44.607706 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:44 crc kubenswrapper[4860]: I1211 08:12:44.607738 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:44 crc kubenswrapper[4860]: I1211 08:12:44.607761 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:44Z","lastTransitionTime":"2025-12-11T08:12:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:44 crc kubenswrapper[4860]: I1211 08:12:44.710707 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:44 crc kubenswrapper[4860]: I1211 08:12:44.710756 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:44 crc kubenswrapper[4860]: I1211 08:12:44.710773 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:44 crc kubenswrapper[4860]: I1211 08:12:44.710795 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:44 crc kubenswrapper[4860]: I1211 08:12:44.710812 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:44Z","lastTransitionTime":"2025-12-11T08:12:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:44 crc kubenswrapper[4860]: I1211 08:12:44.814065 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:44 crc kubenswrapper[4860]: I1211 08:12:44.814192 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:44 crc kubenswrapper[4860]: I1211 08:12:44.814210 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:44 crc kubenswrapper[4860]: I1211 08:12:44.814233 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:44 crc kubenswrapper[4860]: I1211 08:12:44.814250 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:44Z","lastTransitionTime":"2025-12-11T08:12:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:44 crc kubenswrapper[4860]: I1211 08:12:44.917247 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:44 crc kubenswrapper[4860]: I1211 08:12:44.917307 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:44 crc kubenswrapper[4860]: I1211 08:12:44.917318 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:44 crc kubenswrapper[4860]: I1211 08:12:44.917337 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:44 crc kubenswrapper[4860]: I1211 08:12:44.917349 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:44Z","lastTransitionTime":"2025-12-11T08:12:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:45 crc kubenswrapper[4860]: I1211 08:12:45.019852 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:45 crc kubenswrapper[4860]: I1211 08:12:45.019905 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:45 crc kubenswrapper[4860]: I1211 08:12:45.019921 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:45 crc kubenswrapper[4860]: I1211 08:12:45.019943 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:45 crc kubenswrapper[4860]: I1211 08:12:45.019954 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:45Z","lastTransitionTime":"2025-12-11T08:12:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:45 crc kubenswrapper[4860]: I1211 08:12:45.126382 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:45 crc kubenswrapper[4860]: I1211 08:12:45.126425 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:45 crc kubenswrapper[4860]: I1211 08:12:45.126437 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:45 crc kubenswrapper[4860]: I1211 08:12:45.126451 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:45 crc kubenswrapper[4860]: I1211 08:12:45.126462 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:45Z","lastTransitionTime":"2025-12-11T08:12:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:45 crc kubenswrapper[4860]: I1211 08:12:45.229677 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:45 crc kubenswrapper[4860]: I1211 08:12:45.229933 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:45 crc kubenswrapper[4860]: I1211 08:12:45.230022 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:45 crc kubenswrapper[4860]: I1211 08:12:45.230100 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:45 crc kubenswrapper[4860]: I1211 08:12:45.230173 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:45Z","lastTransitionTime":"2025-12-11T08:12:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:45 crc kubenswrapper[4860]: I1211 08:12:45.333567 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:45 crc kubenswrapper[4860]: I1211 08:12:45.333757 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:45 crc kubenswrapper[4860]: I1211 08:12:45.333789 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:45 crc kubenswrapper[4860]: I1211 08:12:45.333820 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:45 crc kubenswrapper[4860]: I1211 08:12:45.333839 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:45Z","lastTransitionTime":"2025-12-11T08:12:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:45 crc kubenswrapper[4860]: I1211 08:12:45.437553 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:45 crc kubenswrapper[4860]: I1211 08:12:45.437610 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:45 crc kubenswrapper[4860]: I1211 08:12:45.437620 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:45 crc kubenswrapper[4860]: I1211 08:12:45.437633 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:45 crc kubenswrapper[4860]: I1211 08:12:45.437657 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:45Z","lastTransitionTime":"2025-12-11T08:12:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:45 crc kubenswrapper[4860]: I1211 08:12:45.540408 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:45 crc kubenswrapper[4860]: I1211 08:12:45.540480 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:45 crc kubenswrapper[4860]: I1211 08:12:45.540491 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:45 crc kubenswrapper[4860]: I1211 08:12:45.540504 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:45 crc kubenswrapper[4860]: I1211 08:12:45.540512 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:45Z","lastTransitionTime":"2025-12-11T08:12:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:45 crc kubenswrapper[4860]: I1211 08:12:45.643311 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:45 crc kubenswrapper[4860]: I1211 08:12:45.643568 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:45 crc kubenswrapper[4860]: I1211 08:12:45.643668 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:45 crc kubenswrapper[4860]: I1211 08:12:45.643757 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:45 crc kubenswrapper[4860]: I1211 08:12:45.643849 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:45Z","lastTransitionTime":"2025-12-11T08:12:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:45 crc kubenswrapper[4860]: I1211 08:12:45.746466 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:45 crc kubenswrapper[4860]: I1211 08:12:45.746853 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:45 crc kubenswrapper[4860]: I1211 08:12:45.747060 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:45 crc kubenswrapper[4860]: I1211 08:12:45.747200 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:45 crc kubenswrapper[4860]: I1211 08:12:45.747485 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:45Z","lastTransitionTime":"2025-12-11T08:12:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:45 crc kubenswrapper[4860]: I1211 08:12:45.850177 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:45 crc kubenswrapper[4860]: I1211 08:12:45.850245 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:45 crc kubenswrapper[4860]: I1211 08:12:45.850268 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:45 crc kubenswrapper[4860]: I1211 08:12:45.850296 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:45 crc kubenswrapper[4860]: I1211 08:12:45.850318 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:45Z","lastTransitionTime":"2025-12-11T08:12:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:45 crc kubenswrapper[4860]: I1211 08:12:45.953247 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:45 crc kubenswrapper[4860]: I1211 08:12:45.953377 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:45 crc kubenswrapper[4860]: I1211 08:12:45.953414 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:45 crc kubenswrapper[4860]: I1211 08:12:45.953441 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:45 crc kubenswrapper[4860]: I1211 08:12:45.953461 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:45Z","lastTransitionTime":"2025-12-11T08:12:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:46 crc kubenswrapper[4860]: I1211 08:12:46.056031 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:46 crc kubenswrapper[4860]: I1211 08:12:46.056093 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:46 crc kubenswrapper[4860]: I1211 08:12:46.056108 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:46 crc kubenswrapper[4860]: I1211 08:12:46.056125 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:46 crc kubenswrapper[4860]: I1211 08:12:46.056137 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:46Z","lastTransitionTime":"2025-12-11T08:12:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:46 crc kubenswrapper[4860]: I1211 08:12:46.157931 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:46 crc kubenswrapper[4860]: I1211 08:12:46.158019 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:46 crc kubenswrapper[4860]: I1211 08:12:46.158038 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:46 crc kubenswrapper[4860]: I1211 08:12:46.158060 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:46 crc kubenswrapper[4860]: I1211 08:12:46.158108 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:46Z","lastTransitionTime":"2025-12-11T08:12:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:46 crc kubenswrapper[4860]: I1211 08:12:46.260933 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:46 crc kubenswrapper[4860]: I1211 08:12:46.260977 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:46 crc kubenswrapper[4860]: I1211 08:12:46.260990 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:46 crc kubenswrapper[4860]: I1211 08:12:46.261009 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:46 crc kubenswrapper[4860]: I1211 08:12:46.261022 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:46Z","lastTransitionTime":"2025-12-11T08:12:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:46 crc kubenswrapper[4860]: I1211 08:12:46.363636 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:46 crc kubenswrapper[4860]: I1211 08:12:46.363711 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:46 crc kubenswrapper[4860]: I1211 08:12:46.363723 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:46 crc kubenswrapper[4860]: I1211 08:12:46.363738 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:46 crc kubenswrapper[4860]: I1211 08:12:46.363755 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:46Z","lastTransitionTime":"2025-12-11T08:12:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:46 crc kubenswrapper[4860]: I1211 08:12:46.466098 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:46 crc kubenswrapper[4860]: I1211 08:12:46.466152 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:46 crc kubenswrapper[4860]: I1211 08:12:46.466163 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:46 crc kubenswrapper[4860]: I1211 08:12:46.466179 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:46 crc kubenswrapper[4860]: I1211 08:12:46.466188 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:46Z","lastTransitionTime":"2025-12-11T08:12:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:46 crc kubenswrapper[4860]: I1211 08:12:46.500201 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5cfd640a-f3d2-4b28-87e1-5f49f6341e21-metrics-certs\") pod \"network-metrics-daemon-gjhlj\" (UID: \"5cfd640a-f3d2-4b28-87e1-5f49f6341e21\") " pod="openshift-multus/network-metrics-daemon-gjhlj" Dec 11 08:12:46 crc kubenswrapper[4860]: E1211 08:12:46.500356 4860 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 11 08:12:46 crc kubenswrapper[4860]: E1211 08:12:46.500414 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5cfd640a-f3d2-4b28-87e1-5f49f6341e21-metrics-certs podName:5cfd640a-f3d2-4b28-87e1-5f49f6341e21 nodeName:}" failed. No retries permitted until 2025-12-11 08:13:50.500399155 +0000 UTC m=+163.228918210 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5cfd640a-f3d2-4b28-87e1-5f49f6341e21-metrics-certs") pod "network-metrics-daemon-gjhlj" (UID: "5cfd640a-f3d2-4b28-87e1-5f49f6341e21") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 11 08:12:46 crc kubenswrapper[4860]: I1211 08:12:46.568831 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:46 crc kubenswrapper[4860]: I1211 08:12:46.568912 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:46 crc kubenswrapper[4860]: I1211 08:12:46.568924 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:46 crc kubenswrapper[4860]: I1211 08:12:46.568974 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:46 crc kubenswrapper[4860]: I1211 08:12:46.568989 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:46Z","lastTransitionTime":"2025-12-11T08:12:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:46 crc kubenswrapper[4860]: I1211 08:12:46.578415 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:12:46 crc kubenswrapper[4860]: I1211 08:12:46.578515 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:12:46 crc kubenswrapper[4860]: I1211 08:12:46.578452 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gjhlj" Dec 11 08:12:46 crc kubenswrapper[4860]: I1211 08:12:46.578437 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:12:46 crc kubenswrapper[4860]: E1211 08:12:46.578745 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:12:46 crc kubenswrapper[4860]: E1211 08:12:46.578794 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:12:46 crc kubenswrapper[4860]: E1211 08:12:46.578882 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:12:46 crc kubenswrapper[4860]: E1211 08:12:46.578986 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gjhlj" podUID="5cfd640a-f3d2-4b28-87e1-5f49f6341e21" Dec 11 08:12:46 crc kubenswrapper[4860]: I1211 08:12:46.672720 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:46 crc kubenswrapper[4860]: I1211 08:12:46.672769 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:46 crc kubenswrapper[4860]: I1211 08:12:46.672823 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:46 crc kubenswrapper[4860]: I1211 08:12:46.672842 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:46 crc kubenswrapper[4860]: I1211 08:12:46.672856 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:46Z","lastTransitionTime":"2025-12-11T08:12:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:46 crc kubenswrapper[4860]: I1211 08:12:46.775521 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:46 crc kubenswrapper[4860]: I1211 08:12:46.775575 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:46 crc kubenswrapper[4860]: I1211 08:12:46.775587 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:46 crc kubenswrapper[4860]: I1211 08:12:46.775606 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:46 crc kubenswrapper[4860]: I1211 08:12:46.775621 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:46Z","lastTransitionTime":"2025-12-11T08:12:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:46 crc kubenswrapper[4860]: I1211 08:12:46.879305 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:46 crc kubenswrapper[4860]: I1211 08:12:46.879360 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:46 crc kubenswrapper[4860]: I1211 08:12:46.879377 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:46 crc kubenswrapper[4860]: I1211 08:12:46.879399 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:46 crc kubenswrapper[4860]: I1211 08:12:46.879411 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:46Z","lastTransitionTime":"2025-12-11T08:12:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:46 crc kubenswrapper[4860]: I1211 08:12:46.981886 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:46 crc kubenswrapper[4860]: I1211 08:12:46.981940 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:46 crc kubenswrapper[4860]: I1211 08:12:46.982035 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:46 crc kubenswrapper[4860]: I1211 08:12:46.982062 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:46 crc kubenswrapper[4860]: I1211 08:12:46.982078 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:46Z","lastTransitionTime":"2025-12-11T08:12:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:47 crc kubenswrapper[4860]: I1211 08:12:47.085671 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:47 crc kubenswrapper[4860]: I1211 08:12:47.086207 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:47 crc kubenswrapper[4860]: I1211 08:12:47.086553 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:47 crc kubenswrapper[4860]: I1211 08:12:47.086679 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:47 crc kubenswrapper[4860]: I1211 08:12:47.086766 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:47Z","lastTransitionTime":"2025-12-11T08:12:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:47 crc kubenswrapper[4860]: I1211 08:12:47.189854 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:47 crc kubenswrapper[4860]: I1211 08:12:47.190293 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:47 crc kubenswrapper[4860]: I1211 08:12:47.190671 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:47 crc kubenswrapper[4860]: I1211 08:12:47.190797 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:47 crc kubenswrapper[4860]: I1211 08:12:47.191117 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:47Z","lastTransitionTime":"2025-12-11T08:12:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:47 crc kubenswrapper[4860]: I1211 08:12:47.293843 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:47 crc kubenswrapper[4860]: I1211 08:12:47.294229 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:47 crc kubenswrapper[4860]: I1211 08:12:47.294297 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:47 crc kubenswrapper[4860]: I1211 08:12:47.294385 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:47 crc kubenswrapper[4860]: I1211 08:12:47.294457 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:47Z","lastTransitionTime":"2025-12-11T08:12:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:47 crc kubenswrapper[4860]: I1211 08:12:47.396938 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:47 crc kubenswrapper[4860]: I1211 08:12:47.397474 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:47 crc kubenswrapper[4860]: I1211 08:12:47.397594 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:47 crc kubenswrapper[4860]: I1211 08:12:47.397749 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:47 crc kubenswrapper[4860]: I1211 08:12:47.397898 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:47Z","lastTransitionTime":"2025-12-11T08:12:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:47 crc kubenswrapper[4860]: I1211 08:12:47.500514 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:47 crc kubenswrapper[4860]: I1211 08:12:47.500607 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:47 crc kubenswrapper[4860]: I1211 08:12:47.500626 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:47 crc kubenswrapper[4860]: I1211 08:12:47.500678 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:47 crc kubenswrapper[4860]: I1211 08:12:47.500697 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:47Z","lastTransitionTime":"2025-12-11T08:12:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:47 crc kubenswrapper[4860]: I1211 08:12:47.598708 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"910ec852-e5d2-4370-806d-77f16a36e630\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://122eb7e14a042027de3aafaa8db89562d80565e7df20f302f872e48d412a26b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1348b83ed7b1e8c25f9dfb1b2d2d035b5adc96dc23854ee90236f06fb155f155\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://867cdac386b0576ee2958a520afd522a2558a8878ea6b5af977b136dc7bba4a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4602fc6635e6188bf4cb75c5c7cf1e71a64fc0c5420626bfafd87beae150722d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4602fc6635e6188bf4cb75c5c7cf1e71a64fc0c5420626bfafd87beae150722d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:47Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:47 crc kubenswrapper[4860]: I1211 08:12:47.605842 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:47 crc kubenswrapper[4860]: I1211 08:12:47.605907 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:47 crc kubenswrapper[4860]: I1211 08:12:47.605929 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:47 crc kubenswrapper[4860]: I1211 08:12:47.605965 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:47 crc kubenswrapper[4860]: I1211 08:12:47.605986 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:47Z","lastTransitionTime":"2025-12-11T08:12:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:47 crc kubenswrapper[4860]: I1211 08:12:47.617945 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:47Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:47 crc kubenswrapper[4860]: I1211 08:12:47.633184 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d8c2f28628dc2f1ffbdd98a55670e22cf6cdc8cf2c6442e9d1f8e5466b2f12e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://074ea63beac52bd196507ee7cfb1c3d7fc2d37d5ec6faa4d47dde62b922e1c43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:47Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:47 crc kubenswrapper[4860]: I1211 08:12:47.646698 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:47Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:47 crc kubenswrapper[4860]: I1211 08:12:47.661912 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25a6a154da2b74952951ae94efc914ba078ea4bbe5893360b3c84d46be6bf973\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:47Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:47 crc kubenswrapper[4860]: I1211 08:12:47.674796 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6xq7j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a75ce3c-5e1e-4b9f-bbde-7a3b0cce5eae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe8d4d675567b41728e476351811a93c47b19868ccbadab8c613a4eb8eca8562\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5c6nb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6xq7j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:47Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:47 crc kubenswrapper[4860]: I1211 08:12:47.691108 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5ec3758-85f1-4a22-a4ec-c805472607a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e831db1a15976d3967d56479689572c26f2599a766baa0402897a3008428c56f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a498aa90f914e1f0370df01dc5ef174df113a2a16d86b2e3f24b19a1b99b4c0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8caebd2f4dfcbb89c597029a07c335171889761d2d200206987537fb0a228db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34e91f615b76445ff3241a304bb2964fde59de7e8a8dfff6841b2e628a31e86d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2090cab02ad74ddfd6bb57ff50eba35d18f56b0cde1c230fe011c01994b693bc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1211 08:11:21.171011 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1211 08:11:21.177674 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-644352688/tls.crt::/tmp/serving-cert-644352688/tls.key\\\\\\\"\\\\nI1211 08:11:26.508606 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 08:11:26.512741 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 08:11:26.512778 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 08:11:26.512839 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 08:11:26.512858 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 08:11:26.521684 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1211 08:11:26.521700 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1211 08:11:26.521722 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:11:26.521732 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 08:11:26.521738 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 08:11:26.521746 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 08:11:26.521749 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 08:11:26.521753 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1211 08:11:26.524999 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2274fb298ca7d2b37b060b4dbf9b166d39ba77a108b08edeeb4d67196e72cb3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf0aff8a6ab816b6617b1feeca9146d41bb5e8808bf1db96f8d493500c135d85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf0aff8a6ab816b6617b1feeca9146d41bb5e8808bf1db96f8d493500c135d85\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:47Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:47 crc kubenswrapper[4860]: I1211 08:12:47.709192 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:47Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:47 crc kubenswrapper[4860]: I1211 08:12:47.710366 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:47 crc kubenswrapper[4860]: I1211 08:12:47.710410 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:47 crc kubenswrapper[4860]: I1211 08:12:47.710425 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:47 crc kubenswrapper[4860]: I1211 08:12:47.710485 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:47 crc kubenswrapper[4860]: I1211 08:12:47.710500 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:47Z","lastTransitionTime":"2025-12-11T08:12:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:47 crc kubenswrapper[4860]: I1211 08:12:47.726457 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a41f35db087418d1be16953f10377042abd3f4488e5ac1bc29ca5dcfc00a37e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:47Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:47 crc kubenswrapper[4860]: I1211 08:12:47.739115 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gjhlj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5cfd640a-f3d2-4b28-87e1-5f49f6341e21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knq7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knq7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:42Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gjhlj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:47Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:47 crc kubenswrapper[4860]: I1211 08:12:47.753965 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8bf0e9df-4e9d-4f45-a5b9-0020fdab2ea6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://77c0c831f046149002cd12564aa9b1544a813da1fcd81e61451eeb504f0450b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3498078e25d1156a89ddc17f4823182b1aeb8b283f8b46bf66df67d67c5c5e44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3498078e25d1156a89ddc17f4823182b1aeb8b283f8b46bf66df67d67c5c5e44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:47Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:47 crc kubenswrapper[4860]: I1211 08:12:47.769161 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9aa0aed-62b3-4ea2-82ae-9a653c1101a5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ad4ba1b769c6cb2b34073798f8e3c25eb0177e6d3059484bf365e6981cab710\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://077aa1dd2fe5189ff60a0f2365e47a6a545cca0de7aa6379370908db4ea5838a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d7c919899b74a2b0fcd9221f3f769a9d013267354b200a282669d342d6413d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5483384f1d2ebc1fefa9706c9e67a3dc5a00ff643f803e3f2707c252e27ad94\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:47Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:47 crc kubenswrapper[4860]: I1211 08:12:47.783046 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31c30642-6e60-41b4-a477-0d802424e0aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79969402f5c33852a26fc839f37a45febf4f2377d77823284fbb01769e5585fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2d4d5de7106175087d91acb1f0c7552014f0c08bfcd65cbbf184d3f8d7826ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qc8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-99qgp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:47Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:47 crc kubenswrapper[4860]: I1211 08:12:47.798514 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-26nc6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1faa733-f408-41fd-b7aa-0cd86283976a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fce5b642b843ed30823c94e0b0c3d8bb1e5d2d93d691f7b8ef901ece03b0ac6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ac45db4f858a1c18a6b87472fc1cea2300f5845f1ad62c22260374995548a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ac45db4f858a1c18a6b87472fc1cea2300f5845f1ad62c22260374995548a0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b35296a5eb17dd2bb5e72c86c1cafdb30b7712cb330264f935d3b4fff05cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c6b35296a5eb17dd2bb5e72c86c1cafdb30b7712cb330264f935d3b4fff05cd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b9e82d6d3a247309ef5bdbf5b3a86ec083add2a138b0b237801e77b09ee96a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b9e82d6d3a247309ef5bdbf5b3a86ec083add2a138b0b237801e77b09ee96a1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://908ee695e78145ef4375c4097d9f1e772502922d02a2fcb4566172a6ad9caaf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://908ee695e78145ef4375c4097d9f1e772502922d02a2fcb4566172a6ad9caaf9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afb0a097f3556a8f9ed7cf8f762cb99c082fe0a36b8d8d3038bf188b08fe9a84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afb0a097f3556a8f9ed7cf8f762cb99c082fe0a36b8d8d3038bf188b08fe9a84\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce18f4f06db091e658684fa0bbc858e19b9f03010acf13b3ffccb1a6e3f88021\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce18f4f06db091e658684fa0bbc858e19b9f03010acf13b3ffccb1a6e3f88021\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9nf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-26nc6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:47Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:47 crc kubenswrapper[4860]: I1211 08:12:47.813464 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:47 crc kubenswrapper[4860]: I1211 08:12:47.813521 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:47 crc kubenswrapper[4860]: I1211 08:12:47.813537 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:47 crc kubenswrapper[4860]: I1211 08:12:47.813560 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:47 crc kubenswrapper[4860]: I1211 08:12:47.813577 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:47Z","lastTransitionTime":"2025-12-11T08:12:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:47 crc kubenswrapper[4860]: I1211 08:12:47.822438 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0589a204-c98c-417d-8256-bfe3bf747660\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3333b4346f59b04017e32d61f457ba7ff457166731128f339d7987ae6b56b06d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52bb8ade314afb9494f4e026d0af2aea7b1970d64dc66102f55e257ae4678154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dfcc4370edaae36a5a32b0a9c29290c50f8e4f427e4e147dc3c779beaa8b780\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d856ae95e9077524c518d82d10fb6b93b7fbcaaf2b720ffafe6f7fae58bd7caa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1426084ac580895c42354630910dd8c8e3a48ddd25f22abce1d28dc3de332a63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9984d5d242bb258405efe91fc9caf9e7ea592834445df2e78add9553a921c76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9e83cda3e40d325604114a37fe6e8b707ac1f91a034b0d979c70b5f4f24f2e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9e83cda3e40d325604114a37fe6e8b707ac1f91a034b0d979c70b5f4f24f2e7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-11T08:12:30Z\\\",\\\"message\\\":\\\"val\\\\nI1211 08:12:30.086873 6903 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1211 08:12:30.086964 6903 handler.go:208] Removed *v1.Node event handler 7\\\\nI1211 08:12:30.086994 6903 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1211 08:12:30.087035 6903 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1211 08:12:30.087066 6903 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1211 08:12:30.087141 6903 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1211 08:12:30.087543 6903 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1211 08:12:30.087550 6903 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1211 08:12:30.087568 6903 handler.go:208] Removed *v1.Node event handler 2\\\\nI1211 08:12:30.087967 6903 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1211 08:12:30.087990 6903 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1211 08:12:30.088004 6903 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1211 08:12:30.088018 6903 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1211 08:12:30.088032 6903 factory.go:656] Stopping watch factory\\\\nI1211 08:12:30.088038 6903 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1211 08:12:30.088050 6903 ovnkube.go:599] Stopped ovnkube\\\\nI1211 08:12:3\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:12:29Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-xvnqp_openshift-ovn-kubernetes(0589a204-c98c-417d-8256-bfe3bf747660)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb80d9928e8349fc5963227d135355ad7ba820d4fd89877af601d005ca03334b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17cbcb49c06f3a9059cd5a0f05eedb8c785da30a3e502e31bae2f57b3486cb14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17cbcb49c06f3a9059cd5a0f05eedb8c785da30a3e502e31bae2f57b3486cb14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tftsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xvnqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:47Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:47 crc kubenswrapper[4860]: I1211 08:12:47.837885 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-gsx9l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aba5b6b9-f88d-4763-9562-6b4dec606df9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://735ef012c523ddc952d46a431f27b2582a2c301d9e9da0c697a20ff6e2fb5d86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6zw9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:30Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-gsx9l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:47Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:47 crc kubenswrapper[4860]: I1211 08:12:47.859086 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9872be1e-b995-4a00-904b-875dff96ad56\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f74a32a8fefea704cc73ffd1114083dab68aa44dc33f8d9d432d2d24216546c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbe00b46063d3968d59c393b87178e135f982abb7d0f32fa798b212d4b83df6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eedc9988e1a4a3c8cae730dc4a352103a6d665535b825e157d0f87c331b137c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2137fee4b7abc9765f9517d7b3b6ea796601a31d03151bf167d17136ced8b25a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f683e48f3577a68862eb6302383538376e41f49b20b1e0ea3f1b3b143f24872a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5de794187456b3576b969069a3b839d75c7918418a97fa35dff8bb40bb76e39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5de794187456b3576b969069a3b839d75c7918418a97fa35dff8bb40bb76e39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08ce42b974fba00c409c0af971f0de351e50f18b2f06f8e722150da3239bb908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08ce42b974fba00c409c0af971f0de351e50f18b2f06f8e722150da3239bb908\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cfac6be03e3cfa0b0f3450b48b67957f2d5172b9c65e27f14c2a260bda5d423f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cfac6be03e3cfa0b0f3450b48b67957f2d5172b9c65e27f14c2a260bda5d423f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T08:11:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:47Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:47 crc kubenswrapper[4860]: I1211 08:12:47.877364 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-pznwl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5310c1fc-66c6-40aa-b1bf-5a59a2410f9c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:12:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34119a6367261d7e12174f914ee8cfaa55703a55210f83c58b0e4694c54a5b03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2d6ff00ce322a28607bc3f6db05c5e4fb14acb6777ea315786ef53963ca504a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-11T08:12:17Z\\\",\\\"message\\\":\\\"2025-12-11T08:11:31+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_f37a703f-1609-4581-ab99-50337287ff33\\\\n2025-12-11T08:11:31+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_f37a703f-1609-4581-ab99-50337287ff33 to /host/opt/cni/bin/\\\\n2025-12-11T08:11:32Z [verbose] multus-daemon started\\\\n2025-12-11T08:11:32Z [verbose] Readiness Indicator file check\\\\n2025-12-11T08:12:17Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T08:11:29Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:12:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hkt48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"multus-pznwl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:47Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:47 crc kubenswrapper[4860]: I1211 08:12:47.888317 4860 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-854v6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e6529ba7-2341-43ac-a187-95ab530fe775\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T08:11:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8986538a421f0184f52f288b0e171da695591caf2b10c67ad2dce1073a6e6016\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kh22b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0d4e3f761ad15ae1424204e40b9d3f60188ab4a9047bd4b5ac6a8a8ca1fb385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T08:11:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kh22b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T08:11:41Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-854v6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-11T08:12:47Z is after 2025-08-24T17:21:41Z" Dec 11 08:12:47 crc kubenswrapper[4860]: I1211 08:12:47.917324 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:47 crc kubenswrapper[4860]: I1211 08:12:47.917383 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:47 crc kubenswrapper[4860]: I1211 08:12:47.917399 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:47 crc kubenswrapper[4860]: I1211 08:12:47.917425 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:47 crc kubenswrapper[4860]: I1211 08:12:47.917445 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:47Z","lastTransitionTime":"2025-12-11T08:12:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:48 crc kubenswrapper[4860]: I1211 08:12:48.020350 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:48 crc kubenswrapper[4860]: I1211 08:12:48.020406 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:48 crc kubenswrapper[4860]: I1211 08:12:48.020417 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:48 crc kubenswrapper[4860]: I1211 08:12:48.020432 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:48 crc kubenswrapper[4860]: I1211 08:12:48.020441 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:48Z","lastTransitionTime":"2025-12-11T08:12:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:48 crc kubenswrapper[4860]: I1211 08:12:48.124033 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:48 crc kubenswrapper[4860]: I1211 08:12:48.124069 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:48 crc kubenswrapper[4860]: I1211 08:12:48.124079 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:48 crc kubenswrapper[4860]: I1211 08:12:48.124093 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:48 crc kubenswrapper[4860]: I1211 08:12:48.124101 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:48Z","lastTransitionTime":"2025-12-11T08:12:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:48 crc kubenswrapper[4860]: I1211 08:12:48.226152 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:48 crc kubenswrapper[4860]: I1211 08:12:48.226199 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:48 crc kubenswrapper[4860]: I1211 08:12:48.226216 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:48 crc kubenswrapper[4860]: I1211 08:12:48.226235 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:48 crc kubenswrapper[4860]: I1211 08:12:48.226246 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:48Z","lastTransitionTime":"2025-12-11T08:12:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:48 crc kubenswrapper[4860]: I1211 08:12:48.329223 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:48 crc kubenswrapper[4860]: I1211 08:12:48.329275 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:48 crc kubenswrapper[4860]: I1211 08:12:48.329288 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:48 crc kubenswrapper[4860]: I1211 08:12:48.329306 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:48 crc kubenswrapper[4860]: I1211 08:12:48.329318 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:48Z","lastTransitionTime":"2025-12-11T08:12:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:48 crc kubenswrapper[4860]: I1211 08:12:48.431484 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:48 crc kubenswrapper[4860]: I1211 08:12:48.431613 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:48 crc kubenswrapper[4860]: I1211 08:12:48.431634 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:48 crc kubenswrapper[4860]: I1211 08:12:48.431709 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:48 crc kubenswrapper[4860]: I1211 08:12:48.431735 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:48Z","lastTransitionTime":"2025-12-11T08:12:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:48 crc kubenswrapper[4860]: I1211 08:12:48.534544 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:48 crc kubenswrapper[4860]: I1211 08:12:48.534591 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:48 crc kubenswrapper[4860]: I1211 08:12:48.534603 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:48 crc kubenswrapper[4860]: I1211 08:12:48.534618 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:48 crc kubenswrapper[4860]: I1211 08:12:48.534630 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:48Z","lastTransitionTime":"2025-12-11T08:12:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:48 crc kubenswrapper[4860]: I1211 08:12:48.578464 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:12:48 crc kubenswrapper[4860]: I1211 08:12:48.578465 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:12:48 crc kubenswrapper[4860]: I1211 08:12:48.578535 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:12:48 crc kubenswrapper[4860]: I1211 08:12:48.578954 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gjhlj" Dec 11 08:12:48 crc kubenswrapper[4860]: E1211 08:12:48.579138 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:12:48 crc kubenswrapper[4860]: E1211 08:12:48.579513 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gjhlj" podUID="5cfd640a-f3d2-4b28-87e1-5f49f6341e21" Dec 11 08:12:48 crc kubenswrapper[4860]: E1211 08:12:48.579768 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:12:48 crc kubenswrapper[4860]: E1211 08:12:48.580257 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:12:48 crc kubenswrapper[4860]: I1211 08:12:48.638374 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:48 crc kubenswrapper[4860]: I1211 08:12:48.638442 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:48 crc kubenswrapper[4860]: I1211 08:12:48.638460 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:48 crc kubenswrapper[4860]: I1211 08:12:48.638486 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:48 crc kubenswrapper[4860]: I1211 08:12:48.638507 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:48Z","lastTransitionTime":"2025-12-11T08:12:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:48 crc kubenswrapper[4860]: I1211 08:12:48.741796 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:48 crc kubenswrapper[4860]: I1211 08:12:48.741875 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:48 crc kubenswrapper[4860]: I1211 08:12:48.741898 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:48 crc kubenswrapper[4860]: I1211 08:12:48.741932 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:48 crc kubenswrapper[4860]: I1211 08:12:48.741955 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:48Z","lastTransitionTime":"2025-12-11T08:12:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:48 crc kubenswrapper[4860]: I1211 08:12:48.844895 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:48 crc kubenswrapper[4860]: I1211 08:12:48.845238 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:48 crc kubenswrapper[4860]: I1211 08:12:48.845350 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:48 crc kubenswrapper[4860]: I1211 08:12:48.845511 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:48 crc kubenswrapper[4860]: I1211 08:12:48.845608 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:48Z","lastTransitionTime":"2025-12-11T08:12:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:48 crc kubenswrapper[4860]: I1211 08:12:48.947936 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:48 crc kubenswrapper[4860]: I1211 08:12:48.947992 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:48 crc kubenswrapper[4860]: I1211 08:12:48.948009 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:48 crc kubenswrapper[4860]: I1211 08:12:48.948033 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:48 crc kubenswrapper[4860]: I1211 08:12:48.948049 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:48Z","lastTransitionTime":"2025-12-11T08:12:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:49 crc kubenswrapper[4860]: I1211 08:12:49.051069 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:49 crc kubenswrapper[4860]: I1211 08:12:49.051434 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:49 crc kubenswrapper[4860]: I1211 08:12:49.051625 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:49 crc kubenswrapper[4860]: I1211 08:12:49.051832 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:49 crc kubenswrapper[4860]: I1211 08:12:49.051969 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:49Z","lastTransitionTime":"2025-12-11T08:12:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:49 crc kubenswrapper[4860]: I1211 08:12:49.155155 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:49 crc kubenswrapper[4860]: I1211 08:12:49.155218 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:49 crc kubenswrapper[4860]: I1211 08:12:49.155272 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:49 crc kubenswrapper[4860]: I1211 08:12:49.155301 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:49 crc kubenswrapper[4860]: I1211 08:12:49.155322 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:49Z","lastTransitionTime":"2025-12-11T08:12:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:49 crc kubenswrapper[4860]: I1211 08:12:49.257736 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:49 crc kubenswrapper[4860]: I1211 08:12:49.257800 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:49 crc kubenswrapper[4860]: I1211 08:12:49.257818 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:49 crc kubenswrapper[4860]: I1211 08:12:49.257842 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:49 crc kubenswrapper[4860]: I1211 08:12:49.257859 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:49Z","lastTransitionTime":"2025-12-11T08:12:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:49 crc kubenswrapper[4860]: I1211 08:12:49.360925 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:49 crc kubenswrapper[4860]: I1211 08:12:49.360984 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:49 crc kubenswrapper[4860]: I1211 08:12:49.361003 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:49 crc kubenswrapper[4860]: I1211 08:12:49.361026 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:49 crc kubenswrapper[4860]: I1211 08:12:49.361045 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:49Z","lastTransitionTime":"2025-12-11T08:12:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:49 crc kubenswrapper[4860]: I1211 08:12:49.463056 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:49 crc kubenswrapper[4860]: I1211 08:12:49.463097 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:49 crc kubenswrapper[4860]: I1211 08:12:49.463111 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:49 crc kubenswrapper[4860]: I1211 08:12:49.463135 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:49 crc kubenswrapper[4860]: I1211 08:12:49.463149 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:49Z","lastTransitionTime":"2025-12-11T08:12:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:49 crc kubenswrapper[4860]: I1211 08:12:49.565571 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:49 crc kubenswrapper[4860]: I1211 08:12:49.565620 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:49 crc kubenswrapper[4860]: I1211 08:12:49.565633 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:49 crc kubenswrapper[4860]: I1211 08:12:49.565676 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:49 crc kubenswrapper[4860]: I1211 08:12:49.565693 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:49Z","lastTransitionTime":"2025-12-11T08:12:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:49 crc kubenswrapper[4860]: I1211 08:12:49.669377 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:49 crc kubenswrapper[4860]: I1211 08:12:49.669422 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:49 crc kubenswrapper[4860]: I1211 08:12:49.669433 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:49 crc kubenswrapper[4860]: I1211 08:12:49.669452 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:49 crc kubenswrapper[4860]: I1211 08:12:49.669465 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:49Z","lastTransitionTime":"2025-12-11T08:12:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:49 crc kubenswrapper[4860]: I1211 08:12:49.772076 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:49 crc kubenswrapper[4860]: I1211 08:12:49.772501 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:49 crc kubenswrapper[4860]: I1211 08:12:49.772690 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:49 crc kubenswrapper[4860]: I1211 08:12:49.772877 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:49 crc kubenswrapper[4860]: I1211 08:12:49.773075 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:49Z","lastTransitionTime":"2025-12-11T08:12:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:49 crc kubenswrapper[4860]: I1211 08:12:49.876075 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:49 crc kubenswrapper[4860]: I1211 08:12:49.876149 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:49 crc kubenswrapper[4860]: I1211 08:12:49.876172 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:49 crc kubenswrapper[4860]: I1211 08:12:49.876204 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:49 crc kubenswrapper[4860]: I1211 08:12:49.876226 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:49Z","lastTransitionTime":"2025-12-11T08:12:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:49 crc kubenswrapper[4860]: I1211 08:12:49.978971 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:49 crc kubenswrapper[4860]: I1211 08:12:49.979020 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:49 crc kubenswrapper[4860]: I1211 08:12:49.979032 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:49 crc kubenswrapper[4860]: I1211 08:12:49.979047 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:49 crc kubenswrapper[4860]: I1211 08:12:49.979058 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:49Z","lastTransitionTime":"2025-12-11T08:12:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:50 crc kubenswrapper[4860]: I1211 08:12:50.081880 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:50 crc kubenswrapper[4860]: I1211 08:12:50.081917 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:50 crc kubenswrapper[4860]: I1211 08:12:50.081929 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:50 crc kubenswrapper[4860]: I1211 08:12:50.081943 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:50 crc kubenswrapper[4860]: I1211 08:12:50.081956 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:50Z","lastTransitionTime":"2025-12-11T08:12:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:50 crc kubenswrapper[4860]: I1211 08:12:50.184725 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:50 crc kubenswrapper[4860]: I1211 08:12:50.185008 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:50 crc kubenswrapper[4860]: I1211 08:12:50.185096 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:50 crc kubenswrapper[4860]: I1211 08:12:50.185201 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:50 crc kubenswrapper[4860]: I1211 08:12:50.185294 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:50Z","lastTransitionTime":"2025-12-11T08:12:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:50 crc kubenswrapper[4860]: I1211 08:12:50.288555 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:50 crc kubenswrapper[4860]: I1211 08:12:50.288910 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:50 crc kubenswrapper[4860]: I1211 08:12:50.289187 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:50 crc kubenswrapper[4860]: I1211 08:12:50.289466 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:50 crc kubenswrapper[4860]: I1211 08:12:50.289678 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:50Z","lastTransitionTime":"2025-12-11T08:12:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:50 crc kubenswrapper[4860]: I1211 08:12:50.392341 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:50 crc kubenswrapper[4860]: I1211 08:12:50.392390 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:50 crc kubenswrapper[4860]: I1211 08:12:50.392403 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:50 crc kubenswrapper[4860]: I1211 08:12:50.392422 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:50 crc kubenswrapper[4860]: I1211 08:12:50.392434 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:50Z","lastTransitionTime":"2025-12-11T08:12:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:50 crc kubenswrapper[4860]: I1211 08:12:50.494976 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:50 crc kubenswrapper[4860]: I1211 08:12:50.495031 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:50 crc kubenswrapper[4860]: I1211 08:12:50.495044 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:50 crc kubenswrapper[4860]: I1211 08:12:50.495061 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:50 crc kubenswrapper[4860]: I1211 08:12:50.495074 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:50Z","lastTransitionTime":"2025-12-11T08:12:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:50 crc kubenswrapper[4860]: I1211 08:12:50.578721 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:12:50 crc kubenswrapper[4860]: E1211 08:12:50.578924 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:12:50 crc kubenswrapper[4860]: I1211 08:12:50.578977 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:12:50 crc kubenswrapper[4860]: E1211 08:12:50.579157 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:12:50 crc kubenswrapper[4860]: I1211 08:12:50.579303 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gjhlj" Dec 11 08:12:50 crc kubenswrapper[4860]: I1211 08:12:50.579494 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:12:50 crc kubenswrapper[4860]: E1211 08:12:50.579593 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gjhlj" podUID="5cfd640a-f3d2-4b28-87e1-5f49f6341e21" Dec 11 08:12:50 crc kubenswrapper[4860]: E1211 08:12:50.580050 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:12:50 crc kubenswrapper[4860]: I1211 08:12:50.597887 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:50 crc kubenswrapper[4860]: I1211 08:12:50.597998 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:50 crc kubenswrapper[4860]: I1211 08:12:50.598012 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:50 crc kubenswrapper[4860]: I1211 08:12:50.598026 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:50 crc kubenswrapper[4860]: I1211 08:12:50.598037 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:50Z","lastTransitionTime":"2025-12-11T08:12:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:50 crc kubenswrapper[4860]: I1211 08:12:50.700999 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:50 crc kubenswrapper[4860]: I1211 08:12:50.701072 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:50 crc kubenswrapper[4860]: I1211 08:12:50.701094 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:50 crc kubenswrapper[4860]: I1211 08:12:50.701121 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:50 crc kubenswrapper[4860]: I1211 08:12:50.701141 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:50Z","lastTransitionTime":"2025-12-11T08:12:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:50 crc kubenswrapper[4860]: I1211 08:12:50.803354 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:50 crc kubenswrapper[4860]: I1211 08:12:50.803723 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:50 crc kubenswrapper[4860]: I1211 08:12:50.804055 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:50 crc kubenswrapper[4860]: I1211 08:12:50.804250 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:50 crc kubenswrapper[4860]: I1211 08:12:50.804439 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:50Z","lastTransitionTime":"2025-12-11T08:12:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:50 crc kubenswrapper[4860]: I1211 08:12:50.907819 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:50 crc kubenswrapper[4860]: I1211 08:12:50.907868 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:50 crc kubenswrapper[4860]: I1211 08:12:50.907885 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:50 crc kubenswrapper[4860]: I1211 08:12:50.907909 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:50 crc kubenswrapper[4860]: I1211 08:12:50.907926 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:50Z","lastTransitionTime":"2025-12-11T08:12:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:51 crc kubenswrapper[4860]: I1211 08:12:51.011056 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:51 crc kubenswrapper[4860]: I1211 08:12:51.011462 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:51 crc kubenswrapper[4860]: I1211 08:12:51.011613 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:51 crc kubenswrapper[4860]: I1211 08:12:51.011812 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:51 crc kubenswrapper[4860]: I1211 08:12:51.011950 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:51Z","lastTransitionTime":"2025-12-11T08:12:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:51 crc kubenswrapper[4860]: I1211 08:12:51.115029 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:51 crc kubenswrapper[4860]: I1211 08:12:51.115409 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:51 crc kubenswrapper[4860]: I1211 08:12:51.115712 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:51 crc kubenswrapper[4860]: I1211 08:12:51.115934 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:51 crc kubenswrapper[4860]: I1211 08:12:51.116121 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:51Z","lastTransitionTime":"2025-12-11T08:12:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:51 crc kubenswrapper[4860]: I1211 08:12:51.218915 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:51 crc kubenswrapper[4860]: I1211 08:12:51.218995 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:51 crc kubenswrapper[4860]: I1211 08:12:51.219007 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:51 crc kubenswrapper[4860]: I1211 08:12:51.219029 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:51 crc kubenswrapper[4860]: I1211 08:12:51.219042 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:51Z","lastTransitionTime":"2025-12-11T08:12:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:51 crc kubenswrapper[4860]: I1211 08:12:51.321898 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:51 crc kubenswrapper[4860]: I1211 08:12:51.322206 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:51 crc kubenswrapper[4860]: I1211 08:12:51.322290 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:51 crc kubenswrapper[4860]: I1211 08:12:51.322392 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:51 crc kubenswrapper[4860]: I1211 08:12:51.322541 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:51Z","lastTransitionTime":"2025-12-11T08:12:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:51 crc kubenswrapper[4860]: I1211 08:12:51.426291 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:51 crc kubenswrapper[4860]: I1211 08:12:51.426546 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:51 crc kubenswrapper[4860]: I1211 08:12:51.426636 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:51 crc kubenswrapper[4860]: I1211 08:12:51.426768 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:51 crc kubenswrapper[4860]: I1211 08:12:51.426869 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:51Z","lastTransitionTime":"2025-12-11T08:12:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:51 crc kubenswrapper[4860]: I1211 08:12:51.529484 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:51 crc kubenswrapper[4860]: I1211 08:12:51.529539 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:51 crc kubenswrapper[4860]: I1211 08:12:51.529549 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:51 crc kubenswrapper[4860]: I1211 08:12:51.529567 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:51 crc kubenswrapper[4860]: I1211 08:12:51.529583 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:51Z","lastTransitionTime":"2025-12-11T08:12:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:51 crc kubenswrapper[4860]: I1211 08:12:51.632463 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:51 crc kubenswrapper[4860]: I1211 08:12:51.632538 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:51 crc kubenswrapper[4860]: I1211 08:12:51.632555 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:51 crc kubenswrapper[4860]: I1211 08:12:51.632585 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:51 crc kubenswrapper[4860]: I1211 08:12:51.632605 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:51Z","lastTransitionTime":"2025-12-11T08:12:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:51 crc kubenswrapper[4860]: I1211 08:12:51.734908 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:51 crc kubenswrapper[4860]: I1211 08:12:51.734980 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:51 crc kubenswrapper[4860]: I1211 08:12:51.734999 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:51 crc kubenswrapper[4860]: I1211 08:12:51.735025 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:51 crc kubenswrapper[4860]: I1211 08:12:51.735042 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:51Z","lastTransitionTime":"2025-12-11T08:12:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:51 crc kubenswrapper[4860]: I1211 08:12:51.838009 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:51 crc kubenswrapper[4860]: I1211 08:12:51.838226 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:51 crc kubenswrapper[4860]: I1211 08:12:51.838330 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:51 crc kubenswrapper[4860]: I1211 08:12:51.838429 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:51 crc kubenswrapper[4860]: I1211 08:12:51.838508 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:51Z","lastTransitionTime":"2025-12-11T08:12:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:51 crc kubenswrapper[4860]: I1211 08:12:51.941212 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:51 crc kubenswrapper[4860]: I1211 08:12:51.941263 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:51 crc kubenswrapper[4860]: I1211 08:12:51.941274 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:51 crc kubenswrapper[4860]: I1211 08:12:51.941293 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:51 crc kubenswrapper[4860]: I1211 08:12:51.941306 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:51Z","lastTransitionTime":"2025-12-11T08:12:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:52 crc kubenswrapper[4860]: I1211 08:12:52.043768 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:52 crc kubenswrapper[4860]: I1211 08:12:52.043864 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:52 crc kubenswrapper[4860]: I1211 08:12:52.043885 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:52 crc kubenswrapper[4860]: I1211 08:12:52.043913 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:52 crc kubenswrapper[4860]: I1211 08:12:52.043934 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:52Z","lastTransitionTime":"2025-12-11T08:12:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:52 crc kubenswrapper[4860]: I1211 08:12:52.147246 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:52 crc kubenswrapper[4860]: I1211 08:12:52.147295 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:52 crc kubenswrapper[4860]: I1211 08:12:52.147309 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:52 crc kubenswrapper[4860]: I1211 08:12:52.147329 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:52 crc kubenswrapper[4860]: I1211 08:12:52.147342 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:52Z","lastTransitionTime":"2025-12-11T08:12:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:52 crc kubenswrapper[4860]: I1211 08:12:52.250004 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:52 crc kubenswrapper[4860]: I1211 08:12:52.250056 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:52 crc kubenswrapper[4860]: I1211 08:12:52.250070 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:52 crc kubenswrapper[4860]: I1211 08:12:52.250086 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:52 crc kubenswrapper[4860]: I1211 08:12:52.250100 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:52Z","lastTransitionTime":"2025-12-11T08:12:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:52 crc kubenswrapper[4860]: I1211 08:12:52.353323 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:52 crc kubenswrapper[4860]: I1211 08:12:52.353395 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:52 crc kubenswrapper[4860]: I1211 08:12:52.353415 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:52 crc kubenswrapper[4860]: I1211 08:12:52.353438 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:52 crc kubenswrapper[4860]: I1211 08:12:52.353456 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:52Z","lastTransitionTime":"2025-12-11T08:12:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:52 crc kubenswrapper[4860]: I1211 08:12:52.455981 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:52 crc kubenswrapper[4860]: I1211 08:12:52.456069 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:52 crc kubenswrapper[4860]: I1211 08:12:52.456088 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:52 crc kubenswrapper[4860]: I1211 08:12:52.456113 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:52 crc kubenswrapper[4860]: I1211 08:12:52.456132 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:52Z","lastTransitionTime":"2025-12-11T08:12:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:52 crc kubenswrapper[4860]: I1211 08:12:52.559507 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:52 crc kubenswrapper[4860]: I1211 08:12:52.559558 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:52 crc kubenswrapper[4860]: I1211 08:12:52.559568 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:52 crc kubenswrapper[4860]: I1211 08:12:52.559585 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:52 crc kubenswrapper[4860]: I1211 08:12:52.559596 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:52Z","lastTransitionTime":"2025-12-11T08:12:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:52 crc kubenswrapper[4860]: I1211 08:12:52.578622 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:12:52 crc kubenswrapper[4860]: I1211 08:12:52.578696 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:12:52 crc kubenswrapper[4860]: I1211 08:12:52.578637 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gjhlj" Dec 11 08:12:52 crc kubenswrapper[4860]: E1211 08:12:52.578863 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:12:52 crc kubenswrapper[4860]: I1211 08:12:52.578661 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:12:52 crc kubenswrapper[4860]: E1211 08:12:52.579021 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:12:52 crc kubenswrapper[4860]: E1211 08:12:52.579112 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:12:52 crc kubenswrapper[4860]: E1211 08:12:52.579188 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gjhlj" podUID="5cfd640a-f3d2-4b28-87e1-5f49f6341e21" Dec 11 08:12:52 crc kubenswrapper[4860]: I1211 08:12:52.662583 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:52 crc kubenswrapper[4860]: I1211 08:12:52.662704 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:52 crc kubenswrapper[4860]: I1211 08:12:52.662729 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:52 crc kubenswrapper[4860]: I1211 08:12:52.662773 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:52 crc kubenswrapper[4860]: I1211 08:12:52.662794 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:52Z","lastTransitionTime":"2025-12-11T08:12:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:52 crc kubenswrapper[4860]: I1211 08:12:52.765225 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:52 crc kubenswrapper[4860]: I1211 08:12:52.765305 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:52 crc kubenswrapper[4860]: I1211 08:12:52.765337 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:52 crc kubenswrapper[4860]: I1211 08:12:52.765369 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:52 crc kubenswrapper[4860]: I1211 08:12:52.765393 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:52Z","lastTransitionTime":"2025-12-11T08:12:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:52 crc kubenswrapper[4860]: I1211 08:12:52.868486 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:52 crc kubenswrapper[4860]: I1211 08:12:52.868580 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:52 crc kubenswrapper[4860]: I1211 08:12:52.868598 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:52 crc kubenswrapper[4860]: I1211 08:12:52.868623 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:52 crc kubenswrapper[4860]: I1211 08:12:52.868686 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:52Z","lastTransitionTime":"2025-12-11T08:12:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:52 crc kubenswrapper[4860]: I1211 08:12:52.925942 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 08:12:52 crc kubenswrapper[4860]: I1211 08:12:52.925994 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 08:12:52 crc kubenswrapper[4860]: I1211 08:12:52.926005 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 08:12:52 crc kubenswrapper[4860]: I1211 08:12:52.926024 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 08:12:52 crc kubenswrapper[4860]: I1211 08:12:52.926040 4860 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T08:12:52Z","lastTransitionTime":"2025-12-11T08:12:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 08:12:52 crc kubenswrapper[4860]: I1211 08:12:52.990265 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-dnnhp"] Dec 11 08:12:52 crc kubenswrapper[4860]: I1211 08:12:52.991160 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dnnhp" Dec 11 08:12:52 crc kubenswrapper[4860]: I1211 08:12:52.993249 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 11 08:12:52 crc kubenswrapper[4860]: I1211 08:12:52.993993 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 11 08:12:52 crc kubenswrapper[4860]: I1211 08:12:52.995909 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 11 08:12:52 crc kubenswrapper[4860]: I1211 08:12:52.996437 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 11 08:12:53 crc kubenswrapper[4860]: I1211 08:12:53.046457 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=86.046431372 podStartE2EDuration="1m26.046431372s" podCreationTimestamp="2025-12-11 08:11:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:12:53.04489485 +0000 UTC m=+105.773413945" watchObservedRunningTime="2025-12-11 08:12:53.046431372 +0000 UTC m=+105.774950437" Dec 11 08:12:53 crc kubenswrapper[4860]: I1211 08:12:53.066940 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-pznwl" podStartSLOduration=85.066913794 podStartE2EDuration="1m25.066913794s" podCreationTimestamp="2025-12-11 08:11:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:12:53.065786495 +0000 UTC m=+105.794305590" watchObservedRunningTime="2025-12-11 08:12:53.066913794 +0000 UTC m=+105.795432869" Dec 11 08:12:53 crc kubenswrapper[4860]: I1211 08:12:53.076521 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/37579930-9925-4460-b6c2-8460db537e26-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-dnnhp\" (UID: \"37579930-9925-4460-b6c2-8460db537e26\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dnnhp" Dec 11 08:12:53 crc kubenswrapper[4860]: I1211 08:12:53.076583 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/37579930-9925-4460-b6c2-8460db537e26-service-ca\") pod \"cluster-version-operator-5c965bbfc6-dnnhp\" (UID: \"37579930-9925-4460-b6c2-8460db537e26\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dnnhp" Dec 11 08:12:53 crc kubenswrapper[4860]: I1211 08:12:53.076673 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/37579930-9925-4460-b6c2-8460db537e26-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-dnnhp\" (UID: \"37579930-9925-4460-b6c2-8460db537e26\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dnnhp" Dec 11 08:12:53 crc kubenswrapper[4860]: I1211 08:12:53.076744 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/37579930-9925-4460-b6c2-8460db537e26-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-dnnhp\" (UID: \"37579930-9925-4460-b6c2-8460db537e26\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dnnhp" Dec 11 08:12:53 crc kubenswrapper[4860]: I1211 08:12:53.076782 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/37579930-9925-4460-b6c2-8460db537e26-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-dnnhp\" (UID: \"37579930-9925-4460-b6c2-8460db537e26\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dnnhp" Dec 11 08:12:53 crc kubenswrapper[4860]: I1211 08:12:53.107493 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-854v6" podStartSLOduration=85.107469314 podStartE2EDuration="1m25.107469314s" podCreationTimestamp="2025-12-11 08:11:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:12:53.084816868 +0000 UTC m=+105.813335933" watchObservedRunningTime="2025-12-11 08:12:53.107469314 +0000 UTC m=+105.835988379" Dec 11 08:12:53 crc kubenswrapper[4860]: I1211 08:12:53.148515 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=59.14849668 podStartE2EDuration="59.14849668s" podCreationTimestamp="2025-12-11 08:11:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:12:53.148109247 +0000 UTC m=+105.876628312" watchObservedRunningTime="2025-12-11 08:12:53.14849668 +0000 UTC m=+105.877015745" Dec 11 08:12:53 crc kubenswrapper[4860]: I1211 08:12:53.148759 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-6xq7j" podStartSLOduration=86.148753029 podStartE2EDuration="1m26.148753029s" podCreationTimestamp="2025-12-11 08:11:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:12:53.133066411 +0000 UTC m=+105.861585476" watchObservedRunningTime="2025-12-11 08:12:53.148753029 +0000 UTC m=+105.877272094" Dec 11 08:12:53 crc kubenswrapper[4860]: I1211 08:12:53.177393 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/37579930-9925-4460-b6c2-8460db537e26-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-dnnhp\" (UID: \"37579930-9925-4460-b6c2-8460db537e26\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dnnhp" Dec 11 08:12:53 crc kubenswrapper[4860]: I1211 08:12:53.177514 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/37579930-9925-4460-b6c2-8460db537e26-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-dnnhp\" (UID: \"37579930-9925-4460-b6c2-8460db537e26\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dnnhp" Dec 11 08:12:53 crc kubenswrapper[4860]: I1211 08:12:53.177567 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/37579930-9925-4460-b6c2-8460db537e26-service-ca\") pod \"cluster-version-operator-5c965bbfc6-dnnhp\" (UID: \"37579930-9925-4460-b6c2-8460db537e26\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dnnhp" Dec 11 08:12:53 crc kubenswrapper[4860]: I1211 08:12:53.177637 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/37579930-9925-4460-b6c2-8460db537e26-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-dnnhp\" (UID: \"37579930-9925-4460-b6c2-8460db537e26\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dnnhp" Dec 11 08:12:53 crc kubenswrapper[4860]: I1211 08:12:53.177723 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/37579930-9925-4460-b6c2-8460db537e26-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-dnnhp\" (UID: \"37579930-9925-4460-b6c2-8460db537e26\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dnnhp" Dec 11 08:12:53 crc kubenswrapper[4860]: I1211 08:12:53.177767 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/37579930-9925-4460-b6c2-8460db537e26-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-dnnhp\" (UID: \"37579930-9925-4460-b6c2-8460db537e26\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dnnhp" Dec 11 08:12:53 crc kubenswrapper[4860]: I1211 08:12:53.177873 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/37579930-9925-4460-b6c2-8460db537e26-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-dnnhp\" (UID: \"37579930-9925-4460-b6c2-8460db537e26\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dnnhp" Dec 11 08:12:53 crc kubenswrapper[4860]: I1211 08:12:53.179614 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/37579930-9925-4460-b6c2-8460db537e26-service-ca\") pod \"cluster-version-operator-5c965bbfc6-dnnhp\" (UID: \"37579930-9925-4460-b6c2-8460db537e26\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dnnhp" Dec 11 08:12:53 crc kubenswrapper[4860]: I1211 08:12:53.191167 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/37579930-9925-4460-b6c2-8460db537e26-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-dnnhp\" (UID: \"37579930-9925-4460-b6c2-8460db537e26\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dnnhp" Dec 11 08:12:53 crc kubenswrapper[4860]: I1211 08:12:53.196481 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/37579930-9925-4460-b6c2-8460db537e26-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-dnnhp\" (UID: \"37579930-9925-4460-b6c2-8460db537e26\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dnnhp" Dec 11 08:12:53 crc kubenswrapper[4860]: I1211 08:12:53.224055 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=87.224024438 podStartE2EDuration="1m27.224024438s" podCreationTimestamp="2025-12-11 08:11:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:12:53.222737244 +0000 UTC m=+105.951256339" watchObservedRunningTime="2025-12-11 08:12:53.224024438 +0000 UTC m=+105.952543533" Dec 11 08:12:53 crc kubenswrapper[4860]: I1211 08:12:53.279442 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-26nc6" podStartSLOduration=85.279422206 podStartE2EDuration="1m25.279422206s" podCreationTimestamp="2025-12-11 08:11:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:12:53.279385825 +0000 UTC m=+106.007904890" watchObservedRunningTime="2025-12-11 08:12:53.279422206 +0000 UTC m=+106.007941251" Dec 11 08:12:53 crc kubenswrapper[4860]: I1211 08:12:53.312327 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-gsx9l" podStartSLOduration=86.312312033 podStartE2EDuration="1m26.312312033s" podCreationTimestamp="2025-12-11 08:11:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:12:53.311908689 +0000 UTC m=+106.040427744" watchObservedRunningTime="2025-12-11 08:12:53.312312033 +0000 UTC m=+106.040831088" Dec 11 08:12:53 crc kubenswrapper[4860]: I1211 08:12:53.319085 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dnnhp" Dec 11 08:12:53 crc kubenswrapper[4860]: I1211 08:12:53.352957 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=36.352936116 podStartE2EDuration="36.352936116s" podCreationTimestamp="2025-12-11 08:12:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:12:53.328620952 +0000 UTC m=+106.057140007" watchObservedRunningTime="2025-12-11 08:12:53.352936116 +0000 UTC m=+106.081455171" Dec 11 08:12:53 crc kubenswrapper[4860]: I1211 08:12:53.353140 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=84.353136113 podStartE2EDuration="1m24.353136113s" podCreationTimestamp="2025-12-11 08:11:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:12:53.352155319 +0000 UTC m=+106.080674384" watchObservedRunningTime="2025-12-11 08:12:53.353136113 +0000 UTC m=+106.081655168" Dec 11 08:12:54 crc kubenswrapper[4860]: I1211 08:12:54.204699 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dnnhp" event={"ID":"37579930-9925-4460-b6c2-8460db537e26","Type":"ContainerStarted","Data":"f1206d52e96d7c2623d83d9617d97c85a7599d5d23efcd07df63e487fde6dea0"} Dec 11 08:12:54 crc kubenswrapper[4860]: I1211 08:12:54.204879 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dnnhp" event={"ID":"37579930-9925-4460-b6c2-8460db537e26","Type":"ContainerStarted","Data":"106a713206ce150dcf197587642face314128d351e83bbc96d257fcdb28e1761"} Dec 11 08:12:54 crc kubenswrapper[4860]: I1211 08:12:54.223331 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podStartSLOduration=86.223310261 podStartE2EDuration="1m26.223310261s" podCreationTimestamp="2025-12-11 08:11:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:12:53.367123502 +0000 UTC m=+106.095642557" watchObservedRunningTime="2025-12-11 08:12:54.223310261 +0000 UTC m=+106.951829356" Dec 11 08:12:54 crc kubenswrapper[4860]: I1211 08:12:54.224481 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dnnhp" podStartSLOduration=86.22447322 podStartE2EDuration="1m26.22447322s" podCreationTimestamp="2025-12-11 08:11:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:12:54.222781812 +0000 UTC m=+106.951300907" watchObservedRunningTime="2025-12-11 08:12:54.22447322 +0000 UTC m=+106.952992305" Dec 11 08:12:54 crc kubenswrapper[4860]: I1211 08:12:54.578496 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:12:54 crc kubenswrapper[4860]: I1211 08:12:54.578509 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gjhlj" Dec 11 08:12:54 crc kubenswrapper[4860]: E1211 08:12:54.578780 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:12:54 crc kubenswrapper[4860]: I1211 08:12:54.578536 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:12:54 crc kubenswrapper[4860]: I1211 08:12:54.578522 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:12:54 crc kubenswrapper[4860]: E1211 08:12:54.579029 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gjhlj" podUID="5cfd640a-f3d2-4b28-87e1-5f49f6341e21" Dec 11 08:12:54 crc kubenswrapper[4860]: E1211 08:12:54.579089 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:12:54 crc kubenswrapper[4860]: E1211 08:12:54.579174 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:12:55 crc kubenswrapper[4860]: I1211 08:12:55.578614 4860 scope.go:117] "RemoveContainer" containerID="d9e83cda3e40d325604114a37fe6e8b707ac1f91a034b0d979c70b5f4f24f2e7" Dec 11 08:12:55 crc kubenswrapper[4860]: E1211 08:12:55.578822 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-xvnqp_openshift-ovn-kubernetes(0589a204-c98c-417d-8256-bfe3bf747660)\"" pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" podUID="0589a204-c98c-417d-8256-bfe3bf747660" Dec 11 08:12:56 crc kubenswrapper[4860]: I1211 08:12:56.578372 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gjhlj" Dec 11 08:12:56 crc kubenswrapper[4860]: I1211 08:12:56.578424 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:12:56 crc kubenswrapper[4860]: I1211 08:12:56.578509 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:12:56 crc kubenswrapper[4860]: E1211 08:12:56.578695 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gjhlj" podUID="5cfd640a-f3d2-4b28-87e1-5f49f6341e21" Dec 11 08:12:56 crc kubenswrapper[4860]: I1211 08:12:56.578725 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:12:56 crc kubenswrapper[4860]: E1211 08:12:56.578930 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:12:56 crc kubenswrapper[4860]: E1211 08:12:56.579036 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:12:56 crc kubenswrapper[4860]: E1211 08:12:56.579092 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:12:58 crc kubenswrapper[4860]: I1211 08:12:58.578180 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gjhlj" Dec 11 08:12:58 crc kubenswrapper[4860]: I1211 08:12:58.578304 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:12:58 crc kubenswrapper[4860]: I1211 08:12:58.578341 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:12:58 crc kubenswrapper[4860]: E1211 08:12:58.579341 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gjhlj" podUID="5cfd640a-f3d2-4b28-87e1-5f49f6341e21" Dec 11 08:12:58 crc kubenswrapper[4860]: E1211 08:12:58.579561 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:12:58 crc kubenswrapper[4860]: I1211 08:12:58.579591 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:12:58 crc kubenswrapper[4860]: E1211 08:12:58.579702 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:12:58 crc kubenswrapper[4860]: E1211 08:12:58.579871 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:13:00 crc kubenswrapper[4860]: I1211 08:13:00.578559 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:13:00 crc kubenswrapper[4860]: E1211 08:13:00.579191 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:13:00 crc kubenswrapper[4860]: I1211 08:13:00.578728 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:13:00 crc kubenswrapper[4860]: E1211 08:13:00.579402 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:13:00 crc kubenswrapper[4860]: I1211 08:13:00.578604 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gjhlj" Dec 11 08:13:00 crc kubenswrapper[4860]: E1211 08:13:00.579577 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gjhlj" podUID="5cfd640a-f3d2-4b28-87e1-5f49f6341e21" Dec 11 08:13:00 crc kubenswrapper[4860]: I1211 08:13:00.578840 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:13:00 crc kubenswrapper[4860]: E1211 08:13:00.579928 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:13:02 crc kubenswrapper[4860]: I1211 08:13:02.578390 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:13:02 crc kubenswrapper[4860]: I1211 08:13:02.578454 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:13:02 crc kubenswrapper[4860]: I1211 08:13:02.578484 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gjhlj" Dec 11 08:13:02 crc kubenswrapper[4860]: E1211 08:13:02.578602 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:13:02 crc kubenswrapper[4860]: I1211 08:13:02.578632 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:13:02 crc kubenswrapper[4860]: E1211 08:13:02.578815 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:13:02 crc kubenswrapper[4860]: E1211 08:13:02.578996 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gjhlj" podUID="5cfd640a-f3d2-4b28-87e1-5f49f6341e21" Dec 11 08:13:02 crc kubenswrapper[4860]: E1211 08:13:02.579142 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:13:04 crc kubenswrapper[4860]: I1211 08:13:04.236542 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-pznwl_5310c1fc-66c6-40aa-b1bf-5a59a2410f9c/kube-multus/1.log" Dec 11 08:13:04 crc kubenswrapper[4860]: I1211 08:13:04.238109 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-pznwl_5310c1fc-66c6-40aa-b1bf-5a59a2410f9c/kube-multus/0.log" Dec 11 08:13:04 crc kubenswrapper[4860]: I1211 08:13:04.238172 4860 generic.go:334] "Generic (PLEG): container finished" podID="5310c1fc-66c6-40aa-b1bf-5a59a2410f9c" containerID="34119a6367261d7e12174f914ee8cfaa55703a55210f83c58b0e4694c54a5b03" exitCode=1 Dec 11 08:13:04 crc kubenswrapper[4860]: I1211 08:13:04.238210 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-pznwl" event={"ID":"5310c1fc-66c6-40aa-b1bf-5a59a2410f9c","Type":"ContainerDied","Data":"34119a6367261d7e12174f914ee8cfaa55703a55210f83c58b0e4694c54a5b03"} Dec 11 08:13:04 crc kubenswrapper[4860]: I1211 08:13:04.238247 4860 scope.go:117] "RemoveContainer" containerID="d2d6ff00ce322a28607bc3f6db05c5e4fb14acb6777ea315786ef53963ca504a" Dec 11 08:13:04 crc kubenswrapper[4860]: I1211 08:13:04.238629 4860 scope.go:117] "RemoveContainer" containerID="34119a6367261d7e12174f914ee8cfaa55703a55210f83c58b0e4694c54a5b03" Dec 11 08:13:04 crc kubenswrapper[4860]: E1211 08:13:04.238908 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-pznwl_openshift-multus(5310c1fc-66c6-40aa-b1bf-5a59a2410f9c)\"" pod="openshift-multus/multus-pznwl" podUID="5310c1fc-66c6-40aa-b1bf-5a59a2410f9c" Dec 11 08:13:04 crc kubenswrapper[4860]: I1211 08:13:04.577865 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:13:04 crc kubenswrapper[4860]: I1211 08:13:04.577949 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gjhlj" Dec 11 08:13:04 crc kubenswrapper[4860]: I1211 08:13:04.577984 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:13:04 crc kubenswrapper[4860]: I1211 08:13:04.577991 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:13:04 crc kubenswrapper[4860]: E1211 08:13:04.578081 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:13:04 crc kubenswrapper[4860]: E1211 08:13:04.578180 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:13:04 crc kubenswrapper[4860]: E1211 08:13:04.578267 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gjhlj" podUID="5cfd640a-f3d2-4b28-87e1-5f49f6341e21" Dec 11 08:13:04 crc kubenswrapper[4860]: E1211 08:13:04.578442 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:13:05 crc kubenswrapper[4860]: I1211 08:13:05.243192 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-pznwl_5310c1fc-66c6-40aa-b1bf-5a59a2410f9c/kube-multus/1.log" Dec 11 08:13:06 crc kubenswrapper[4860]: I1211 08:13:06.578883 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:13:06 crc kubenswrapper[4860]: E1211 08:13:06.579170 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:13:06 crc kubenswrapper[4860]: I1211 08:13:06.579598 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:13:06 crc kubenswrapper[4860]: E1211 08:13:06.579817 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:13:06 crc kubenswrapper[4860]: I1211 08:13:06.580009 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:13:06 crc kubenswrapper[4860]: I1211 08:13:06.580080 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gjhlj" Dec 11 08:13:06 crc kubenswrapper[4860]: E1211 08:13:06.580163 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:13:06 crc kubenswrapper[4860]: E1211 08:13:06.580307 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gjhlj" podUID="5cfd640a-f3d2-4b28-87e1-5f49f6341e21" Dec 11 08:13:07 crc kubenswrapper[4860]: E1211 08:13:07.514496 4860 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Dec 11 08:13:07 crc kubenswrapper[4860]: I1211 08:13:07.581395 4860 scope.go:117] "RemoveContainer" containerID="d9e83cda3e40d325604114a37fe6e8b707ac1f91a034b0d979c70b5f4f24f2e7" Dec 11 08:13:07 crc kubenswrapper[4860]: E1211 08:13:07.581756 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-xvnqp_openshift-ovn-kubernetes(0589a204-c98c-417d-8256-bfe3bf747660)\"" pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" podUID="0589a204-c98c-417d-8256-bfe3bf747660" Dec 11 08:13:07 crc kubenswrapper[4860]: E1211 08:13:07.736498 4860 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 11 08:13:08 crc kubenswrapper[4860]: I1211 08:13:08.577959 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gjhlj" Dec 11 08:13:08 crc kubenswrapper[4860]: I1211 08:13:08.578023 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:13:08 crc kubenswrapper[4860]: I1211 08:13:08.577972 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:13:08 crc kubenswrapper[4860]: I1211 08:13:08.578146 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:13:08 crc kubenswrapper[4860]: E1211 08:13:08.578300 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gjhlj" podUID="5cfd640a-f3d2-4b28-87e1-5f49f6341e21" Dec 11 08:13:08 crc kubenswrapper[4860]: E1211 08:13:08.578457 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:13:08 crc kubenswrapper[4860]: E1211 08:13:08.578564 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:13:08 crc kubenswrapper[4860]: E1211 08:13:08.578630 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:13:10 crc kubenswrapper[4860]: I1211 08:13:10.578016 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:13:10 crc kubenswrapper[4860]: I1211 08:13:10.578148 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:13:10 crc kubenswrapper[4860]: E1211 08:13:10.578192 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:13:10 crc kubenswrapper[4860]: I1211 08:13:10.578306 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:13:10 crc kubenswrapper[4860]: E1211 08:13:10.578480 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:13:10 crc kubenswrapper[4860]: E1211 08:13:10.578602 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:13:10 crc kubenswrapper[4860]: I1211 08:13:10.580007 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gjhlj" Dec 11 08:13:10 crc kubenswrapper[4860]: E1211 08:13:10.580438 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gjhlj" podUID="5cfd640a-f3d2-4b28-87e1-5f49f6341e21" Dec 11 08:13:12 crc kubenswrapper[4860]: I1211 08:13:12.578545 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:13:12 crc kubenswrapper[4860]: I1211 08:13:12.578572 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:13:12 crc kubenswrapper[4860]: I1211 08:13:12.578572 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gjhlj" Dec 11 08:13:12 crc kubenswrapper[4860]: I1211 08:13:12.578916 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:13:12 crc kubenswrapper[4860]: E1211 08:13:12.579120 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:13:12 crc kubenswrapper[4860]: E1211 08:13:12.579537 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:13:12 crc kubenswrapper[4860]: E1211 08:13:12.579870 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gjhlj" podUID="5cfd640a-f3d2-4b28-87e1-5f49f6341e21" Dec 11 08:13:12 crc kubenswrapper[4860]: E1211 08:13:12.580095 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:13:12 crc kubenswrapper[4860]: E1211 08:13:12.738337 4860 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 11 08:13:14 crc kubenswrapper[4860]: I1211 08:13:14.578469 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:13:14 crc kubenswrapper[4860]: I1211 08:13:14.578569 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:13:14 crc kubenswrapper[4860]: E1211 08:13:14.578751 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:13:14 crc kubenswrapper[4860]: I1211 08:13:14.578790 4860 scope.go:117] "RemoveContainer" containerID="34119a6367261d7e12174f914ee8cfaa55703a55210f83c58b0e4694c54a5b03" Dec 11 08:13:14 crc kubenswrapper[4860]: I1211 08:13:14.578884 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gjhlj" Dec 11 08:13:14 crc kubenswrapper[4860]: E1211 08:13:14.579033 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:13:14 crc kubenswrapper[4860]: I1211 08:13:14.579095 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:13:14 crc kubenswrapper[4860]: E1211 08:13:14.579180 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gjhlj" podUID="5cfd640a-f3d2-4b28-87e1-5f49f6341e21" Dec 11 08:13:14 crc kubenswrapper[4860]: E1211 08:13:14.579265 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:13:15 crc kubenswrapper[4860]: I1211 08:13:15.281947 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-pznwl_5310c1fc-66c6-40aa-b1bf-5a59a2410f9c/kube-multus/1.log" Dec 11 08:13:15 crc kubenswrapper[4860]: I1211 08:13:15.282028 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-pznwl" event={"ID":"5310c1fc-66c6-40aa-b1bf-5a59a2410f9c","Type":"ContainerStarted","Data":"c50698b594e2b9e096d97dd750433d216996139113b1349d3ba87a70d057511b"} Dec 11 08:13:16 crc kubenswrapper[4860]: I1211 08:13:16.578336 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gjhlj" Dec 11 08:13:16 crc kubenswrapper[4860]: I1211 08:13:16.578483 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:13:16 crc kubenswrapper[4860]: I1211 08:13:16.578525 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:13:16 crc kubenswrapper[4860]: I1211 08:13:16.578556 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:13:16 crc kubenswrapper[4860]: E1211 08:13:16.578928 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gjhlj" podUID="5cfd640a-f3d2-4b28-87e1-5f49f6341e21" Dec 11 08:13:16 crc kubenswrapper[4860]: E1211 08:13:16.579059 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:13:16 crc kubenswrapper[4860]: E1211 08:13:16.579193 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:13:16 crc kubenswrapper[4860]: E1211 08:13:16.579346 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:13:17 crc kubenswrapper[4860]: E1211 08:13:17.739817 4860 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 11 08:13:18 crc kubenswrapper[4860]: I1211 08:13:18.578271 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gjhlj" Dec 11 08:13:18 crc kubenswrapper[4860]: I1211 08:13:18.578346 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:13:18 crc kubenswrapper[4860]: I1211 08:13:18.578380 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:13:18 crc kubenswrapper[4860]: I1211 08:13:18.578276 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:13:18 crc kubenswrapper[4860]: E1211 08:13:18.578425 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gjhlj" podUID="5cfd640a-f3d2-4b28-87e1-5f49f6341e21" Dec 11 08:13:18 crc kubenswrapper[4860]: E1211 08:13:18.578534 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:13:18 crc kubenswrapper[4860]: E1211 08:13:18.578602 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:13:18 crc kubenswrapper[4860]: E1211 08:13:18.578633 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:13:20 crc kubenswrapper[4860]: I1211 08:13:20.578005 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gjhlj" Dec 11 08:13:20 crc kubenswrapper[4860]: I1211 08:13:20.578693 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:13:20 crc kubenswrapper[4860]: I1211 08:13:20.578799 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:13:20 crc kubenswrapper[4860]: I1211 08:13:20.578807 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:13:20 crc kubenswrapper[4860]: E1211 08:13:20.578744 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gjhlj" podUID="5cfd640a-f3d2-4b28-87e1-5f49f6341e21" Dec 11 08:13:20 crc kubenswrapper[4860]: E1211 08:13:20.578932 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:13:20 crc kubenswrapper[4860]: E1211 08:13:20.579030 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:13:20 crc kubenswrapper[4860]: E1211 08:13:20.579092 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:13:20 crc kubenswrapper[4860]: I1211 08:13:20.580368 4860 scope.go:117] "RemoveContainer" containerID="d9e83cda3e40d325604114a37fe6e8b707ac1f91a034b0d979c70b5f4f24f2e7" Dec 11 08:13:21 crc kubenswrapper[4860]: I1211 08:13:21.302716 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-xvnqp_0589a204-c98c-417d-8256-bfe3bf747660/ovnkube-controller/3.log" Dec 11 08:13:21 crc kubenswrapper[4860]: I1211 08:13:21.306059 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" event={"ID":"0589a204-c98c-417d-8256-bfe3bf747660","Type":"ContainerStarted","Data":"4513cf6bc605dd49d55acb42fa24068e7e61a7b40104f7c5a17dc78cc5694388"} Dec 11 08:13:21 crc kubenswrapper[4860]: I1211 08:13:21.306801 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" Dec 11 08:13:21 crc kubenswrapper[4860]: I1211 08:13:21.334714 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" podStartSLOduration=113.33469437 podStartE2EDuration="1m53.33469437s" podCreationTimestamp="2025-12-11 08:11:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:13:21.332521476 +0000 UTC m=+134.061040541" watchObservedRunningTime="2025-12-11 08:13:21.33469437 +0000 UTC m=+134.063213425" Dec 11 08:13:21 crc kubenswrapper[4860]: I1211 08:13:21.591424 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-gjhlj"] Dec 11 08:13:21 crc kubenswrapper[4860]: I1211 08:13:21.591547 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gjhlj" Dec 11 08:13:21 crc kubenswrapper[4860]: E1211 08:13:21.591797 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gjhlj" podUID="5cfd640a-f3d2-4b28-87e1-5f49f6341e21" Dec 11 08:13:22 crc kubenswrapper[4860]: I1211 08:13:22.578227 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:13:22 crc kubenswrapper[4860]: I1211 08:13:22.578261 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:13:22 crc kubenswrapper[4860]: E1211 08:13:22.578346 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:13:22 crc kubenswrapper[4860]: I1211 08:13:22.578263 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:13:22 crc kubenswrapper[4860]: E1211 08:13:22.578440 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:13:22 crc kubenswrapper[4860]: E1211 08:13:22.578521 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:13:22 crc kubenswrapper[4860]: E1211 08:13:22.740903 4860 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 11 08:13:23 crc kubenswrapper[4860]: I1211 08:13:23.578068 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gjhlj" Dec 11 08:13:23 crc kubenswrapper[4860]: E1211 08:13:23.578847 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gjhlj" podUID="5cfd640a-f3d2-4b28-87e1-5f49f6341e21" Dec 11 08:13:24 crc kubenswrapper[4860]: I1211 08:13:24.577936 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:13:24 crc kubenswrapper[4860]: I1211 08:13:24.577966 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:13:24 crc kubenswrapper[4860]: E1211 08:13:24.578043 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:13:24 crc kubenswrapper[4860]: I1211 08:13:24.577943 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:13:24 crc kubenswrapper[4860]: E1211 08:13:24.578263 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:13:24 crc kubenswrapper[4860]: E1211 08:13:24.578287 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:13:25 crc kubenswrapper[4860]: I1211 08:13:25.578358 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gjhlj" Dec 11 08:13:25 crc kubenswrapper[4860]: E1211 08:13:25.578555 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gjhlj" podUID="5cfd640a-f3d2-4b28-87e1-5f49f6341e21" Dec 11 08:13:26 crc kubenswrapper[4860]: I1211 08:13:26.577765 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:13:26 crc kubenswrapper[4860]: I1211 08:13:26.577866 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:13:26 crc kubenswrapper[4860]: E1211 08:13:26.577938 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 11 08:13:26 crc kubenswrapper[4860]: I1211 08:13:26.577782 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:13:26 crc kubenswrapper[4860]: E1211 08:13:26.578092 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 11 08:13:26 crc kubenswrapper[4860]: E1211 08:13:26.578134 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 11 08:13:27 crc kubenswrapper[4860]: I1211 08:13:27.578295 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gjhlj" Dec 11 08:13:27 crc kubenswrapper[4860]: E1211 08:13:27.581017 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gjhlj" podUID="5cfd640a-f3d2-4b28-87e1-5f49f6341e21" Dec 11 08:13:28 crc kubenswrapper[4860]: I1211 08:13:28.578489 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:13:28 crc kubenswrapper[4860]: I1211 08:13:28.578586 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:13:28 crc kubenswrapper[4860]: I1211 08:13:28.578636 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:13:28 crc kubenswrapper[4860]: I1211 08:13:28.580261 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 11 08:13:28 crc kubenswrapper[4860]: I1211 08:13:28.580361 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 11 08:13:28 crc kubenswrapper[4860]: I1211 08:13:28.580429 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 11 08:13:28 crc kubenswrapper[4860]: I1211 08:13:28.580711 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 11 08:13:29 crc kubenswrapper[4860]: I1211 08:13:29.578567 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gjhlj" Dec 11 08:13:29 crc kubenswrapper[4860]: I1211 08:13:29.581995 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 11 08:13:29 crc kubenswrapper[4860]: I1211 08:13:29.582115 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 11 08:13:33 crc kubenswrapper[4860]: I1211 08:13:33.846418 4860 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Dec 11 08:13:33 crc kubenswrapper[4860]: I1211 08:13:33.902089 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-jxft9"] Dec 11 08:13:33 crc kubenswrapper[4860]: I1211 08:13:33.903572 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-jxft9" Dec 11 08:13:33 crc kubenswrapper[4860]: I1211 08:13:33.904305 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-cpflv"] Dec 11 08:13:33 crc kubenswrapper[4860]: I1211 08:13:33.905506 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cpflv" Dec 11 08:13:33 crc kubenswrapper[4860]: I1211 08:13:33.907442 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-nc65d"] Dec 11 08:13:33 crc kubenswrapper[4860]: I1211 08:13:33.908384 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-nc65d" Dec 11 08:13:33 crc kubenswrapper[4860]: I1211 08:13:33.913075 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-24dmg"] Dec 11 08:13:33 crc kubenswrapper[4860]: I1211 08:13:33.913840 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-24dmg" Dec 11 08:13:33 crc kubenswrapper[4860]: I1211 08:13:33.915816 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 11 08:13:33 crc kubenswrapper[4860]: I1211 08:13:33.917557 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 11 08:13:33 crc kubenswrapper[4860]: I1211 08:13:33.917621 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 11 08:13:33 crc kubenswrapper[4860]: I1211 08:13:33.917796 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 11 08:13:33 crc kubenswrapper[4860]: I1211 08:13:33.919040 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-rvtb7"] Dec 11 08:13:33 crc kubenswrapper[4860]: I1211 08:13:33.920002 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rvtb7" Dec 11 08:13:33 crc kubenswrapper[4860]: I1211 08:13:33.932284 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-rhl5v"] Dec 11 08:13:33 crc kubenswrapper[4860]: I1211 08:13:33.932888 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-rhl5v" Dec 11 08:13:33 crc kubenswrapper[4860]: I1211 08:13:33.934764 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gkg77"] Dec 11 08:13:33 crc kubenswrapper[4860]: I1211 08:13:33.935388 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gkg77" Dec 11 08:13:33 crc kubenswrapper[4860]: I1211 08:13:33.936199 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-b4q7q"] Dec 11 08:13:33 crc kubenswrapper[4860]: I1211 08:13:33.936736 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-b4q7q" Dec 11 08:13:33 crc kubenswrapper[4860]: I1211 08:13:33.938949 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-dz7kr"] Dec 11 08:13:33 crc kubenswrapper[4860]: I1211 08:13:33.939417 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-dz7kr" Dec 11 08:13:33 crc kubenswrapper[4860]: I1211 08:13:33.940709 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-z5lg5"] Dec 11 08:13:33 crc kubenswrapper[4860]: I1211 08:13:33.941561 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-z5lg5" Dec 11 08:13:33 crc kubenswrapper[4860]: I1211 08:13:33.948674 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-j5kxr"] Dec 11 08:13:33 crc kubenswrapper[4860]: I1211 08:13:33.949256 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lbrj7"] Dec 11 08:13:33 crc kubenswrapper[4860]: I1211 08:13:33.949534 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 11 08:13:33 crc kubenswrapper[4860]: I1211 08:13:33.949561 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-j5kxr" Dec 11 08:13:33 crc kubenswrapper[4860]: I1211 08:13:33.949780 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lbrj7" Dec 11 08:13:33 crc kubenswrapper[4860]: I1211 08:13:33.963409 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 11 08:13:33 crc kubenswrapper[4860]: I1211 08:13:33.963684 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 11 08:13:33 crc kubenswrapper[4860]: I1211 08:13:33.968161 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 11 08:13:33 crc kubenswrapper[4860]: I1211 08:13:33.968532 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 11 08:13:33 crc kubenswrapper[4860]: I1211 08:13:33.970512 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-wrqbs"] Dec 11 08:13:33 crc kubenswrapper[4860]: I1211 08:13:33.971331 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wrqbs" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.012274 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.012885 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.016414 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.024554 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.024807 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.027227 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.027243 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.027238 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.027958 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.028350 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.028675 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.028732 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.029385 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.031822 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.031964 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.032105 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.032156 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.032284 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.032865 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.033029 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.033510 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.033558 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.033710 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.033889 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.033907 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8x4xt"] Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.034002 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.034109 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.034336 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.034431 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8x4xt" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.034447 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.034548 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.035103 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.035160 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.035196 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.035299 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.035302 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.035408 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.035419 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.035894 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.035896 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.036014 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.036132 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.038715 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-zf572"] Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.039447 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-zf572" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.049030 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.049278 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.049476 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.049695 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.049737 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.049928 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.050030 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.050056 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.050252 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.050765 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.050818 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.055822 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q6ck8"] Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.056262 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q6ck8" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.056629 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.056823 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.057022 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.058770 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.059157 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.060738 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qww7r\" (UniqueName: \"kubernetes.io/projected/e55e20fb-bef2-4fee-b6a7-60e2d9d61faa-kube-api-access-qww7r\") pod \"cluster-samples-operator-665b6dd947-z5lg5\" (UID: \"e55e20fb-bef2-4fee-b6a7-60e2d9d61faa\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-z5lg5" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.060774 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/d7d1536c-ed04-4199-b839-b94e9482a63c-oauth-serving-cert\") pod \"console-f9d7485db-rhl5v\" (UID: \"d7d1536c-ed04-4199-b839-b94e9482a63c\") " pod="openshift-console/console-f9d7485db-rhl5v" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.060804 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/e86ba495-65e4-4539-928e-366376df64d8-image-import-ca\") pod \"apiserver-76f77b778f-jxft9\" (UID: \"e86ba495-65e4-4539-928e-366376df64d8\") " pod="openshift-apiserver/apiserver-76f77b778f-jxft9" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.060824 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/cb55e751-9913-4a3c-a0b4-3e681129d052-images\") pod \"machine-api-operator-5694c8668f-nc65d\" (UID: \"cb55e751-9913-4a3c-a0b4-3e681129d052\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-nc65d" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.060846 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x9wpg\" (UniqueName: \"kubernetes.io/projected/09073ec8-36df-4ffb-a43a-10611999d8cb-kube-api-access-x9wpg\") pod \"openshift-apiserver-operator-796bbdcf4f-8x4xt\" (UID: \"09073ec8-36df-4ffb-a43a-10611999d8cb\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8x4xt" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.060870 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/e55e20fb-bef2-4fee-b6a7-60e2d9d61faa-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-z5lg5\" (UID: \"e55e20fb-bef2-4fee-b6a7-60e2d9d61faa\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-z5lg5" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.060891 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e86ba495-65e4-4539-928e-366376df64d8-config\") pod \"apiserver-76f77b778f-jxft9\" (UID: \"e86ba495-65e4-4539-928e-366376df64d8\") " pod="openshift-apiserver/apiserver-76f77b778f-jxft9" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.060913 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e86ba495-65e4-4539-928e-366376df64d8-serving-cert\") pod \"apiserver-76f77b778f-jxft9\" (UID: \"e86ba495-65e4-4539-928e-366376df64d8\") " pod="openshift-apiserver/apiserver-76f77b778f-jxft9" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.060934 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/ccbceeb2-58ab-4409-82df-5c99189569b8-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-dz7kr\" (UID: \"ccbceeb2-58ab-4409-82df-5c99189569b8\") " pod="openshift-authentication/oauth-openshift-558db77b4-dz7kr" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.060954 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/fb3441e4-e612-4cfe-9f12-6497cd9176a9-trusted-ca\") pod \"console-operator-58897d9998-b4q7q\" (UID: \"fb3441e4-e612-4cfe-9f12-6497cd9176a9\") " pod="openshift-console-operator/console-operator-58897d9998-b4q7q" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.060974 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d7d1536c-ed04-4199-b839-b94e9482a63c-trusted-ca-bundle\") pod \"console-f9d7485db-rhl5v\" (UID: \"d7d1536c-ed04-4199-b839-b94e9482a63c\") " pod="openshift-console/console-f9d7485db-rhl5v" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.060994 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2c04dbaf-f385-4cee-bccf-8506ecdeb680-serving-cert\") pod \"authentication-operator-69f744f599-j5kxr\" (UID: \"2c04dbaf-f385-4cee-bccf-8506ecdeb680\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-j5kxr" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.061015 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ngxkg\" (UniqueName: \"kubernetes.io/projected/2c04dbaf-f385-4cee-bccf-8506ecdeb680-kube-api-access-ngxkg\") pod \"authentication-operator-69f744f599-j5kxr\" (UID: \"2c04dbaf-f385-4cee-bccf-8506ecdeb680\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-j5kxr" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.061036 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6hs5b\" (UniqueName: \"kubernetes.io/projected/fb3441e4-e612-4cfe-9f12-6497cd9176a9-kube-api-access-6hs5b\") pod \"console-operator-58897d9998-b4q7q\" (UID: \"fb3441e4-e612-4cfe-9f12-6497cd9176a9\") " pod="openshift-console-operator/console-operator-58897d9998-b4q7q" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.061054 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/e86ba495-65e4-4539-928e-366376df64d8-etcd-serving-ca\") pod \"apiserver-76f77b778f-jxft9\" (UID: \"e86ba495-65e4-4539-928e-366376df64d8\") " pod="openshift-apiserver/apiserver-76f77b778f-jxft9" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.061074 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/d7d1536c-ed04-4199-b839-b94e9482a63c-console-config\") pod \"console-f9d7485db-rhl5v\" (UID: \"d7d1536c-ed04-4199-b839-b94e9482a63c\") " pod="openshift-console/console-f9d7485db-rhl5v" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.061094 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ccbceeb2-58ab-4409-82df-5c99189569b8-audit-dir\") pod \"oauth-openshift-558db77b4-dz7kr\" (UID: \"ccbceeb2-58ab-4409-82df-5c99189569b8\") " pod="openshift-authentication/oauth-openshift-558db77b4-dz7kr" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.061114 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b7feb2dc-f1f2-4228-9974-8a771fe98f3c-serving-cert\") pod \"controller-manager-879f6c89f-24dmg\" (UID: \"b7feb2dc-f1f2-4228-9974-8a771fe98f3c\") " pod="openshift-controller-manager/controller-manager-879f6c89f-24dmg" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.061133 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/e86ba495-65e4-4539-928e-366376df64d8-encryption-config\") pod \"apiserver-76f77b778f-jxft9\" (UID: \"e86ba495-65e4-4539-928e-366376df64d8\") " pod="openshift-apiserver/apiserver-76f77b778f-jxft9" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.061153 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2c04dbaf-f385-4cee-bccf-8506ecdeb680-config\") pod \"authentication-operator-69f744f599-j5kxr\" (UID: \"2c04dbaf-f385-4cee-bccf-8506ecdeb680\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-j5kxr" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.061186 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/ccbceeb2-58ab-4409-82df-5c99189569b8-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-dz7kr\" (UID: \"ccbceeb2-58ab-4409-82df-5c99189569b8\") " pod="openshift-authentication/oauth-openshift-558db77b4-dz7kr" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.061207 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c094602b-e1ec-4b9a-9640-cd6250a1fd94-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-gkg77\" (UID: \"c094602b-e1ec-4b9a-9640-cd6250a1fd94\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gkg77" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.061225 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09073ec8-36df-4ffb-a43a-10611999d8cb-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-8x4xt\" (UID: \"09073ec8-36df-4ffb-a43a-10611999d8cb\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8x4xt" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.061247 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c465q\" (UniqueName: \"kubernetes.io/projected/d7d1536c-ed04-4199-b839-b94e9482a63c-kube-api-access-c465q\") pod \"console-f9d7485db-rhl5v\" (UID: \"d7d1536c-ed04-4199-b839-b94e9482a63c\") " pod="openshift-console/console-f9d7485db-rhl5v" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.061269 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/e84a804f-75e1-46a2-b37a-f868de516379-available-featuregates\") pod \"openshift-config-operator-7777fb866f-rvtb7\" (UID: \"e84a804f-75e1-46a2-b37a-f868de516379\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rvtb7" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.061310 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9da229ee-dedc-417f-aec7-d2d1d9407a41-config\") pod \"kube-controller-manager-operator-78b949d7b-lbrj7\" (UID: \"9da229ee-dedc-417f-aec7-d2d1d9407a41\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lbrj7" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.061330 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/e86ba495-65e4-4539-928e-366376df64d8-node-pullsecrets\") pod \"apiserver-76f77b778f-jxft9\" (UID: \"e86ba495-65e4-4539-928e-366376df64d8\") " pod="openshift-apiserver/apiserver-76f77b778f-jxft9" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.061349 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/df36e993-8a70-42c5-ac22-508485572d35-etcd-client\") pod \"apiserver-7bbb656c7d-cpflv\" (UID: \"df36e993-8a70-42c5-ac22-508485572d35\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cpflv" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.061368 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cb55e751-9913-4a3c-a0b4-3e681129d052-config\") pod \"machine-api-operator-5694c8668f-nc65d\" (UID: \"cb55e751-9913-4a3c-a0b4-3e681129d052\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-nc65d" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.061387 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/e86ba495-65e4-4539-928e-366376df64d8-audit\") pod \"apiserver-76f77b778f-jxft9\" (UID: \"e86ba495-65e4-4539-928e-366376df64d8\") " pod="openshift-apiserver/apiserver-76f77b778f-jxft9" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.061406 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e86ba495-65e4-4539-928e-366376df64d8-etcd-client\") pod \"apiserver-76f77b778f-jxft9\" (UID: \"e86ba495-65e4-4539-928e-366376df64d8\") " pod="openshift-apiserver/apiserver-76f77b778f-jxft9" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.061425 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d7d1536c-ed04-4199-b839-b94e9482a63c-service-ca\") pod \"console-f9d7485db-rhl5v\" (UID: \"d7d1536c-ed04-4199-b839-b94e9482a63c\") " pod="openshift-console/console-f9d7485db-rhl5v" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.061445 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/df36e993-8a70-42c5-ac22-508485572d35-audit-dir\") pod \"apiserver-7bbb656c7d-cpflv\" (UID: \"df36e993-8a70-42c5-ac22-508485572d35\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cpflv" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.061465 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9da229ee-dedc-417f-aec7-d2d1d9407a41-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-lbrj7\" (UID: \"9da229ee-dedc-417f-aec7-d2d1d9407a41\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lbrj7" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.061484 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/d7d1536c-ed04-4199-b839-b94e9482a63c-console-oauth-config\") pod \"console-f9d7485db-rhl5v\" (UID: \"d7d1536c-ed04-4199-b839-b94e9482a63c\") " pod="openshift-console/console-f9d7485db-rhl5v" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.061504 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m96cx\" (UniqueName: \"kubernetes.io/projected/cb55e751-9913-4a3c-a0b4-3e681129d052-kube-api-access-m96cx\") pod \"machine-api-operator-5694c8668f-nc65d\" (UID: \"cb55e751-9913-4a3c-a0b4-3e681129d052\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-nc65d" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.061523 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2c04dbaf-f385-4cee-bccf-8506ecdeb680-service-ca-bundle\") pod \"authentication-operator-69f744f599-j5kxr\" (UID: \"2c04dbaf-f385-4cee-bccf-8506ecdeb680\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-j5kxr" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.061542 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/ccbceeb2-58ab-4409-82df-5c99189569b8-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-dz7kr\" (UID: \"ccbceeb2-58ab-4409-82df-5c99189569b8\") " pod="openshift-authentication/oauth-openshift-558db77b4-dz7kr" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.061561 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b7feb2dc-f1f2-4228-9974-8a771fe98f3c-client-ca\") pod \"controller-manager-879f6c89f-24dmg\" (UID: \"b7feb2dc-f1f2-4228-9974-8a771fe98f3c\") " pod="openshift-controller-manager/controller-manager-879f6c89f-24dmg" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.061581 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/d7d1536c-ed04-4199-b839-b94e9482a63c-console-serving-cert\") pod \"console-f9d7485db-rhl5v\" (UID: \"d7d1536c-ed04-4199-b839-b94e9482a63c\") " pod="openshift-console/console-f9d7485db-rhl5v" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.061599 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/df36e993-8a70-42c5-ac22-508485572d35-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-cpflv\" (UID: \"df36e993-8a70-42c5-ac22-508485572d35\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cpflv" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.061628 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/df36e993-8a70-42c5-ac22-508485572d35-audit-policies\") pod \"apiserver-7bbb656c7d-cpflv\" (UID: \"df36e993-8a70-42c5-ac22-508485572d35\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cpflv" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.061666 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-67q4s\" (UniqueName: \"kubernetes.io/projected/e84a804f-75e1-46a2-b37a-f868de516379-kube-api-access-67q4s\") pod \"openshift-config-operator-7777fb866f-rvtb7\" (UID: \"e84a804f-75e1-46a2-b37a-f868de516379\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rvtb7" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.061692 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/ccbceeb2-58ab-4409-82df-5c99189569b8-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-dz7kr\" (UID: \"ccbceeb2-58ab-4409-82df-5c99189569b8\") " pod="openshift-authentication/oauth-openshift-558db77b4-dz7kr" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.061713 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e84a804f-75e1-46a2-b37a-f868de516379-serving-cert\") pod \"openshift-config-operator-7777fb866f-rvtb7\" (UID: \"e84a804f-75e1-46a2-b37a-f868de516379\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rvtb7" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.061732 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9da229ee-dedc-417f-aec7-d2d1d9407a41-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-lbrj7\" (UID: \"9da229ee-dedc-417f-aec7-d2d1d9407a41\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lbrj7" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.061752 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fb3441e4-e612-4cfe-9f12-6497cd9176a9-config\") pod \"console-operator-58897d9998-b4q7q\" (UID: \"fb3441e4-e612-4cfe-9f12-6497cd9176a9\") " pod="openshift-console-operator/console-operator-58897d9998-b4q7q" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.061773 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e86ba495-65e4-4539-928e-366376df64d8-audit-dir\") pod \"apiserver-76f77b778f-jxft9\" (UID: \"e86ba495-65e4-4539-928e-366376df64d8\") " pod="openshift-apiserver/apiserver-76f77b778f-jxft9" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.061791 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/cb55e751-9913-4a3c-a0b4-3e681129d052-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-nc65d\" (UID: \"cb55e751-9913-4a3c-a0b4-3e681129d052\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-nc65d" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.061821 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ccbceeb2-58ab-4409-82df-5c99189569b8-audit-policies\") pod \"oauth-openshift-558db77b4-dz7kr\" (UID: \"ccbceeb2-58ab-4409-82df-5c99189569b8\") " pod="openshift-authentication/oauth-openshift-558db77b4-dz7kr" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.061852 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/ccbceeb2-58ab-4409-82df-5c99189569b8-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-dz7kr\" (UID: \"ccbceeb2-58ab-4409-82df-5c99189569b8\") " pod="openshift-authentication/oauth-openshift-558db77b4-dz7kr" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.061872 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/ccbceeb2-58ab-4409-82df-5c99189569b8-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-dz7kr\" (UID: \"ccbceeb2-58ab-4409-82df-5c99189569b8\") " pod="openshift-authentication/oauth-openshift-558db77b4-dz7kr" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.061892 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c094602b-e1ec-4b9a-9640-cd6250a1fd94-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-gkg77\" (UID: \"c094602b-e1ec-4b9a-9640-cd6250a1fd94\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gkg77" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.061913 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/ccbceeb2-58ab-4409-82df-5c99189569b8-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-dz7kr\" (UID: \"ccbceeb2-58ab-4409-82df-5c99189569b8\") " pod="openshift-authentication/oauth-openshift-558db77b4-dz7kr" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.061933 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/ccbceeb2-58ab-4409-82df-5c99189569b8-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-dz7kr\" (UID: \"ccbceeb2-58ab-4409-82df-5c99189569b8\") " pod="openshift-authentication/oauth-openshift-558db77b4-dz7kr" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.061954 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fb3441e4-e612-4cfe-9f12-6497cd9176a9-serving-cert\") pod \"console-operator-58897d9998-b4q7q\" (UID: \"fb3441e4-e612-4cfe-9f12-6497cd9176a9\") " pod="openshift-console-operator/console-operator-58897d9998-b4q7q" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.061976 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rkfd4\" (UniqueName: \"kubernetes.io/projected/c094602b-e1ec-4b9a-9640-cd6250a1fd94-kube-api-access-rkfd4\") pod \"openshift-controller-manager-operator-756b6f6bc6-gkg77\" (UID: \"c094602b-e1ec-4b9a-9640-cd6250a1fd94\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gkg77" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.061996 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8k4gt\" (UniqueName: \"kubernetes.io/projected/ccbceeb2-58ab-4409-82df-5c99189569b8-kube-api-access-8k4gt\") pod \"oauth-openshift-558db77b4-dz7kr\" (UID: \"ccbceeb2-58ab-4409-82df-5c99189569b8\") " pod="openshift-authentication/oauth-openshift-558db77b4-dz7kr" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.062017 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/df36e993-8a70-42c5-ac22-508485572d35-encryption-config\") pod \"apiserver-7bbb656c7d-cpflv\" (UID: \"df36e993-8a70-42c5-ac22-508485572d35\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cpflv" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.062036 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e86ba495-65e4-4539-928e-366376df64d8-trusted-ca-bundle\") pod \"apiserver-76f77b778f-jxft9\" (UID: \"e86ba495-65e4-4539-928e-366376df64d8\") " pod="openshift-apiserver/apiserver-76f77b778f-jxft9" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.062055 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/ccbceeb2-58ab-4409-82df-5c99189569b8-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-dz7kr\" (UID: \"ccbceeb2-58ab-4409-82df-5c99189569b8\") " pod="openshift-authentication/oauth-openshift-558db77b4-dz7kr" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.062075 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/df36e993-8a70-42c5-ac22-508485572d35-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-cpflv\" (UID: \"df36e993-8a70-42c5-ac22-508485572d35\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cpflv" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.062095 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b7feb2dc-f1f2-4228-9974-8a771fe98f3c-config\") pod \"controller-manager-879f6c89f-24dmg\" (UID: \"b7feb2dc-f1f2-4228-9974-8a771fe98f3c\") " pod="openshift-controller-manager/controller-manager-879f6c89f-24dmg" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.062118 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/ccbceeb2-58ab-4409-82df-5c99189569b8-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-dz7kr\" (UID: \"ccbceeb2-58ab-4409-82df-5c99189569b8\") " pod="openshift-authentication/oauth-openshift-558db77b4-dz7kr" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.062137 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mtwdk\" (UniqueName: \"kubernetes.io/projected/df36e993-8a70-42c5-ac22-508485572d35-kube-api-access-mtwdk\") pod \"apiserver-7bbb656c7d-cpflv\" (UID: \"df36e993-8a70-42c5-ac22-508485572d35\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cpflv" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.062158 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2c04dbaf-f385-4cee-bccf-8506ecdeb680-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-j5kxr\" (UID: \"2c04dbaf-f385-4cee-bccf-8506ecdeb680\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-j5kxr" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.062177 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ccbceeb2-58ab-4409-82df-5c99189569b8-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-dz7kr\" (UID: \"ccbceeb2-58ab-4409-82df-5c99189569b8\") " pod="openshift-authentication/oauth-openshift-558db77b4-dz7kr" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.062196 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b7feb2dc-f1f2-4228-9974-8a771fe98f3c-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-24dmg\" (UID: \"b7feb2dc-f1f2-4228-9974-8a771fe98f3c\") " pod="openshift-controller-manager/controller-manager-879f6c89f-24dmg" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.062220 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x629f\" (UniqueName: \"kubernetes.io/projected/e86ba495-65e4-4539-928e-366376df64d8-kube-api-access-x629f\") pod \"apiserver-76f77b778f-jxft9\" (UID: \"e86ba495-65e4-4539-928e-366376df64d8\") " pod="openshift-apiserver/apiserver-76f77b778f-jxft9" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.062249 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/df36e993-8a70-42c5-ac22-508485572d35-serving-cert\") pod \"apiserver-7bbb656c7d-cpflv\" (UID: \"df36e993-8a70-42c5-ac22-508485572d35\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cpflv" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.062279 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09073ec8-36df-4ffb-a43a-10611999d8cb-config\") pod \"openshift-apiserver-operator-796bbdcf4f-8x4xt\" (UID: \"09073ec8-36df-4ffb-a43a-10611999d8cb\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8x4xt" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.062309 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hkdsn\" (UniqueName: \"kubernetes.io/projected/b7feb2dc-f1f2-4228-9974-8a771fe98f3c-kube-api-access-hkdsn\") pod \"controller-manager-879f6c89f-24dmg\" (UID: \"b7feb2dc-f1f2-4228-9974-8a771fe98f3c\") " pod="openshift-controller-manager/controller-manager-879f6c89f-24dmg" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.066927 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.067186 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.068277 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.068496 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.068668 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.068893 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.073060 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.074065 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.074451 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.074594 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.078753 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.078801 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.078909 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.079003 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.079084 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.079191 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.079338 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.079480 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.079583 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.079842 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.079942 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.079979 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.080064 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.080274 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.080428 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.080452 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.083558 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.084719 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.085236 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.086505 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.104876 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.106440 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-mj2tr"] Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.118186 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.118397 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.118862 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.119570 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-6q4dc"] Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.120041 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6q4dc" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.120316 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-mj2tr" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.121008 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.121298 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.121672 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.121920 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-87b7k"] Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.122400 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-87b7k" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.125449 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-x28lp"] Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.125928 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.126087 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-x28lp" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.126214 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-g4bql"] Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.126912 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.127036 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-g4bql" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.128248 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-h7zxk"] Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.128742 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-h7zxk" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.128739 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-rl2ff"] Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.131379 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9tkth"] Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.131728 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-n8wrd"] Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.131740 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.131809 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-rl2ff" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.132139 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.134052 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9tkth" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.135190 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-jgk9d"] Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.135681 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-ttrm2"] Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.137312 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-n8wrd" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.137421 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vc92z"] Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.137539 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-ttrm2" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.137810 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jgk9d" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.138249 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5zqnk"] Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.138703 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-l4m2j"] Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.138978 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vc92z" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.139219 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5zqnk" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.139346 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-76lnq"] Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.139704 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-l4m2j" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.139784 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fsfs7"] Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.139856 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-76lnq" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.142519 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gkg77"] Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.142547 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-2z7ml"] Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.143136 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29424000-nf7zl"] Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.143229 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-2z7ml" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.143245 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fsfs7" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.144132 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29424000-nf7zl" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.144683 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-6f4m5"] Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.145719 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-cpflv"] Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.145737 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.145813 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-6f4m5" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.146746 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-24dmg"] Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.149836 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-nc65d"] Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.150726 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-svdpj"] Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.151412 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-svdpj" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.151539 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-rvtb7"] Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.157183 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-s6qsz"] Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.158829 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-s6qsz" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.162747 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-jxft9"] Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.163585 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hkdsn\" (UniqueName: \"kubernetes.io/projected/b7feb2dc-f1f2-4228-9974-8a771fe98f3c-kube-api-access-hkdsn\") pod \"controller-manager-879f6c89f-24dmg\" (UID: \"b7feb2dc-f1f2-4228-9974-8a771fe98f3c\") " pod="openshift-controller-manager/controller-manager-879f6c89f-24dmg" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.163656 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qww7r\" (UniqueName: \"kubernetes.io/projected/e55e20fb-bef2-4fee-b6a7-60e2d9d61faa-kube-api-access-qww7r\") pod \"cluster-samples-operator-665b6dd947-z5lg5\" (UID: \"e55e20fb-bef2-4fee-b6a7-60e2d9d61faa\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-z5lg5" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.163695 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/e86ba495-65e4-4539-928e-366376df64d8-image-import-ca\") pod \"apiserver-76f77b778f-jxft9\" (UID: \"e86ba495-65e4-4539-928e-366376df64d8\") " pod="openshift-apiserver/apiserver-76f77b778f-jxft9" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.163765 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/d7d1536c-ed04-4199-b839-b94e9482a63c-oauth-serving-cert\") pod \"console-f9d7485db-rhl5v\" (UID: \"d7d1536c-ed04-4199-b839-b94e9482a63c\") " pod="openshift-console/console-f9d7485db-rhl5v" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.163807 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/cb55e751-9913-4a3c-a0b4-3e681129d052-images\") pod \"machine-api-operator-5694c8668f-nc65d\" (UID: \"cb55e751-9913-4a3c-a0b4-3e681129d052\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-nc65d" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.163831 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x9wpg\" (UniqueName: \"kubernetes.io/projected/09073ec8-36df-4ffb-a43a-10611999d8cb-kube-api-access-x9wpg\") pod \"openshift-apiserver-operator-796bbdcf4f-8x4xt\" (UID: \"09073ec8-36df-4ffb-a43a-10611999d8cb\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8x4xt" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.163857 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/e55e20fb-bef2-4fee-b6a7-60e2d9d61faa-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-z5lg5\" (UID: \"e55e20fb-bef2-4fee-b6a7-60e2d9d61faa\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-z5lg5" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.164133 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e86ba495-65e4-4539-928e-366376df64d8-config\") pod \"apiserver-76f77b778f-jxft9\" (UID: \"e86ba495-65e4-4539-928e-366376df64d8\") " pod="openshift-apiserver/apiserver-76f77b778f-jxft9" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.164221 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e86ba495-65e4-4539-928e-366376df64d8-serving-cert\") pod \"apiserver-76f77b778f-jxft9\" (UID: \"e86ba495-65e4-4539-928e-366376df64d8\") " pod="openshift-apiserver/apiserver-76f77b778f-jxft9" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.164255 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/ccbceeb2-58ab-4409-82df-5c99189569b8-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-dz7kr\" (UID: \"ccbceeb2-58ab-4409-82df-5c99189569b8\") " pod="openshift-authentication/oauth-openshift-558db77b4-dz7kr" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.164303 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/fb3441e4-e612-4cfe-9f12-6497cd9176a9-trusted-ca\") pod \"console-operator-58897d9998-b4q7q\" (UID: \"fb3441e4-e612-4cfe-9f12-6497cd9176a9\") " pod="openshift-console-operator/console-operator-58897d9998-b4q7q" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.165347 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d7d1536c-ed04-4199-b839-b94e9482a63c-trusted-ca-bundle\") pod \"console-f9d7485db-rhl5v\" (UID: \"d7d1536c-ed04-4199-b839-b94e9482a63c\") " pod="openshift-console/console-f9d7485db-rhl5v" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.165379 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/d7d1536c-ed04-4199-b839-b94e9482a63c-oauth-serving-cert\") pod \"console-f9d7485db-rhl5v\" (UID: \"d7d1536c-ed04-4199-b839-b94e9482a63c\") " pod="openshift-console/console-f9d7485db-rhl5v" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.165399 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2c04dbaf-f385-4cee-bccf-8506ecdeb680-serving-cert\") pod \"authentication-operator-69f744f599-j5kxr\" (UID: \"2c04dbaf-f385-4cee-bccf-8506ecdeb680\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-j5kxr" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.165427 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ngxkg\" (UniqueName: \"kubernetes.io/projected/2c04dbaf-f385-4cee-bccf-8506ecdeb680-kube-api-access-ngxkg\") pod \"authentication-operator-69f744f599-j5kxr\" (UID: \"2c04dbaf-f385-4cee-bccf-8506ecdeb680\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-j5kxr" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.165443 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/e86ba495-65e4-4539-928e-366376df64d8-image-import-ca\") pod \"apiserver-76f77b778f-jxft9\" (UID: \"e86ba495-65e4-4539-928e-366376df64d8\") " pod="openshift-apiserver/apiserver-76f77b778f-jxft9" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.165461 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/38005ee2-6635-498c-b156-ac57b4f2f4d9-trusted-ca\") pod \"image-registry-697d97f7c8-zf572\" (UID: \"38005ee2-6635-498c-b156-ac57b4f2f4d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-zf572" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.165548 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9xdkf\" (UniqueName: \"kubernetes.io/projected/e963dd44-250d-4467-980c-1b67ef1bdd93-kube-api-access-9xdkf\") pod \"machine-approver-56656f9798-wrqbs\" (UID: \"e963dd44-250d-4467-980c-1b67ef1bdd93\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wrqbs" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.165581 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/d7d1536c-ed04-4199-b839-b94e9482a63c-console-config\") pod \"console-f9d7485db-rhl5v\" (UID: \"d7d1536c-ed04-4199-b839-b94e9482a63c\") " pod="openshift-console/console-f9d7485db-rhl5v" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.165605 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ccbceeb2-58ab-4409-82df-5c99189569b8-audit-dir\") pod \"oauth-openshift-558db77b4-dz7kr\" (UID: \"ccbceeb2-58ab-4409-82df-5c99189569b8\") " pod="openshift-authentication/oauth-openshift-558db77b4-dz7kr" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.165626 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6hs5b\" (UniqueName: \"kubernetes.io/projected/fb3441e4-e612-4cfe-9f12-6497cd9176a9-kube-api-access-6hs5b\") pod \"console-operator-58897d9998-b4q7q\" (UID: \"fb3441e4-e612-4cfe-9f12-6497cd9176a9\") " pod="openshift-console-operator/console-operator-58897d9998-b4q7q" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.165665 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/adb430a2-20a7-4e9d-8f92-446b56e36e4f-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-q6ck8\" (UID: \"adb430a2-20a7-4e9d-8f92-446b56e36e4f\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q6ck8" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.165691 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/e86ba495-65e4-4539-928e-366376df64d8-etcd-serving-ca\") pod \"apiserver-76f77b778f-jxft9\" (UID: \"e86ba495-65e4-4539-928e-366376df64d8\") " pod="openshift-apiserver/apiserver-76f77b778f-jxft9" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.165715 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b7feb2dc-f1f2-4228-9974-8a771fe98f3c-serving-cert\") pod \"controller-manager-879f6c89f-24dmg\" (UID: \"b7feb2dc-f1f2-4228-9974-8a771fe98f3c\") " pod="openshift-controller-manager/controller-manager-879f6c89f-24dmg" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.165743 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/e86ba495-65e4-4539-928e-366376df64d8-encryption-config\") pod \"apiserver-76f77b778f-jxft9\" (UID: \"e86ba495-65e4-4539-928e-366376df64d8\") " pod="openshift-apiserver/apiserver-76f77b778f-jxft9" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.165774 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2c04dbaf-f385-4cee-bccf-8506ecdeb680-config\") pod \"authentication-operator-69f744f599-j5kxr\" (UID: \"2c04dbaf-f385-4cee-bccf-8506ecdeb680\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-j5kxr" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.165819 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zf572\" (UID: \"38005ee2-6635-498c-b156-ac57b4f2f4d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-zf572" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.165843 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d7d1536c-ed04-4199-b839-b94e9482a63c-trusted-ca-bundle\") pod \"console-f9d7485db-rhl5v\" (UID: \"d7d1536c-ed04-4199-b839-b94e9482a63c\") " pod="openshift-console/console-f9d7485db-rhl5v" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.165876 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c094602b-e1ec-4b9a-9640-cd6250a1fd94-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-gkg77\" (UID: \"c094602b-e1ec-4b9a-9640-cd6250a1fd94\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gkg77" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.165914 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/ccbceeb2-58ab-4409-82df-5c99189569b8-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-dz7kr\" (UID: \"ccbceeb2-58ab-4409-82df-5c99189569b8\") " pod="openshift-authentication/oauth-openshift-558db77b4-dz7kr" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.165954 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09073ec8-36df-4ffb-a43a-10611999d8cb-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-8x4xt\" (UID: \"09073ec8-36df-4ffb-a43a-10611999d8cb\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8x4xt" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.165988 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c465q\" (UniqueName: \"kubernetes.io/projected/d7d1536c-ed04-4199-b839-b94e9482a63c-kube-api-access-c465q\") pod \"console-f9d7485db-rhl5v\" (UID: \"d7d1536c-ed04-4199-b839-b94e9482a63c\") " pod="openshift-console/console-f9d7485db-rhl5v" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.166018 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/e84a804f-75e1-46a2-b37a-f868de516379-available-featuregates\") pod \"openshift-config-operator-7777fb866f-rvtb7\" (UID: \"e84a804f-75e1-46a2-b37a-f868de516379\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rvtb7" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.166052 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/adb430a2-20a7-4e9d-8f92-446b56e36e4f-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-q6ck8\" (UID: \"adb430a2-20a7-4e9d-8f92-446b56e36e4f\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q6ck8" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.166087 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9da229ee-dedc-417f-aec7-d2d1d9407a41-config\") pod \"kube-controller-manager-operator-78b949d7b-lbrj7\" (UID: \"9da229ee-dedc-417f-aec7-d2d1d9407a41\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lbrj7" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.166112 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/e86ba495-65e4-4539-928e-366376df64d8-node-pullsecrets\") pod \"apiserver-76f77b778f-jxft9\" (UID: \"e86ba495-65e4-4539-928e-366376df64d8\") " pod="openshift-apiserver/apiserver-76f77b778f-jxft9" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.166134 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/df36e993-8a70-42c5-ac22-508485572d35-etcd-client\") pod \"apiserver-7bbb656c7d-cpflv\" (UID: \"df36e993-8a70-42c5-ac22-508485572d35\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cpflv" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.166158 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e86ba495-65e4-4539-928e-366376df64d8-etcd-client\") pod \"apiserver-76f77b778f-jxft9\" (UID: \"e86ba495-65e4-4539-928e-366376df64d8\") " pod="openshift-apiserver/apiserver-76f77b778f-jxft9" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.166179 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d7d1536c-ed04-4199-b839-b94e9482a63c-service-ca\") pod \"console-f9d7485db-rhl5v\" (UID: \"d7d1536c-ed04-4199-b839-b94e9482a63c\") " pod="openshift-console/console-f9d7485db-rhl5v" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.166199 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cb55e751-9913-4a3c-a0b4-3e681129d052-config\") pod \"machine-api-operator-5694c8668f-nc65d\" (UID: \"cb55e751-9913-4a3c-a0b4-3e681129d052\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-nc65d" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.166224 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/e86ba495-65e4-4539-928e-366376df64d8-audit\") pod \"apiserver-76f77b778f-jxft9\" (UID: \"e86ba495-65e4-4539-928e-366376df64d8\") " pod="openshift-apiserver/apiserver-76f77b778f-jxft9" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.166252 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9da229ee-dedc-417f-aec7-d2d1d9407a41-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-lbrj7\" (UID: \"9da229ee-dedc-417f-aec7-d2d1d9407a41\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lbrj7" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.166294 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/d7d1536c-ed04-4199-b839-b94e9482a63c-console-oauth-config\") pod \"console-f9d7485db-rhl5v\" (UID: \"d7d1536c-ed04-4199-b839-b94e9482a63c\") " pod="openshift-console/console-f9d7485db-rhl5v" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.166322 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/df36e993-8a70-42c5-ac22-508485572d35-audit-dir\") pod \"apiserver-7bbb656c7d-cpflv\" (UID: \"df36e993-8a70-42c5-ac22-508485572d35\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cpflv" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.166353 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/ccbceeb2-58ab-4409-82df-5c99189569b8-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-dz7kr\" (UID: \"ccbceeb2-58ab-4409-82df-5c99189569b8\") " pod="openshift-authentication/oauth-openshift-558db77b4-dz7kr" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.166375 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m96cx\" (UniqueName: \"kubernetes.io/projected/cb55e751-9913-4a3c-a0b4-3e681129d052-kube-api-access-m96cx\") pod \"machine-api-operator-5694c8668f-nc65d\" (UID: \"cb55e751-9913-4a3c-a0b4-3e681129d052\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-nc65d" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.166397 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2c04dbaf-f385-4cee-bccf-8506ecdeb680-service-ca-bundle\") pod \"authentication-operator-69f744f599-j5kxr\" (UID: \"2c04dbaf-f385-4cee-bccf-8506ecdeb680\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-j5kxr" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.166416 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/38005ee2-6635-498c-b156-ac57b4f2f4d9-bound-sa-token\") pod \"image-registry-697d97f7c8-zf572\" (UID: \"38005ee2-6635-498c-b156-ac57b4f2f4d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-zf572" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.166445 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b7feb2dc-f1f2-4228-9974-8a771fe98f3c-client-ca\") pod \"controller-manager-879f6c89f-24dmg\" (UID: \"b7feb2dc-f1f2-4228-9974-8a771fe98f3c\") " pod="openshift-controller-manager/controller-manager-879f6c89f-24dmg" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.166474 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/d7d1536c-ed04-4199-b839-b94e9482a63c-console-serving-cert\") pod \"console-f9d7485db-rhl5v\" (UID: \"d7d1536c-ed04-4199-b839-b94e9482a63c\") " pod="openshift-console/console-f9d7485db-rhl5v" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.166497 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e86ba495-65e4-4539-928e-366376df64d8-config\") pod \"apiserver-76f77b778f-jxft9\" (UID: \"e86ba495-65e4-4539-928e-366376df64d8\") " pod="openshift-apiserver/apiserver-76f77b778f-jxft9" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.166500 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/df36e993-8a70-42c5-ac22-508485572d35-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-cpflv\" (UID: \"df36e993-8a70-42c5-ac22-508485572d35\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cpflv" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.166575 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e963dd44-250d-4467-980c-1b67ef1bdd93-config\") pod \"machine-approver-56656f9798-wrqbs\" (UID: \"e963dd44-250d-4467-980c-1b67ef1bdd93\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wrqbs" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.166601 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/38005ee2-6635-498c-b156-ac57b4f2f4d9-registry-tls\") pod \"image-registry-697d97f7c8-zf572\" (UID: \"38005ee2-6635-498c-b156-ac57b4f2f4d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-zf572" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.166675 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/df36e993-8a70-42c5-ac22-508485572d35-audit-policies\") pod \"apiserver-7bbb656c7d-cpflv\" (UID: \"df36e993-8a70-42c5-ac22-508485572d35\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cpflv" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.166801 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-67q4s\" (UniqueName: \"kubernetes.io/projected/e84a804f-75e1-46a2-b37a-f868de516379-kube-api-access-67q4s\") pod \"openshift-config-operator-7777fb866f-rvtb7\" (UID: \"e84a804f-75e1-46a2-b37a-f868de516379\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rvtb7" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.166874 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/ccbceeb2-58ab-4409-82df-5c99189569b8-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-dz7kr\" (UID: \"ccbceeb2-58ab-4409-82df-5c99189569b8\") " pod="openshift-authentication/oauth-openshift-558db77b4-dz7kr" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.167367 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/df36e993-8a70-42c5-ac22-508485572d35-audit-policies\") pod \"apiserver-7bbb656c7d-cpflv\" (UID: \"df36e993-8a70-42c5-ac22-508485572d35\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cpflv" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.167585 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-rhl5v"] Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.167981 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ccbceeb2-58ab-4409-82df-5c99189569b8-audit-dir\") pod \"oauth-openshift-558db77b4-dz7kr\" (UID: \"ccbceeb2-58ab-4409-82df-5c99189569b8\") " pod="openshift-authentication/oauth-openshift-558db77b4-dz7kr" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.168659 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/df36e993-8a70-42c5-ac22-508485572d35-audit-dir\") pod \"apiserver-7bbb656c7d-cpflv\" (UID: \"df36e993-8a70-42c5-ac22-508485572d35\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cpflv" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.168809 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9da229ee-dedc-417f-aec7-d2d1d9407a41-config\") pod \"kube-controller-manager-operator-78b949d7b-lbrj7\" (UID: \"9da229ee-dedc-417f-aec7-d2d1d9407a41\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lbrj7" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.168875 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/d7d1536c-ed04-4199-b839-b94e9482a63c-console-config\") pod \"console-f9d7485db-rhl5v\" (UID: \"d7d1536c-ed04-4199-b839-b94e9482a63c\") " pod="openshift-console/console-f9d7485db-rhl5v" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.169063 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/ccbceeb2-58ab-4409-82df-5c99189569b8-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-dz7kr\" (UID: \"ccbceeb2-58ab-4409-82df-5c99189569b8\") " pod="openshift-authentication/oauth-openshift-558db77b4-dz7kr" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.169279 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2c04dbaf-f385-4cee-bccf-8506ecdeb680-service-ca-bundle\") pod \"authentication-operator-69f744f599-j5kxr\" (UID: \"2c04dbaf-f385-4cee-bccf-8506ecdeb680\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-j5kxr" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.169359 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/ccbceeb2-58ab-4409-82df-5c99189569b8-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-dz7kr\" (UID: \"ccbceeb2-58ab-4409-82df-5c99189569b8\") " pod="openshift-authentication/oauth-openshift-558db77b4-dz7kr" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.170030 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c094602b-e1ec-4b9a-9640-cd6250a1fd94-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-gkg77\" (UID: \"c094602b-e1ec-4b9a-9640-cd6250a1fd94\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gkg77" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.170031 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b7feb2dc-f1f2-4228-9974-8a771fe98f3c-client-ca\") pod \"controller-manager-879f6c89f-24dmg\" (UID: \"b7feb2dc-f1f2-4228-9974-8a771fe98f3c\") " pod="openshift-controller-manager/controller-manager-879f6c89f-24dmg" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.170434 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/e86ba495-65e4-4539-928e-366376df64d8-etcd-serving-ca\") pod \"apiserver-76f77b778f-jxft9\" (UID: \"e86ba495-65e4-4539-928e-366376df64d8\") " pod="openshift-apiserver/apiserver-76f77b778f-jxft9" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.170615 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2c04dbaf-f385-4cee-bccf-8506ecdeb680-serving-cert\") pod \"authentication-operator-69f744f599-j5kxr\" (UID: \"2c04dbaf-f385-4cee-bccf-8506ecdeb680\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-j5kxr" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.171172 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2c04dbaf-f385-4cee-bccf-8506ecdeb680-config\") pod \"authentication-operator-69f744f599-j5kxr\" (UID: \"2c04dbaf-f385-4cee-bccf-8506ecdeb680\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-j5kxr" Dec 11 08:13:34 crc kubenswrapper[4860]: E1211 08:13:34.171220 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:13:34.671205732 +0000 UTC m=+147.399724887 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zf572" (UID: "38005ee2-6635-498c-b156-ac57b4f2f4d9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.171376 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/e86ba495-65e4-4539-928e-366376df64d8-node-pullsecrets\") pod \"apiserver-76f77b778f-jxft9\" (UID: \"e86ba495-65e4-4539-928e-366376df64d8\") " pod="openshift-apiserver/apiserver-76f77b778f-jxft9" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.171767 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d7d1536c-ed04-4199-b839-b94e9482a63c-service-ca\") pod \"console-f9d7485db-rhl5v\" (UID: \"d7d1536c-ed04-4199-b839-b94e9482a63c\") " pod="openshift-console/console-f9d7485db-rhl5v" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.171848 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/e84a804f-75e1-46a2-b37a-f868de516379-available-featuregates\") pod \"openshift-config-operator-7777fb866f-rvtb7\" (UID: \"e84a804f-75e1-46a2-b37a-f868de516379\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rvtb7" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.172765 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-qx9pf"] Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.173069 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/ccbceeb2-58ab-4409-82df-5c99189569b8-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-dz7kr\" (UID: \"ccbceeb2-58ab-4409-82df-5c99189569b8\") " pod="openshift-authentication/oauth-openshift-558db77b4-dz7kr" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.173422 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-qx9pf" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.173450 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09073ec8-36df-4ffb-a43a-10611999d8cb-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-8x4xt\" (UID: \"09073ec8-36df-4ffb-a43a-10611999d8cb\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8x4xt" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.173526 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/e86ba495-65e4-4539-928e-366376df64d8-audit\") pod \"apiserver-76f77b778f-jxft9\" (UID: \"e86ba495-65e4-4539-928e-366376df64d8\") " pod="openshift-apiserver/apiserver-76f77b778f-jxft9" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.174206 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cb55e751-9913-4a3c-a0b4-3e681129d052-config\") pod \"machine-api-operator-5694c8668f-nc65d\" (UID: \"cb55e751-9913-4a3c-a0b4-3e681129d052\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-nc65d" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.174238 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/cb55e751-9913-4a3c-a0b4-3e681129d052-images\") pod \"machine-api-operator-5694c8668f-nc65d\" (UID: \"cb55e751-9913-4a3c-a0b4-3e681129d052\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-nc65d" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.167674 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-24srs\" (UniqueName: \"kubernetes.io/projected/38005ee2-6635-498c-b156-ac57b4f2f4d9-kube-api-access-24srs\") pod \"image-registry-697d97f7c8-zf572\" (UID: \"38005ee2-6635-498c-b156-ac57b4f2f4d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-zf572" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.174522 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9da229ee-dedc-417f-aec7-d2d1d9407a41-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-lbrj7\" (UID: \"9da229ee-dedc-417f-aec7-d2d1d9407a41\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lbrj7" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.174544 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fb3441e4-e612-4cfe-9f12-6497cd9176a9-config\") pod \"console-operator-58897d9998-b4q7q\" (UID: \"fb3441e4-e612-4cfe-9f12-6497cd9176a9\") " pod="openshift-console-operator/console-operator-58897d9998-b4q7q" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.174562 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e84a804f-75e1-46a2-b37a-f868de516379-serving-cert\") pod \"openshift-config-operator-7777fb866f-rvtb7\" (UID: \"e84a804f-75e1-46a2-b37a-f868de516379\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rvtb7" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.174591 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/adb430a2-20a7-4e9d-8f92-446b56e36e4f-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-q6ck8\" (UID: \"adb430a2-20a7-4e9d-8f92-446b56e36e4f\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q6ck8" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.174627 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e86ba495-65e4-4539-928e-366376df64d8-audit-dir\") pod \"apiserver-76f77b778f-jxft9\" (UID: \"e86ba495-65e4-4539-928e-366376df64d8\") " pod="openshift-apiserver/apiserver-76f77b778f-jxft9" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.174670 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/cb55e751-9913-4a3c-a0b4-3e681129d052-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-nc65d\" (UID: \"cb55e751-9913-4a3c-a0b4-3e681129d052\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-nc65d" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.174697 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/38005ee2-6635-498c-b156-ac57b4f2f4d9-registry-certificates\") pod \"image-registry-697d97f7c8-zf572\" (UID: \"38005ee2-6635-498c-b156-ac57b4f2f4d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-zf572" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.174715 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ccbceeb2-58ab-4409-82df-5c99189569b8-audit-policies\") pod \"oauth-openshift-558db77b4-dz7kr\" (UID: \"ccbceeb2-58ab-4409-82df-5c99189569b8\") " pod="openshift-authentication/oauth-openshift-558db77b4-dz7kr" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.174734 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/38005ee2-6635-498c-b156-ac57b4f2f4d9-ca-trust-extracted\") pod \"image-registry-697d97f7c8-zf572\" (UID: \"38005ee2-6635-498c-b156-ac57b4f2f4d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-zf572" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.174761 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/ccbceeb2-58ab-4409-82df-5c99189569b8-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-dz7kr\" (UID: \"ccbceeb2-58ab-4409-82df-5c99189569b8\") " pod="openshift-authentication/oauth-openshift-558db77b4-dz7kr" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.174781 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/fb3441e4-e612-4cfe-9f12-6497cd9176a9-trusted-ca\") pod \"console-operator-58897d9998-b4q7q\" (UID: \"fb3441e4-e612-4cfe-9f12-6497cd9176a9\") " pod="openshift-console-operator/console-operator-58897d9998-b4q7q" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.174788 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/ccbceeb2-58ab-4409-82df-5c99189569b8-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-dz7kr\" (UID: \"ccbceeb2-58ab-4409-82df-5c99189569b8\") " pod="openshift-authentication/oauth-openshift-558db77b4-dz7kr" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.174807 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c094602b-e1ec-4b9a-9640-cd6250a1fd94-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-gkg77\" (UID: \"c094602b-e1ec-4b9a-9640-cd6250a1fd94\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gkg77" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.174828 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/e963dd44-250d-4467-980c-1b67ef1bdd93-machine-approver-tls\") pod \"machine-approver-56656f9798-wrqbs\" (UID: \"e963dd44-250d-4467-980c-1b67ef1bdd93\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wrqbs" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.174833 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e86ba495-65e4-4539-928e-366376df64d8-audit-dir\") pod \"apiserver-76f77b778f-jxft9\" (UID: \"e86ba495-65e4-4539-928e-366376df64d8\") " pod="openshift-apiserver/apiserver-76f77b778f-jxft9" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.174854 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/ccbceeb2-58ab-4409-82df-5c99189569b8-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-dz7kr\" (UID: \"ccbceeb2-58ab-4409-82df-5c99189569b8\") " pod="openshift-authentication/oauth-openshift-558db77b4-dz7kr" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.174877 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/ccbceeb2-58ab-4409-82df-5c99189569b8-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-dz7kr\" (UID: \"ccbceeb2-58ab-4409-82df-5c99189569b8\") " pod="openshift-authentication/oauth-openshift-558db77b4-dz7kr" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.174898 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fb3441e4-e612-4cfe-9f12-6497cd9176a9-serving-cert\") pod \"console-operator-58897d9998-b4q7q\" (UID: \"fb3441e4-e612-4cfe-9f12-6497cd9176a9\") " pod="openshift-console-operator/console-operator-58897d9998-b4q7q" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.174923 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rkfd4\" (UniqueName: \"kubernetes.io/projected/c094602b-e1ec-4b9a-9640-cd6250a1fd94-kube-api-access-rkfd4\") pod \"openshift-controller-manager-operator-756b6f6bc6-gkg77\" (UID: \"c094602b-e1ec-4b9a-9640-cd6250a1fd94\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gkg77" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.174947 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e963dd44-250d-4467-980c-1b67ef1bdd93-auth-proxy-config\") pod \"machine-approver-56656f9798-wrqbs\" (UID: \"e963dd44-250d-4467-980c-1b67ef1bdd93\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wrqbs" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.175004 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8k4gt\" (UniqueName: \"kubernetes.io/projected/ccbceeb2-58ab-4409-82df-5c99189569b8-kube-api-access-8k4gt\") pod \"oauth-openshift-558db77b4-dz7kr\" (UID: \"ccbceeb2-58ab-4409-82df-5c99189569b8\") " pod="openshift-authentication/oauth-openshift-558db77b4-dz7kr" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.175031 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/38005ee2-6635-498c-b156-ac57b4f2f4d9-installation-pull-secrets\") pod \"image-registry-697d97f7c8-zf572\" (UID: \"38005ee2-6635-498c-b156-ac57b4f2f4d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-zf572" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.175111 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fp9sr\" (UniqueName: \"kubernetes.io/projected/adb430a2-20a7-4e9d-8f92-446b56e36e4f-kube-api-access-fp9sr\") pod \"cluster-image-registry-operator-dc59b4c8b-q6ck8\" (UID: \"adb430a2-20a7-4e9d-8f92-446b56e36e4f\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q6ck8" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.175143 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/df36e993-8a70-42c5-ac22-508485572d35-encryption-config\") pod \"apiserver-7bbb656c7d-cpflv\" (UID: \"df36e993-8a70-42c5-ac22-508485572d35\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cpflv" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.175626 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b7feb2dc-f1f2-4228-9974-8a771fe98f3c-config\") pod \"controller-manager-879f6c89f-24dmg\" (UID: \"b7feb2dc-f1f2-4228-9974-8a771fe98f3c\") " pod="openshift-controller-manager/controller-manager-879f6c89f-24dmg" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.175690 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e86ba495-65e4-4539-928e-366376df64d8-trusted-ca-bundle\") pod \"apiserver-76f77b778f-jxft9\" (UID: \"e86ba495-65e4-4539-928e-366376df64d8\") " pod="openshift-apiserver/apiserver-76f77b778f-jxft9" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.175743 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/ccbceeb2-58ab-4409-82df-5c99189569b8-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-dz7kr\" (UID: \"ccbceeb2-58ab-4409-82df-5c99189569b8\") " pod="openshift-authentication/oauth-openshift-558db77b4-dz7kr" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.175774 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.175808 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/df36e993-8a70-42c5-ac22-508485572d35-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-cpflv\" (UID: \"df36e993-8a70-42c5-ac22-508485572d35\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cpflv" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.175858 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/ccbceeb2-58ab-4409-82df-5c99189569b8-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-dz7kr\" (UID: \"ccbceeb2-58ab-4409-82df-5c99189569b8\") " pod="openshift-authentication/oauth-openshift-558db77b4-dz7kr" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.175885 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mtwdk\" (UniqueName: \"kubernetes.io/projected/df36e993-8a70-42c5-ac22-508485572d35-kube-api-access-mtwdk\") pod \"apiserver-7bbb656c7d-cpflv\" (UID: \"df36e993-8a70-42c5-ac22-508485572d35\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cpflv" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.175933 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ccbceeb2-58ab-4409-82df-5c99189569b8-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-dz7kr\" (UID: \"ccbceeb2-58ab-4409-82df-5c99189569b8\") " pod="openshift-authentication/oauth-openshift-558db77b4-dz7kr" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.175959 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2c04dbaf-f385-4cee-bccf-8506ecdeb680-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-j5kxr\" (UID: \"2c04dbaf-f385-4cee-bccf-8506ecdeb680\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-j5kxr" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.176012 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b7feb2dc-f1f2-4228-9974-8a771fe98f3c-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-24dmg\" (UID: \"b7feb2dc-f1f2-4228-9974-8a771fe98f3c\") " pod="openshift-controller-manager/controller-manager-879f6c89f-24dmg" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.176036 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x629f\" (UniqueName: \"kubernetes.io/projected/e86ba495-65e4-4539-928e-366376df64d8-kube-api-access-x629f\") pod \"apiserver-76f77b778f-jxft9\" (UID: \"e86ba495-65e4-4539-928e-366376df64d8\") " pod="openshift-apiserver/apiserver-76f77b778f-jxft9" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.176059 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/df36e993-8a70-42c5-ac22-508485572d35-serving-cert\") pod \"apiserver-7bbb656c7d-cpflv\" (UID: \"df36e993-8a70-42c5-ac22-508485572d35\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cpflv" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.176084 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09073ec8-36df-4ffb-a43a-10611999d8cb-config\") pod \"openshift-apiserver-operator-796bbdcf4f-8x4xt\" (UID: \"09073ec8-36df-4ffb-a43a-10611999d8cb\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8x4xt" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.176539 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e86ba495-65e4-4539-928e-366376df64d8-serving-cert\") pod \"apiserver-76f77b778f-jxft9\" (UID: \"e86ba495-65e4-4539-928e-366376df64d8\") " pod="openshift-apiserver/apiserver-76f77b778f-jxft9" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.176685 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09073ec8-36df-4ffb-a43a-10611999d8cb-config\") pod \"openshift-apiserver-operator-796bbdcf4f-8x4xt\" (UID: \"09073ec8-36df-4ffb-a43a-10611999d8cb\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8x4xt" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.178557 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fb3441e4-e612-4cfe-9f12-6497cd9176a9-config\") pod \"console-operator-58897d9998-b4q7q\" (UID: \"fb3441e4-e612-4cfe-9f12-6497cd9176a9\") " pod="openshift-console-operator/console-operator-58897d9998-b4q7q" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.167715 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/df36e993-8a70-42c5-ac22-508485572d35-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-cpflv\" (UID: \"df36e993-8a70-42c5-ac22-508485572d35\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cpflv" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.175504 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/e55e20fb-bef2-4fee-b6a7-60e2d9d61faa-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-z5lg5\" (UID: \"e55e20fb-bef2-4fee-b6a7-60e2d9d61faa\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-z5lg5" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.175431 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b7feb2dc-f1f2-4228-9974-8a771fe98f3c-serving-cert\") pod \"controller-manager-879f6c89f-24dmg\" (UID: \"b7feb2dc-f1f2-4228-9974-8a771fe98f3c\") " pod="openshift-controller-manager/controller-manager-879f6c89f-24dmg" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.179430 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/cb55e751-9913-4a3c-a0b4-3e681129d052-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-nc65d\" (UID: \"cb55e751-9913-4a3c-a0b4-3e681129d052\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-nc65d" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.179942 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ccbceeb2-58ab-4409-82df-5c99189569b8-audit-policies\") pod \"oauth-openshift-558db77b4-dz7kr\" (UID: \"ccbceeb2-58ab-4409-82df-5c99189569b8\") " pod="openshift-authentication/oauth-openshift-558db77b4-dz7kr" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.179989 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b7feb2dc-f1f2-4228-9974-8a771fe98f3c-config\") pod \"controller-manager-879f6c89f-24dmg\" (UID: \"b7feb2dc-f1f2-4228-9974-8a771fe98f3c\") " pod="openshift-controller-manager/controller-manager-879f6c89f-24dmg" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.180458 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e84a804f-75e1-46a2-b37a-f868de516379-serving-cert\") pod \"openshift-config-operator-7777fb866f-rvtb7\" (UID: \"e84a804f-75e1-46a2-b37a-f868de516379\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rvtb7" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.180542 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9da229ee-dedc-417f-aec7-d2d1d9407a41-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-lbrj7\" (UID: \"9da229ee-dedc-417f-aec7-d2d1d9407a41\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lbrj7" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.180696 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-j5kxr"] Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.181000 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/ccbceeb2-58ab-4409-82df-5c99189569b8-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-dz7kr\" (UID: \"ccbceeb2-58ab-4409-82df-5c99189569b8\") " pod="openshift-authentication/oauth-openshift-558db77b4-dz7kr" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.181435 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2c04dbaf-f385-4cee-bccf-8506ecdeb680-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-j5kxr\" (UID: \"2c04dbaf-f385-4cee-bccf-8506ecdeb680\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-j5kxr" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.181839 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e86ba495-65e4-4539-928e-366376df64d8-trusted-ca-bundle\") pod \"apiserver-76f77b778f-jxft9\" (UID: \"e86ba495-65e4-4539-928e-366376df64d8\") " pod="openshift-apiserver/apiserver-76f77b778f-jxft9" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.182123 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ccbceeb2-58ab-4409-82df-5c99189569b8-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-dz7kr\" (UID: \"ccbceeb2-58ab-4409-82df-5c99189569b8\") " pod="openshift-authentication/oauth-openshift-558db77b4-dz7kr" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.182792 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/df36e993-8a70-42c5-ac22-508485572d35-encryption-config\") pod \"apiserver-7bbb656c7d-cpflv\" (UID: \"df36e993-8a70-42c5-ac22-508485572d35\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cpflv" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.182916 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-b4q7q"] Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.183161 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fb3441e4-e612-4cfe-9f12-6497cd9176a9-serving-cert\") pod \"console-operator-58897d9998-b4q7q\" (UID: \"fb3441e4-e612-4cfe-9f12-6497cd9176a9\") " pod="openshift-console-operator/console-operator-58897d9998-b4q7q" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.184033 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b7feb2dc-f1f2-4228-9974-8a771fe98f3c-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-24dmg\" (UID: \"b7feb2dc-f1f2-4228-9974-8a771fe98f3c\") " pod="openshift-controller-manager/controller-manager-879f6c89f-24dmg" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.184035 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/df36e993-8a70-42c5-ac22-508485572d35-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-cpflv\" (UID: \"df36e993-8a70-42c5-ac22-508485572d35\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cpflv" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.184395 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c094602b-e1ec-4b9a-9640-cd6250a1fd94-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-gkg77\" (UID: \"c094602b-e1ec-4b9a-9640-cd6250a1fd94\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gkg77" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.184436 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q6ck8"] Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.184613 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/ccbceeb2-58ab-4409-82df-5c99189569b8-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-dz7kr\" (UID: \"ccbceeb2-58ab-4409-82df-5c99189569b8\") " pod="openshift-authentication/oauth-openshift-558db77b4-dz7kr" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.184923 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/ccbceeb2-58ab-4409-82df-5c99189569b8-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-dz7kr\" (UID: \"ccbceeb2-58ab-4409-82df-5c99189569b8\") " pod="openshift-authentication/oauth-openshift-558db77b4-dz7kr" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.186007 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-87b7k"] Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.186409 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/ccbceeb2-58ab-4409-82df-5c99189569b8-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-dz7kr\" (UID: \"ccbceeb2-58ab-4409-82df-5c99189569b8\") " pod="openshift-authentication/oauth-openshift-558db77b4-dz7kr" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.186528 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/ccbceeb2-58ab-4409-82df-5c99189569b8-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-dz7kr\" (UID: \"ccbceeb2-58ab-4409-82df-5c99189569b8\") " pod="openshift-authentication/oauth-openshift-558db77b4-dz7kr" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.186610 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/ccbceeb2-58ab-4409-82df-5c99189569b8-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-dz7kr\" (UID: \"ccbceeb2-58ab-4409-82df-5c99189569b8\") " pod="openshift-authentication/oauth-openshift-558db77b4-dz7kr" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.186661 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-z5lg5"] Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.186789 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.188261 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/ccbceeb2-58ab-4409-82df-5c99189569b8-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-dz7kr\" (UID: \"ccbceeb2-58ab-4409-82df-5c99189569b8\") " pod="openshift-authentication/oauth-openshift-558db77b4-dz7kr" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.190880 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-zf572"] Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.190911 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-g4bql"] Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.192220 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-6q4dc"] Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.193233 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-rl2ff"] Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.193771 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lbrj7"] Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.194751 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/df36e993-8a70-42c5-ac22-508485572d35-serving-cert\") pod \"apiserver-7bbb656c7d-cpflv\" (UID: \"df36e993-8a70-42c5-ac22-508485572d35\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cpflv" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.194843 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-dz7kr"] Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.195570 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/e86ba495-65e4-4539-928e-366376df64d8-encryption-config\") pod \"apiserver-76f77b778f-jxft9\" (UID: \"e86ba495-65e4-4539-928e-366376df64d8\") " pod="openshift-apiserver/apiserver-76f77b778f-jxft9" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.195724 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/d7d1536c-ed04-4199-b839-b94e9482a63c-console-oauth-config\") pod \"console-f9d7485db-rhl5v\" (UID: \"d7d1536c-ed04-4199-b839-b94e9482a63c\") " pod="openshift-console/console-f9d7485db-rhl5v" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.196609 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-mj2tr"] Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.197780 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/d7d1536c-ed04-4199-b839-b94e9482a63c-console-serving-cert\") pod \"console-f9d7485db-rhl5v\" (UID: \"d7d1536c-ed04-4199-b839-b94e9482a63c\") " pod="openshift-console/console-f9d7485db-rhl5v" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.197846 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-h7zxk"] Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.199304 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-jgk9d"] Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.200935 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-x28lp"] Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.201090 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e86ba495-65e4-4539-928e-366376df64d8-etcd-client\") pod \"apiserver-76f77b778f-jxft9\" (UID: \"e86ba495-65e4-4539-928e-366376df64d8\") " pod="openshift-apiserver/apiserver-76f77b778f-jxft9" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.201478 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8x4xt"] Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.202045 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/df36e993-8a70-42c5-ac22-508485572d35-etcd-client\") pod \"apiserver-7bbb656c7d-cpflv\" (UID: \"df36e993-8a70-42c5-ac22-508485572d35\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cpflv" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.203021 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-kbz2w"] Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.204303 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29424000-nf7zl"] Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.204440 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-kbz2w" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.205388 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9tkth"] Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.206519 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-l4m2j"] Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.207487 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-2z7ml"] Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.208160 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.208526 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vc92z"] Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.209566 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-76lnq"] Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.210553 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-sbcxd"] Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.211397 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-sbcxd" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.211789 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-qx9pf"] Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.212859 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-6f4m5"] Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.216190 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-svdpj"] Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.216219 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fsfs7"] Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.216232 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-s6qsz"] Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.217678 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-kbz2w"] Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.220072 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-ttrm2"] Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.222224 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5zqnk"] Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.223324 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-t8ffw"] Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.223932 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-t8ffw" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.224754 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-9wb7z"] Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.225236 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-9wb7z" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.226514 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.228810 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-sbcxd"] Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.230007 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-t8ffw"] Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.247231 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.265825 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.276490 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.276706 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/38005ee2-6635-498c-b156-ac57b4f2f4d9-installation-pull-secrets\") pod \"image-registry-697d97f7c8-zf572\" (UID: \"38005ee2-6635-498c-b156-ac57b4f2f4d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-zf572" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.276739 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/ad724960-18d1-4e83-944c-4cdeb033a436-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-ttrm2\" (UID: \"ad724960-18d1-4e83-944c-4cdeb033a436\") " pod="openshift-marketplace/marketplace-operator-79b997595-ttrm2" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.276765 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/cefdf543-6d3a-4e62-9875-ca26b060f437-auth-proxy-config\") pod \"machine-config-operator-74547568cd-6q4dc\" (UID: \"cefdf543-6d3a-4e62-9875-ca26b060f437\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6q4dc" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.276801 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmgfp\" (UniqueName: \"kubernetes.io/projected/73ab8adf-cf47-4482-bbdb-8fb6c3edc87a-kube-api-access-vmgfp\") pod \"route-controller-manager-6576b87f9c-87b7k\" (UID: \"73ab8adf-cf47-4482-bbdb-8fb6c3edc87a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-87b7k" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.276824 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g6g9q\" (UniqueName: \"kubernetes.io/projected/b08141e2-41b8-45f4-bb40-c319621719c8-kube-api-access-g6g9q\") pod \"ingress-operator-5b745b69d9-h7zxk\" (UID: \"b08141e2-41b8-45f4-bb40-c319621719c8\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-h7zxk" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.276853 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pt56k\" (UniqueName: \"kubernetes.io/projected/220b6958-d8ee-4733-a93a-35d2ced9f150-kube-api-access-pt56k\") pod \"dns-default-sbcxd\" (UID: \"220b6958-d8ee-4733-a93a-35d2ced9f150\") " pod="openshift-dns/dns-default-sbcxd" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.276873 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/30bec2ef-4635-4766-a34c-423c2e7f471b-metrics-certs\") pod \"router-default-5444994796-n8wrd\" (UID: \"30bec2ef-4635-4766-a34c-423c2e7f471b\") " pod="openshift-ingress/router-default-5444994796-n8wrd" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.276916 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bmjqc\" (UniqueName: \"kubernetes.io/projected/5c0bfb74-4985-47a1-ade8-17afad50fabe-kube-api-access-bmjqc\") pod \"machine-config-controller-84d6567774-jgk9d\" (UID: \"5c0bfb74-4985-47a1-ade8-17afad50fabe\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jgk9d" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.276942 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-grc8b\" (UniqueName: \"kubernetes.io/projected/93550904-9295-4c4a-98bb-7a228c2699bb-kube-api-access-grc8b\") pod \"machine-config-server-9wb7z\" (UID: \"93550904-9295-4c4a-98bb-7a228c2699bb\") " pod="openshift-machine-config-operator/machine-config-server-9wb7z" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.276958 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3e5d90ab-884d-420b-a284-74f4b578a864-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-5zqnk\" (UID: \"3e5d90ab-884d-420b-a284-74f4b578a864\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5zqnk" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.276979 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9xdkf\" (UniqueName: \"kubernetes.io/projected/e963dd44-250d-4467-980c-1b67ef1bdd93-kube-api-access-9xdkf\") pod \"machine-approver-56656f9798-wrqbs\" (UID: \"e963dd44-250d-4467-980c-1b67ef1bdd93\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wrqbs" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.276997 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/adb430a2-20a7-4e9d-8f92-446b56e36e4f-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-q6ck8\" (UID: \"adb430a2-20a7-4e9d-8f92-446b56e36e4f\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q6ck8" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.277332 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/220b6958-d8ee-4733-a93a-35d2ced9f150-metrics-tls\") pod \"dns-default-sbcxd\" (UID: \"220b6958-d8ee-4733-a93a-35d2ced9f150\") " pod="openshift-dns/dns-default-sbcxd" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.277428 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m8497\" (UniqueName: \"kubernetes.io/projected/0f6c4496-7751-49db-adc2-18099589c708-kube-api-access-m8497\") pod \"collect-profiles-29424000-nf7zl\" (UID: \"0f6c4496-7751-49db-adc2-18099589c708\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424000-nf7zl" Dec 11 08:13:34 crc kubenswrapper[4860]: E1211 08:13:34.277470 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:13:34.777441305 +0000 UTC m=+147.505960360 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.277495 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/1fff60d5-55fd-41d8-9225-251277e8e3c2-metrics-tls\") pod \"dns-operator-744455d44c-g4bql\" (UID: \"1fff60d5-55fd-41d8-9225-251277e8e3c2\") " pod="openshift-dns-operator/dns-operator-744455d44c-g4bql" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.277853 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/30bec2ef-4635-4766-a34c-423c2e7f471b-stats-auth\") pod \"router-default-5444994796-n8wrd\" (UID: \"30bec2ef-4635-4766-a34c-423c2e7f471b\") " pod="openshift-ingress/router-default-5444994796-n8wrd" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.277897 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h6gns\" (UniqueName: \"kubernetes.io/projected/c3ce3c34-9b72-4964-b59a-f6b3c0a0e5c8-kube-api-access-h6gns\") pod \"kube-storage-version-migrator-operator-b67b599dd-x28lp\" (UID: \"c3ce3c34-9b72-4964-b59a-f6b3c0a0e5c8\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-x28lp" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.277935 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b08141e2-41b8-45f4-bb40-c319621719c8-trusted-ca\") pod \"ingress-operator-5b745b69d9-h7zxk\" (UID: \"b08141e2-41b8-45f4-bb40-c319621719c8\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-h7zxk" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.277971 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c3ce3c34-9b72-4964-b59a-f6b3c0a0e5c8-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-x28lp\" (UID: \"c3ce3c34-9b72-4964-b59a-f6b3c0a0e5c8\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-x28lp" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.278024 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/93550904-9295-4c4a-98bb-7a228c2699bb-certs\") pod \"machine-config-server-9wb7z\" (UID: \"93550904-9295-4c4a-98bb-7a228c2699bb\") " pod="openshift-machine-config-operator/machine-config-server-9wb7z" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.278051 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/eceb74a4-c695-46e9-ba5f-b5145d20d968-plugins-dir\") pod \"csi-hostpathplugin-kbz2w\" (UID: \"eceb74a4-c695-46e9-ba5f-b5145d20d968\") " pod="hostpath-provisioner/csi-hostpathplugin-kbz2w" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.278092 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/38005ee2-6635-498c-b156-ac57b4f2f4d9-bound-sa-token\") pod \"image-registry-697d97f7c8-zf572\" (UID: \"38005ee2-6635-498c-b156-ac57b4f2f4d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-zf572" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.278134 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/73ab8adf-cf47-4482-bbdb-8fb6c3edc87a-serving-cert\") pod \"route-controller-manager-6576b87f9c-87b7k\" (UID: \"73ab8adf-cf47-4482-bbdb-8fb6c3edc87a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-87b7k" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.278293 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/73ab8adf-cf47-4482-bbdb-8fb6c3edc87a-client-ca\") pod \"route-controller-manager-6576b87f9c-87b7k\" (UID: \"73ab8adf-cf47-4482-bbdb-8fb6c3edc87a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-87b7k" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.278337 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/30bec2ef-4635-4766-a34c-423c2e7f471b-service-ca-bundle\") pod \"router-default-5444994796-n8wrd\" (UID: \"30bec2ef-4635-4766-a34c-423c2e7f471b\") " pod="openshift-ingress/router-default-5444994796-n8wrd" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.278358 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/eceb74a4-c695-46e9-ba5f-b5145d20d968-socket-dir\") pod \"csi-hostpathplugin-kbz2w\" (UID: \"eceb74a4-c695-46e9-ba5f-b5145d20d968\") " pod="hostpath-provisioner/csi-hostpathplugin-kbz2w" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.278381 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/adb430a2-20a7-4e9d-8f92-446b56e36e4f-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-q6ck8\" (UID: \"adb430a2-20a7-4e9d-8f92-446b56e36e4f\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q6ck8" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.278402 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/38005ee2-6635-498c-b156-ac57b4f2f4d9-registry-certificates\") pod \"image-registry-697d97f7c8-zf572\" (UID: \"38005ee2-6635-498c-b156-ac57b4f2f4d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-zf572" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.278424 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/220b6958-d8ee-4733-a93a-35d2ced9f150-config-volume\") pod \"dns-default-sbcxd\" (UID: \"220b6958-d8ee-4733-a93a-35d2ced9f150\") " pod="openshift-dns/dns-default-sbcxd" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.278446 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/38005ee2-6635-498c-b156-ac57b4f2f4d9-ca-trust-extracted\") pod \"image-registry-697d97f7c8-zf572\" (UID: \"38005ee2-6635-498c-b156-ac57b4f2f4d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-zf572" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.278465 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mk9kz\" (UniqueName: \"kubernetes.io/projected/cefdf543-6d3a-4e62-9875-ca26b060f437-kube-api-access-mk9kz\") pod \"machine-config-operator-74547568cd-6q4dc\" (UID: \"cefdf543-6d3a-4e62-9875-ca26b060f437\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6q4dc" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.278487 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/e963dd44-250d-4467-980c-1b67ef1bdd93-machine-approver-tls\") pod \"machine-approver-56656f9798-wrqbs\" (UID: \"e963dd44-250d-4467-980c-1b67ef1bdd93\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wrqbs" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.278509 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/eceb74a4-c695-46e9-ba5f-b5145d20d968-mountpoint-dir\") pod \"csi-hostpathplugin-kbz2w\" (UID: \"eceb74a4-c695-46e9-ba5f-b5145d20d968\") " pod="hostpath-provisioner/csi-hostpathplugin-kbz2w" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.278529 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/cefdf543-6d3a-4e62-9875-ca26b060f437-images\") pod \"machine-config-operator-74547568cd-6q4dc\" (UID: \"cefdf543-6d3a-4e62-9875-ca26b060f437\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6q4dc" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.278549 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c78b27aa-097e-4308-ad5f-1d609b6b1b6c-webhook-cert\") pod \"packageserver-d55dfcdfc-fsfs7\" (UID: \"c78b27aa-097e-4308-ad5f-1d609b6b1b6c\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fsfs7" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.278570 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/b08141e2-41b8-45f4-bb40-c319621719c8-metrics-tls\") pod \"ingress-operator-5b745b69d9-h7zxk\" (UID: \"b08141e2-41b8-45f4-bb40-c319621719c8\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-h7zxk" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.278598 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e963dd44-250d-4467-980c-1b67ef1bdd93-auth-proxy-config\") pod \"machine-approver-56656f9798-wrqbs\" (UID: \"e963dd44-250d-4467-980c-1b67ef1bdd93\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wrqbs" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.278620 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fp9sr\" (UniqueName: \"kubernetes.io/projected/adb430a2-20a7-4e9d-8f92-446b56e36e4f-kube-api-access-fp9sr\") pod \"cluster-image-registry-operator-dc59b4c8b-q6ck8\" (UID: \"adb430a2-20a7-4e9d-8f92-446b56e36e4f\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q6ck8" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.278657 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/cefdf543-6d3a-4e62-9875-ca26b060f437-proxy-tls\") pod \"machine-config-operator-74547568cd-6q4dc\" (UID: \"cefdf543-6d3a-4e62-9875-ca26b060f437\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6q4dc" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.278682 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2vgdm\" (UniqueName: \"kubernetes.io/projected/cf1bd6cb-c047-423c-9d36-7246a6108844-kube-api-access-2vgdm\") pod \"downloads-7954f5f757-mj2tr\" (UID: \"cf1bd6cb-c047-423c-9d36-7246a6108844\") " pod="openshift-console/downloads-7954f5f757-mj2tr" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.278703 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/eceb74a4-c695-46e9-ba5f-b5145d20d968-registration-dir\") pod \"csi-hostpathplugin-kbz2w\" (UID: \"eceb74a4-c695-46e9-ba5f-b5145d20d968\") " pod="hostpath-provisioner/csi-hostpathplugin-kbz2w" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.278720 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gspwb\" (UniqueName: \"kubernetes.io/projected/5b8b374c-b131-402c-8e3b-02d93d416489-kube-api-access-gspwb\") pod \"migrator-59844c95c7-l4m2j\" (UID: \"5b8b374c-b131-402c-8e3b-02d93d416489\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-l4m2j" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.278751 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c3ce3c34-9b72-4964-b59a-f6b3c0a0e5c8-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-x28lp\" (UID: \"c3ce3c34-9b72-4964-b59a-f6b3c0a0e5c8\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-x28lp" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.278774 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8dd6t\" (UniqueName: \"kubernetes.io/projected/30bec2ef-4635-4766-a34c-423c2e7f471b-kube-api-access-8dd6t\") pod \"router-default-5444994796-n8wrd\" (UID: \"30bec2ef-4635-4766-a34c-423c2e7f471b\") " pod="openshift-ingress/router-default-5444994796-n8wrd" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.278791 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/73ab8adf-cf47-4482-bbdb-8fb6c3edc87a-config\") pod \"route-controller-manager-6576b87f9c-87b7k\" (UID: \"73ab8adf-cf47-4482-bbdb-8fb6c3edc87a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-87b7k" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.278808 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hckxn\" (UniqueName: \"kubernetes.io/projected/3e5d90ab-884d-420b-a284-74f4b578a864-kube-api-access-hckxn\") pod \"package-server-manager-789f6589d5-5zqnk\" (UID: \"3e5d90ab-884d-420b-a284-74f4b578a864\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5zqnk" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.278830 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/eceb74a4-c695-46e9-ba5f-b5145d20d968-csi-data-dir\") pod \"csi-hostpathplugin-kbz2w\" (UID: \"eceb74a4-c695-46e9-ba5f-b5145d20d968\") " pod="hostpath-provisioner/csi-hostpathplugin-kbz2w" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.278855 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/03384260-7d59-42d5-ab3c-b411e9ed88d1-profile-collector-cert\") pod \"olm-operator-6b444d44fb-svdpj\" (UID: \"03384260-7d59-42d5-ab3c-b411e9ed88d1\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-svdpj" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.278877 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0f6c4496-7751-49db-adc2-18099589c708-secret-volume\") pod \"collect-profiles-29424000-nf7zl\" (UID: \"0f6c4496-7751-49db-adc2-18099589c708\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424000-nf7zl" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.278899 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/38005ee2-6635-498c-b156-ac57b4f2f4d9-trusted-ca\") pod \"image-registry-697d97f7c8-zf572\" (UID: \"38005ee2-6635-498c-b156-ac57b4f2f4d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-zf572" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.278921 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/64e3a9e4-d299-4349-b493-994630f747a5-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-2z7ml\" (UID: \"64e3a9e4-d299-4349-b493-994630f747a5\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-2z7ml" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.278947 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0f6c4496-7751-49db-adc2-18099589c708-config-volume\") pod \"collect-profiles-29424000-nf7zl\" (UID: \"0f6c4496-7751-49db-adc2-18099589c708\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424000-nf7zl" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.278970 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/c78b27aa-097e-4308-ad5f-1d609b6b1b6c-tmpfs\") pod \"packageserver-d55dfcdfc-fsfs7\" (UID: \"c78b27aa-097e-4308-ad5f-1d609b6b1b6c\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fsfs7" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.278988 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/5c0bfb74-4985-47a1-ade8-17afad50fabe-proxy-tls\") pod \"machine-config-controller-84d6567774-jgk9d\" (UID: \"5c0bfb74-4985-47a1-ade8-17afad50fabe\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jgk9d" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.279010 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f7673917-9864-467b-b1f8-af4b2bac9a48-serving-cert\") pod \"service-ca-operator-777779d784-6f4m5\" (UID: \"f7673917-9864-467b-b1f8-af4b2bac9a48\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-6f4m5" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.279031 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rcfdr\" (UniqueName: \"kubernetes.io/projected/1fff60d5-55fd-41d8-9225-251277e8e3c2-kube-api-access-rcfdr\") pod \"dns-operator-744455d44c-g4bql\" (UID: \"1fff60d5-55fd-41d8-9225-251277e8e3c2\") " pod="openshift-dns-operator/dns-operator-744455d44c-g4bql" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.279048 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ad724960-18d1-4e83-944c-4cdeb033a436-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-ttrm2\" (UID: \"ad724960-18d1-4e83-944c-4cdeb033a436\") " pod="openshift-marketplace/marketplace-operator-79b997595-ttrm2" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.279068 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/30bec2ef-4635-4766-a34c-423c2e7f471b-default-certificate\") pod \"router-default-5444994796-n8wrd\" (UID: \"30bec2ef-4635-4766-a34c-423c2e7f471b\") " pod="openshift-ingress/router-default-5444994796-n8wrd" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.280181 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/38005ee2-6635-498c-b156-ac57b4f2f4d9-ca-trust-extracted\") pod \"image-registry-697d97f7c8-zf572\" (UID: \"38005ee2-6635-498c-b156-ac57b4f2f4d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-zf572" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.280537 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e963dd44-250d-4467-980c-1b67ef1bdd93-auth-proxy-config\") pod \"machine-approver-56656f9798-wrqbs\" (UID: \"e963dd44-250d-4467-980c-1b67ef1bdd93\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wrqbs" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.280603 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/c4c320f7-3b40-4edd-8e25-40b477e1d374-profile-collector-cert\") pod \"catalog-operator-68c6474976-s6qsz\" (UID: \"c4c320f7-3b40-4edd-8e25-40b477e1d374\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-s6qsz" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.280736 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/38005ee2-6635-498c-b156-ac57b4f2f4d9-registry-certificates\") pod \"image-registry-697d97f7c8-zf572\" (UID: \"38005ee2-6635-498c-b156-ac57b4f2f4d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-zf572" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.280790 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c78b27aa-097e-4308-ad5f-1d609b6b1b6c-apiservice-cert\") pod \"packageserver-d55dfcdfc-fsfs7\" (UID: \"c78b27aa-097e-4308-ad5f-1d609b6b1b6c\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fsfs7" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.280815 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/c4c320f7-3b40-4edd-8e25-40b477e1d374-srv-cert\") pod \"catalog-operator-68c6474976-s6qsz\" (UID: \"c4c320f7-3b40-4edd-8e25-40b477e1d374\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-s6qsz" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.281116 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-47grh\" (UniqueName: \"kubernetes.io/projected/03384260-7d59-42d5-ab3c-b411e9ed88d1-kube-api-access-47grh\") pod \"olm-operator-6b444d44fb-svdpj\" (UID: \"03384260-7d59-42d5-ab3c-b411e9ed88d1\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-svdpj" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.281202 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/adb430a2-20a7-4e9d-8f92-446b56e36e4f-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-q6ck8\" (UID: \"adb430a2-20a7-4e9d-8f92-446b56e36e4f\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q6ck8" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.281312 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9zrk8\" (UniqueName: \"kubernetes.io/projected/f7673917-9864-467b-b1f8-af4b2bac9a48-kube-api-access-9zrk8\") pod \"service-ca-operator-777779d784-6f4m5\" (UID: \"f7673917-9864-467b-b1f8-af4b2bac9a48\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-6f4m5" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.281613 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/adb430a2-20a7-4e9d-8f92-446b56e36e4f-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-q6ck8\" (UID: \"adb430a2-20a7-4e9d-8f92-446b56e36e4f\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q6ck8" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.281691 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/03384260-7d59-42d5-ab3c-b411e9ed88d1-srv-cert\") pod \"olm-operator-6b444d44fb-svdpj\" (UID: \"03384260-7d59-42d5-ab3c-b411e9ed88d1\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-svdpj" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.281891 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/93550904-9295-4c4a-98bb-7a228c2699bb-node-bootstrap-token\") pod \"machine-config-server-9wb7z\" (UID: \"93550904-9295-4c4a-98bb-7a228c2699bb\") " pod="openshift-machine-config-operator/machine-config-server-9wb7z" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.282071 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b08141e2-41b8-45f4-bb40-c319621719c8-bound-sa-token\") pod \"ingress-operator-5b745b69d9-h7zxk\" (UID: \"b08141e2-41b8-45f4-bb40-c319621719c8\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-h7zxk" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.282244 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/38005ee2-6635-498c-b156-ac57b4f2f4d9-registry-tls\") pod \"image-registry-697d97f7c8-zf572\" (UID: \"38005ee2-6635-498c-b156-ac57b4f2f4d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-zf572" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.282294 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e963dd44-250d-4467-980c-1b67ef1bdd93-config\") pod \"machine-approver-56656f9798-wrqbs\" (UID: \"e963dd44-250d-4467-980c-1b67ef1bdd93\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wrqbs" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.282798 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e963dd44-250d-4467-980c-1b67ef1bdd93-config\") pod \"machine-approver-56656f9798-wrqbs\" (UID: \"e963dd44-250d-4467-980c-1b67ef1bdd93\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wrqbs" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.283000 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4bzjr\" (UniqueName: \"kubernetes.io/projected/c78b27aa-097e-4308-ad5f-1d609b6b1b6c-kube-api-access-4bzjr\") pod \"packageserver-d55dfcdfc-fsfs7\" (UID: \"c78b27aa-097e-4308-ad5f-1d609b6b1b6c\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fsfs7" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.283056 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h8x5s\" (UniqueName: \"kubernetes.io/projected/64e3a9e4-d299-4349-b493-994630f747a5-kube-api-access-h8x5s\") pod \"multus-admission-controller-857f4d67dd-2z7ml\" (UID: \"64e3a9e4-d299-4349-b493-994630f747a5\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-2z7ml" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.283130 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/e963dd44-250d-4467-980c-1b67ef1bdd93-machine-approver-tls\") pod \"machine-approver-56656f9798-wrqbs\" (UID: \"e963dd44-250d-4467-980c-1b67ef1bdd93\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wrqbs" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.283161 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-24srs\" (UniqueName: \"kubernetes.io/projected/38005ee2-6635-498c-b156-ac57b4f2f4d9-kube-api-access-24srs\") pod \"image-registry-697d97f7c8-zf572\" (UID: \"38005ee2-6635-498c-b156-ac57b4f2f4d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-zf572" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.283236 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5c0bfb74-4985-47a1-ade8-17afad50fabe-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-jgk9d\" (UID: \"5c0bfb74-4985-47a1-ade8-17afad50fabe\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jgk9d" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.283368 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-krqtd\" (UniqueName: \"kubernetes.io/projected/ad724960-18d1-4e83-944c-4cdeb033a436-kube-api-access-krqtd\") pod \"marketplace-operator-79b997595-ttrm2\" (UID: \"ad724960-18d1-4e83-944c-4cdeb033a436\") " pod="openshift-marketplace/marketplace-operator-79b997595-ttrm2" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.283456 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f7673917-9864-467b-b1f8-af4b2bac9a48-config\") pod \"service-ca-operator-777779d784-6f4m5\" (UID: \"f7673917-9864-467b-b1f8-af4b2bac9a48\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-6f4m5" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.283530 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-58dpf\" (UniqueName: \"kubernetes.io/projected/c4c320f7-3b40-4edd-8e25-40b477e1d374-kube-api-access-58dpf\") pod \"catalog-operator-68c6474976-s6qsz\" (UID: \"c4c320f7-3b40-4edd-8e25-40b477e1d374\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-s6qsz" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.283563 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xgz7t\" (UniqueName: \"kubernetes.io/projected/eceb74a4-c695-46e9-ba5f-b5145d20d968-kube-api-access-xgz7t\") pod \"csi-hostpathplugin-kbz2w\" (UID: \"eceb74a4-c695-46e9-ba5f-b5145d20d968\") " pod="hostpath-provisioner/csi-hostpathplugin-kbz2w" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.283942 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/38005ee2-6635-498c-b156-ac57b4f2f4d9-trusted-ca\") pod \"image-registry-697d97f7c8-zf572\" (UID: \"38005ee2-6635-498c-b156-ac57b4f2f4d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-zf572" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.284244 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/38005ee2-6635-498c-b156-ac57b4f2f4d9-installation-pull-secrets\") pod \"image-registry-697d97f7c8-zf572\" (UID: \"38005ee2-6635-498c-b156-ac57b4f2f4d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-zf572" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.284338 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/adb430a2-20a7-4e9d-8f92-446b56e36e4f-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-q6ck8\" (UID: \"adb430a2-20a7-4e9d-8f92-446b56e36e4f\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q6ck8" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.286576 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.288389 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/38005ee2-6635-498c-b156-ac57b4f2f4d9-registry-tls\") pod \"image-registry-697d97f7c8-zf572\" (UID: \"38005ee2-6635-498c-b156-ac57b4f2f4d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-zf572" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.307398 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.327050 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.346198 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.365905 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.385300 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/73ab8adf-cf47-4482-bbdb-8fb6c3edc87a-serving-cert\") pod \"route-controller-manager-6576b87f9c-87b7k\" (UID: \"73ab8adf-cf47-4482-bbdb-8fb6c3edc87a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-87b7k" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.385340 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/73ab8adf-cf47-4482-bbdb-8fb6c3edc87a-client-ca\") pod \"route-controller-manager-6576b87f9c-87b7k\" (UID: \"73ab8adf-cf47-4482-bbdb-8fb6c3edc87a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-87b7k" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.385383 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/30bec2ef-4635-4766-a34c-423c2e7f471b-service-ca-bundle\") pod \"router-default-5444994796-n8wrd\" (UID: \"30bec2ef-4635-4766-a34c-423c2e7f471b\") " pod="openshift-ingress/router-default-5444994796-n8wrd" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.385408 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/eceb74a4-c695-46e9-ba5f-b5145d20d968-socket-dir\") pod \"csi-hostpathplugin-kbz2w\" (UID: \"eceb74a4-c695-46e9-ba5f-b5145d20d968\") " pod="hostpath-provisioner/csi-hostpathplugin-kbz2w" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.385428 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mk9kz\" (UniqueName: \"kubernetes.io/projected/cefdf543-6d3a-4e62-9875-ca26b060f437-kube-api-access-mk9kz\") pod \"machine-config-operator-74547568cd-6q4dc\" (UID: \"cefdf543-6d3a-4e62-9875-ca26b060f437\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6q4dc" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.385465 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/220b6958-d8ee-4733-a93a-35d2ced9f150-config-volume\") pod \"dns-default-sbcxd\" (UID: \"220b6958-d8ee-4733-a93a-35d2ced9f150\") " pod="openshift-dns/dns-default-sbcxd" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.385481 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c78b27aa-097e-4308-ad5f-1d609b6b1b6c-webhook-cert\") pod \"packageserver-d55dfcdfc-fsfs7\" (UID: \"c78b27aa-097e-4308-ad5f-1d609b6b1b6c\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fsfs7" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.385497 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/eceb74a4-c695-46e9-ba5f-b5145d20d968-mountpoint-dir\") pod \"csi-hostpathplugin-kbz2w\" (UID: \"eceb74a4-c695-46e9-ba5f-b5145d20d968\") " pod="hostpath-provisioner/csi-hostpathplugin-kbz2w" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.385530 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/cefdf543-6d3a-4e62-9875-ca26b060f437-images\") pod \"machine-config-operator-74547568cd-6q4dc\" (UID: \"cefdf543-6d3a-4e62-9875-ca26b060f437\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6q4dc" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.385556 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/b08141e2-41b8-45f4-bb40-c319621719c8-metrics-tls\") pod \"ingress-operator-5b745b69d9-h7zxk\" (UID: \"b08141e2-41b8-45f4-bb40-c319621719c8\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-h7zxk" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.385714 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/eceb74a4-c695-46e9-ba5f-b5145d20d968-mountpoint-dir\") pod \"csi-hostpathplugin-kbz2w\" (UID: \"eceb74a4-c695-46e9-ba5f-b5145d20d968\") " pod="hostpath-provisioner/csi-hostpathplugin-kbz2w" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.385972 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/eceb74a4-c695-46e9-ba5f-b5145d20d968-socket-dir\") pod \"csi-hostpathplugin-kbz2w\" (UID: \"eceb74a4-c695-46e9-ba5f-b5145d20d968\") " pod="hostpath-provisioner/csi-hostpathplugin-kbz2w" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.386534 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/cefdf543-6d3a-4e62-9875-ca26b060f437-proxy-tls\") pod \"machine-config-operator-74547568cd-6q4dc\" (UID: \"cefdf543-6d3a-4e62-9875-ca26b060f437\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6q4dc" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.386662 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2vgdm\" (UniqueName: \"kubernetes.io/projected/cf1bd6cb-c047-423c-9d36-7246a6108844-kube-api-access-2vgdm\") pod \"downloads-7954f5f757-mj2tr\" (UID: \"cf1bd6cb-c047-423c-9d36-7246a6108844\") " pod="openshift-console/downloads-7954f5f757-mj2tr" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.386690 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/eceb74a4-c695-46e9-ba5f-b5145d20d968-registration-dir\") pod \"csi-hostpathplugin-kbz2w\" (UID: \"eceb74a4-c695-46e9-ba5f-b5145d20d968\") " pod="hostpath-provisioner/csi-hostpathplugin-kbz2w" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.386708 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gspwb\" (UniqueName: \"kubernetes.io/projected/5b8b374c-b131-402c-8e3b-02d93d416489-kube-api-access-gspwb\") pod \"migrator-59844c95c7-l4m2j\" (UID: \"5b8b374c-b131-402c-8e3b-02d93d416489\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-l4m2j" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.386823 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8dd6t\" (UniqueName: \"kubernetes.io/projected/30bec2ef-4635-4766-a34c-423c2e7f471b-kube-api-access-8dd6t\") pod \"router-default-5444994796-n8wrd\" (UID: \"30bec2ef-4635-4766-a34c-423c2e7f471b\") " pod="openshift-ingress/router-default-5444994796-n8wrd" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.386846 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c3ce3c34-9b72-4964-b59a-f6b3c0a0e5c8-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-x28lp\" (UID: \"c3ce3c34-9b72-4964-b59a-f6b3c0a0e5c8\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-x28lp" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.386867 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hckxn\" (UniqueName: \"kubernetes.io/projected/3e5d90ab-884d-420b-a284-74f4b578a864-kube-api-access-hckxn\") pod \"package-server-manager-789f6589d5-5zqnk\" (UID: \"3e5d90ab-884d-420b-a284-74f4b578a864\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5zqnk" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.386932 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/cefdf543-6d3a-4e62-9875-ca26b060f437-images\") pod \"machine-config-operator-74547568cd-6q4dc\" (UID: \"cefdf543-6d3a-4e62-9875-ca26b060f437\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6q4dc" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.386981 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/73ab8adf-cf47-4482-bbdb-8fb6c3edc87a-config\") pod \"route-controller-manager-6576b87f9c-87b7k\" (UID: \"73ab8adf-cf47-4482-bbdb-8fb6c3edc87a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-87b7k" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.387003 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/eceb74a4-c695-46e9-ba5f-b5145d20d968-csi-data-dir\") pod \"csi-hostpathplugin-kbz2w\" (UID: \"eceb74a4-c695-46e9-ba5f-b5145d20d968\") " pod="hostpath-provisioner/csi-hostpathplugin-kbz2w" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.387026 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/03384260-7d59-42d5-ab3c-b411e9ed88d1-profile-collector-cert\") pod \"olm-operator-6b444d44fb-svdpj\" (UID: \"03384260-7d59-42d5-ab3c-b411e9ed88d1\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-svdpj" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.387165 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0f6c4496-7751-49db-adc2-18099589c708-secret-volume\") pod \"collect-profiles-29424000-nf7zl\" (UID: \"0f6c4496-7751-49db-adc2-18099589c708\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424000-nf7zl" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.387251 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/eceb74a4-c695-46e9-ba5f-b5145d20d968-registration-dir\") pod \"csi-hostpathplugin-kbz2w\" (UID: \"eceb74a4-c695-46e9-ba5f-b5145d20d968\") " pod="hostpath-provisioner/csi-hostpathplugin-kbz2w" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.387319 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/64e3a9e4-d299-4349-b493-994630f747a5-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-2z7ml\" (UID: \"64e3a9e4-d299-4349-b493-994630f747a5\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-2z7ml" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.387345 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/c78b27aa-097e-4308-ad5f-1d609b6b1b6c-tmpfs\") pod \"packageserver-d55dfcdfc-fsfs7\" (UID: \"c78b27aa-097e-4308-ad5f-1d609b6b1b6c\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fsfs7" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.387362 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0f6c4496-7751-49db-adc2-18099589c708-config-volume\") pod \"collect-profiles-29424000-nf7zl\" (UID: \"0f6c4496-7751-49db-adc2-18099589c708\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424000-nf7zl" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.387485 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f7673917-9864-467b-b1f8-af4b2bac9a48-serving-cert\") pod \"service-ca-operator-777779d784-6f4m5\" (UID: \"f7673917-9864-467b-b1f8-af4b2bac9a48\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-6f4m5" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.387512 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/5c0bfb74-4985-47a1-ade8-17afad50fabe-proxy-tls\") pod \"machine-config-controller-84d6567774-jgk9d\" (UID: \"5c0bfb74-4985-47a1-ade8-17afad50fabe\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jgk9d" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.387529 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ad724960-18d1-4e83-944c-4cdeb033a436-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-ttrm2\" (UID: \"ad724960-18d1-4e83-944c-4cdeb033a436\") " pod="openshift-marketplace/marketplace-operator-79b997595-ttrm2" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.387659 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/30bec2ef-4635-4766-a34c-423c2e7f471b-default-certificate\") pod \"router-default-5444994796-n8wrd\" (UID: \"30bec2ef-4635-4766-a34c-423c2e7f471b\") " pod="openshift-ingress/router-default-5444994796-n8wrd" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.387678 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rcfdr\" (UniqueName: \"kubernetes.io/projected/1fff60d5-55fd-41d8-9225-251277e8e3c2-kube-api-access-rcfdr\") pod \"dns-operator-744455d44c-g4bql\" (UID: \"1fff60d5-55fd-41d8-9225-251277e8e3c2\") " pod="openshift-dns-operator/dns-operator-744455d44c-g4bql" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.387729 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/eceb74a4-c695-46e9-ba5f-b5145d20d968-csi-data-dir\") pod \"csi-hostpathplugin-kbz2w\" (UID: \"eceb74a4-c695-46e9-ba5f-b5145d20d968\") " pod="hostpath-provisioner/csi-hostpathplugin-kbz2w" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.387796 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c78b27aa-097e-4308-ad5f-1d609b6b1b6c-apiservice-cert\") pod \"packageserver-d55dfcdfc-fsfs7\" (UID: \"c78b27aa-097e-4308-ad5f-1d609b6b1b6c\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fsfs7" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.387821 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/c4c320f7-3b40-4edd-8e25-40b477e1d374-profile-collector-cert\") pod \"catalog-operator-68c6474976-s6qsz\" (UID: \"c4c320f7-3b40-4edd-8e25-40b477e1d374\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-s6qsz" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.387841 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/c4c320f7-3b40-4edd-8e25-40b477e1d374-srv-cert\") pod \"catalog-operator-68c6474976-s6qsz\" (UID: \"c4c320f7-3b40-4edd-8e25-40b477e1d374\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-s6qsz" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.387874 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.387894 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/73ab8adf-cf47-4482-bbdb-8fb6c3edc87a-client-ca\") pod \"route-controller-manager-6576b87f9c-87b7k\" (UID: \"73ab8adf-cf47-4482-bbdb-8fb6c3edc87a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-87b7k" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.387973 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9zrk8\" (UniqueName: \"kubernetes.io/projected/f7673917-9864-467b-b1f8-af4b2bac9a48-kube-api-access-9zrk8\") pod \"service-ca-operator-777779d784-6f4m5\" (UID: \"f7673917-9864-467b-b1f8-af4b2bac9a48\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-6f4m5" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.388001 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/03384260-7d59-42d5-ab3c-b411e9ed88d1-srv-cert\") pod \"olm-operator-6b444d44fb-svdpj\" (UID: \"03384260-7d59-42d5-ab3c-b411e9ed88d1\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-svdpj" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.388134 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-47grh\" (UniqueName: \"kubernetes.io/projected/03384260-7d59-42d5-ab3c-b411e9ed88d1-kube-api-access-47grh\") pod \"olm-operator-6b444d44fb-svdpj\" (UID: \"03384260-7d59-42d5-ab3c-b411e9ed88d1\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-svdpj" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.388166 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/93550904-9295-4c4a-98bb-7a228c2699bb-node-bootstrap-token\") pod \"machine-config-server-9wb7z\" (UID: \"93550904-9295-4c4a-98bb-7a228c2699bb\") " pod="openshift-machine-config-operator/machine-config-server-9wb7z" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.388309 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b08141e2-41b8-45f4-bb40-c319621719c8-bound-sa-token\") pod \"ingress-operator-5b745b69d9-h7zxk\" (UID: \"b08141e2-41b8-45f4-bb40-c319621719c8\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-h7zxk" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.388455 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4bzjr\" (UniqueName: \"kubernetes.io/projected/c78b27aa-097e-4308-ad5f-1d609b6b1b6c-kube-api-access-4bzjr\") pod \"packageserver-d55dfcdfc-fsfs7\" (UID: \"c78b27aa-097e-4308-ad5f-1d609b6b1b6c\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fsfs7" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.388495 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h8x5s\" (UniqueName: \"kubernetes.io/projected/64e3a9e4-d299-4349-b493-994630f747a5-kube-api-access-h8x5s\") pod \"multus-admission-controller-857f4d67dd-2z7ml\" (UID: \"64e3a9e4-d299-4349-b493-994630f747a5\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-2z7ml" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.388565 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5c0bfb74-4985-47a1-ade8-17afad50fabe-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-jgk9d\" (UID: \"5c0bfb74-4985-47a1-ade8-17afad50fabe\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jgk9d" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.388601 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/c78b27aa-097e-4308-ad5f-1d609b6b1b6c-tmpfs\") pod \"packageserver-d55dfcdfc-fsfs7\" (UID: \"c78b27aa-097e-4308-ad5f-1d609b6b1b6c\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fsfs7" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.388686 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-krqtd\" (UniqueName: \"kubernetes.io/projected/ad724960-18d1-4e83-944c-4cdeb033a436-kube-api-access-krqtd\") pod \"marketplace-operator-79b997595-ttrm2\" (UID: \"ad724960-18d1-4e83-944c-4cdeb033a436\") " pod="openshift-marketplace/marketplace-operator-79b997595-ttrm2" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.388716 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/73ab8adf-cf47-4482-bbdb-8fb6c3edc87a-config\") pod \"route-controller-manager-6576b87f9c-87b7k\" (UID: \"73ab8adf-cf47-4482-bbdb-8fb6c3edc87a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-87b7k" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.388775 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f7673917-9864-467b-b1f8-af4b2bac9a48-config\") pod \"service-ca-operator-777779d784-6f4m5\" (UID: \"f7673917-9864-467b-b1f8-af4b2bac9a48\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-6f4m5" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.388832 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-58dpf\" (UniqueName: \"kubernetes.io/projected/c4c320f7-3b40-4edd-8e25-40b477e1d374-kube-api-access-58dpf\") pod \"catalog-operator-68c6474976-s6qsz\" (UID: \"c4c320f7-3b40-4edd-8e25-40b477e1d374\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-s6qsz" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.388879 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xgz7t\" (UniqueName: \"kubernetes.io/projected/eceb74a4-c695-46e9-ba5f-b5145d20d968-kube-api-access-xgz7t\") pod \"csi-hostpathplugin-kbz2w\" (UID: \"eceb74a4-c695-46e9-ba5f-b5145d20d968\") " pod="hostpath-provisioner/csi-hostpathplugin-kbz2w" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.388938 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/ad724960-18d1-4e83-944c-4cdeb033a436-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-ttrm2\" (UID: \"ad724960-18d1-4e83-944c-4cdeb033a436\") " pod="openshift-marketplace/marketplace-operator-79b997595-ttrm2" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.388984 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/cefdf543-6d3a-4e62-9875-ca26b060f437-auth-proxy-config\") pod \"machine-config-operator-74547568cd-6q4dc\" (UID: \"cefdf543-6d3a-4e62-9875-ca26b060f437\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6q4dc" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.389021 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vmgfp\" (UniqueName: \"kubernetes.io/projected/73ab8adf-cf47-4482-bbdb-8fb6c3edc87a-kube-api-access-vmgfp\") pod \"route-controller-manager-6576b87f9c-87b7k\" (UID: \"73ab8adf-cf47-4482-bbdb-8fb6c3edc87a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-87b7k" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.389075 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g6g9q\" (UniqueName: \"kubernetes.io/projected/b08141e2-41b8-45f4-bb40-c319621719c8-kube-api-access-g6g9q\") pod \"ingress-operator-5b745b69d9-h7zxk\" (UID: \"b08141e2-41b8-45f4-bb40-c319621719c8\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-h7zxk" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.389150 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pt56k\" (UniqueName: \"kubernetes.io/projected/220b6958-d8ee-4733-a93a-35d2ced9f150-kube-api-access-pt56k\") pod \"dns-default-sbcxd\" (UID: \"220b6958-d8ee-4733-a93a-35d2ced9f150\") " pod="openshift-dns/dns-default-sbcxd" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.389222 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/30bec2ef-4635-4766-a34c-423c2e7f471b-metrics-certs\") pod \"router-default-5444994796-n8wrd\" (UID: \"30bec2ef-4635-4766-a34c-423c2e7f471b\") " pod="openshift-ingress/router-default-5444994796-n8wrd" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.389269 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5c0bfb74-4985-47a1-ade8-17afad50fabe-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-jgk9d\" (UID: \"5c0bfb74-4985-47a1-ade8-17afad50fabe\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jgk9d" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.389276 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bmjqc\" (UniqueName: \"kubernetes.io/projected/5c0bfb74-4985-47a1-ade8-17afad50fabe-kube-api-access-bmjqc\") pod \"machine-config-controller-84d6567774-jgk9d\" (UID: \"5c0bfb74-4985-47a1-ade8-17afad50fabe\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jgk9d" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.389339 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-grc8b\" (UniqueName: \"kubernetes.io/projected/93550904-9295-4c4a-98bb-7a228c2699bb-kube-api-access-grc8b\") pod \"machine-config-server-9wb7z\" (UID: \"93550904-9295-4c4a-98bb-7a228c2699bb\") " pod="openshift-machine-config-operator/machine-config-server-9wb7z" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.389399 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3e5d90ab-884d-420b-a284-74f4b578a864-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-5zqnk\" (UID: \"3e5d90ab-884d-420b-a284-74f4b578a864\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5zqnk" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.389476 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zf572\" (UID: \"38005ee2-6635-498c-b156-ac57b4f2f4d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-zf572" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.389522 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m8497\" (UniqueName: \"kubernetes.io/projected/0f6c4496-7751-49db-adc2-18099589c708-kube-api-access-m8497\") pod \"collect-profiles-29424000-nf7zl\" (UID: \"0f6c4496-7751-49db-adc2-18099589c708\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424000-nf7zl" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.389558 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/220b6958-d8ee-4733-a93a-35d2ced9f150-metrics-tls\") pod \"dns-default-sbcxd\" (UID: \"220b6958-d8ee-4733-a93a-35d2ced9f150\") " pod="openshift-dns/dns-default-sbcxd" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.389607 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/30bec2ef-4635-4766-a34c-423c2e7f471b-stats-auth\") pod \"router-default-5444994796-n8wrd\" (UID: \"30bec2ef-4635-4766-a34c-423c2e7f471b\") " pod="openshift-ingress/router-default-5444994796-n8wrd" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.389697 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h6gns\" (UniqueName: \"kubernetes.io/projected/c3ce3c34-9b72-4964-b59a-f6b3c0a0e5c8-kube-api-access-h6gns\") pod \"kube-storage-version-migrator-operator-b67b599dd-x28lp\" (UID: \"c3ce3c34-9b72-4964-b59a-f6b3c0a0e5c8\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-x28lp" Dec 11 08:13:34 crc kubenswrapper[4860]: E1211 08:13:34.389734 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:13:34.889719926 +0000 UTC m=+147.618238981 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zf572" (UID: "38005ee2-6635-498c-b156-ac57b4f2f4d9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.389767 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/1fff60d5-55fd-41d8-9225-251277e8e3c2-metrics-tls\") pod \"dns-operator-744455d44c-g4bql\" (UID: \"1fff60d5-55fd-41d8-9225-251277e8e3c2\") " pod="openshift-dns-operator/dns-operator-744455d44c-g4bql" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.389795 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b08141e2-41b8-45f4-bb40-c319621719c8-trusted-ca\") pod \"ingress-operator-5b745b69d9-h7zxk\" (UID: \"b08141e2-41b8-45f4-bb40-c319621719c8\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-h7zxk" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.389819 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c3ce3c34-9b72-4964-b59a-f6b3c0a0e5c8-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-x28lp\" (UID: \"c3ce3c34-9b72-4964-b59a-f6b3c0a0e5c8\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-x28lp" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.389850 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/93550904-9295-4c4a-98bb-7a228c2699bb-certs\") pod \"machine-config-server-9wb7z\" (UID: \"93550904-9295-4c4a-98bb-7a228c2699bb\") " pod="openshift-machine-config-operator/machine-config-server-9wb7z" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.389865 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/cefdf543-6d3a-4e62-9875-ca26b060f437-auth-proxy-config\") pod \"machine-config-operator-74547568cd-6q4dc\" (UID: \"cefdf543-6d3a-4e62-9875-ca26b060f437\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6q4dc" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.389880 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/eceb74a4-c695-46e9-ba5f-b5145d20d968-plugins-dir\") pod \"csi-hostpathplugin-kbz2w\" (UID: \"eceb74a4-c695-46e9-ba5f-b5145d20d968\") " pod="hostpath-provisioner/csi-hostpathplugin-kbz2w" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.389916 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/eceb74a4-c695-46e9-ba5f-b5145d20d968-plugins-dir\") pod \"csi-hostpathplugin-kbz2w\" (UID: \"eceb74a4-c695-46e9-ba5f-b5145d20d968\") " pod="hostpath-provisioner/csi-hostpathplugin-kbz2w" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.390425 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c3ce3c34-9b72-4964-b59a-f6b3c0a0e5c8-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-x28lp\" (UID: \"c3ce3c34-9b72-4964-b59a-f6b3c0a0e5c8\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-x28lp" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.390863 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/cefdf543-6d3a-4e62-9875-ca26b060f437-proxy-tls\") pod \"machine-config-operator-74547568cd-6q4dc\" (UID: \"cefdf543-6d3a-4e62-9875-ca26b060f437\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6q4dc" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.391273 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c3ce3c34-9b72-4964-b59a-f6b3c0a0e5c8-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-x28lp\" (UID: \"c3ce3c34-9b72-4964-b59a-f6b3c0a0e5c8\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-x28lp" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.395407 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/73ab8adf-cf47-4482-bbdb-8fb6c3edc87a-serving-cert\") pod \"route-controller-manager-6576b87f9c-87b7k\" (UID: \"73ab8adf-cf47-4482-bbdb-8fb6c3edc87a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-87b7k" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.405799 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.427276 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.446222 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.455227 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/1fff60d5-55fd-41d8-9225-251277e8e3c2-metrics-tls\") pod \"dns-operator-744455d44c-g4bql\" (UID: \"1fff60d5-55fd-41d8-9225-251277e8e3c2\") " pod="openshift-dns-operator/dns-operator-744455d44c-g4bql" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.466555 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.487293 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.490719 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:13:34 crc kubenswrapper[4860]: E1211 08:13:34.491111 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:13:34.991072332 +0000 UTC m=+147.719591427 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.491346 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zf572\" (UID: \"38005ee2-6635-498c-b156-ac57b4f2f4d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-zf572" Dec 11 08:13:34 crc kubenswrapper[4860]: E1211 08:13:34.491856 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:13:34.991834658 +0000 UTC m=+147.720353743 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zf572" (UID: "38005ee2-6635-498c-b156-ac57b4f2f4d9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.518379 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.521964 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b08141e2-41b8-45f4-bb40-c319621719c8-trusted-ca\") pod \"ingress-operator-5b745b69d9-h7zxk\" (UID: \"b08141e2-41b8-45f4-bb40-c319621719c8\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-h7zxk" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.526097 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.547267 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.561130 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/b08141e2-41b8-45f4-bb40-c319621719c8-metrics-tls\") pod \"ingress-operator-5b745b69d9-h7zxk\" (UID: \"b08141e2-41b8-45f4-bb40-c319621719c8\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-h7zxk" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.566911 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.593606 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:13:34 crc kubenswrapper[4860]: E1211 08:13:34.593729 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:13:35.093712433 +0000 UTC m=+147.822231478 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.594011 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zf572\" (UID: \"38005ee2-6635-498c-b156-ac57b4f2f4d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-zf572" Dec 11 08:13:34 crc kubenswrapper[4860]: E1211 08:13:34.594318 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:13:35.094308902 +0000 UTC m=+147.822827957 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zf572" (UID: "38005ee2-6635-498c-b156-ac57b4f2f4d9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.594347 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.597188 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.606383 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.626234 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.646352 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.666594 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.686677 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.695891 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:13:34 crc kubenswrapper[4860]: E1211 08:13:34.696008 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:13:35.19597924 +0000 UTC m=+147.924498305 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.696476 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.696817 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zf572\" (UID: \"38005ee2-6635-498c-b156-ac57b4f2f4d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-zf572" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.696863 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.696970 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:13:34 crc kubenswrapper[4860]: E1211 08:13:34.697215 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:13:35.197194782 +0000 UTC m=+147.925713847 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zf572" (UID: "38005ee2-6635-498c-b156-ac57b4f2f4d9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.697752 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.701731 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.701880 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.705375 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.726961 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.745537 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.766800 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.787159 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.798824 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:13:34 crc kubenswrapper[4860]: E1211 08:13:34.799052 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:13:35.299026034 +0000 UTC m=+148.027545099 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.799541 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zf572\" (UID: \"38005ee2-6635-498c-b156-ac57b4f2f4d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-zf572" Dec 11 08:13:34 crc kubenswrapper[4860]: E1211 08:13:34.799893 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:13:35.299877943 +0000 UTC m=+148.028397008 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zf572" (UID: "38005ee2-6635-498c-b156-ac57b4f2f4d9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.807051 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.826931 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.845907 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.867093 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.873779 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/30bec2ef-4635-4766-a34c-423c2e7f471b-stats-auth\") pod \"router-default-5444994796-n8wrd\" (UID: \"30bec2ef-4635-4766-a34c-423c2e7f471b\") " pod="openshift-ingress/router-default-5444994796-n8wrd" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.887246 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.892424 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/30bec2ef-4635-4766-a34c-423c2e7f471b-metrics-certs\") pod \"router-default-5444994796-n8wrd\" (UID: \"30bec2ef-4635-4766-a34c-423c2e7f471b\") " pod="openshift-ingress/router-default-5444994796-n8wrd" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.897288 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.900689 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:13:34 crc kubenswrapper[4860]: E1211 08:13:34.900815 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:13:35.400793004 +0000 UTC m=+148.129312069 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.901614 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zf572\" (UID: \"38005ee2-6635-498c-b156-ac57b4f2f4d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-zf572" Dec 11 08:13:34 crc kubenswrapper[4860]: E1211 08:13:34.901967 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:13:35.401953824 +0000 UTC m=+148.130472899 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zf572" (UID: "38005ee2-6635-498c-b156-ac57b4f2f4d9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.906297 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.906679 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.912058 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/30bec2ef-4635-4766-a34c-423c2e7f471b-default-certificate\") pod \"router-default-5444994796-n8wrd\" (UID: \"30bec2ef-4635-4766-a34c-423c2e7f471b\") " pod="openshift-ingress/router-default-5444994796-n8wrd" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.914718 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.927046 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.937835 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/30bec2ef-4635-4766-a34c-423c2e7f471b-service-ca-bundle\") pod \"router-default-5444994796-n8wrd\" (UID: \"30bec2ef-4635-4766-a34c-423c2e7f471b\") " pod="openshift-ingress/router-default-5444994796-n8wrd" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.946892 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.967498 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.986078 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 11 08:13:34 crc kubenswrapper[4860]: I1211 08:13:34.995909 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/ad724960-18d1-4e83-944c-4cdeb033a436-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-ttrm2\" (UID: \"ad724960-18d1-4e83-944c-4cdeb033a436\") " pod="openshift-marketplace/marketplace-operator-79b997595-ttrm2" Dec 11 08:13:35 crc kubenswrapper[4860]: I1211 08:13:35.004247 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:13:35 crc kubenswrapper[4860]: E1211 08:13:35.005454 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:13:35.505412843 +0000 UTC m=+148.233931948 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:13:35 crc kubenswrapper[4860]: I1211 08:13:35.030310 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 11 08:13:35 crc kubenswrapper[4860]: I1211 08:13:35.030671 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 11 08:13:35 crc kubenswrapper[4860]: I1211 08:13:35.043061 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ad724960-18d1-4e83-944c-4cdeb033a436-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-ttrm2\" (UID: \"ad724960-18d1-4e83-944c-4cdeb033a436\") " pod="openshift-marketplace/marketplace-operator-79b997595-ttrm2" Dec 11 08:13:35 crc kubenswrapper[4860]: I1211 08:13:35.047887 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 11 08:13:35 crc kubenswrapper[4860]: I1211 08:13:35.054479 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/5c0bfb74-4985-47a1-ade8-17afad50fabe-proxy-tls\") pod \"machine-config-controller-84d6567774-jgk9d\" (UID: \"5c0bfb74-4985-47a1-ade8-17afad50fabe\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jgk9d" Dec 11 08:13:35 crc kubenswrapper[4860]: I1211 08:13:35.067385 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 11 08:13:35 crc kubenswrapper[4860]: I1211 08:13:35.087720 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 11 08:13:35 crc kubenswrapper[4860]: I1211 08:13:35.106087 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 11 08:13:35 crc kubenswrapper[4860]: I1211 08:13:35.107302 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zf572\" (UID: \"38005ee2-6635-498c-b156-ac57b4f2f4d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-zf572" Dec 11 08:13:35 crc kubenswrapper[4860]: E1211 08:13:35.107756 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:13:35.607736492 +0000 UTC m=+148.336255607 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zf572" (UID: "38005ee2-6635-498c-b156-ac57b4f2f4d9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:13:35 crc kubenswrapper[4860]: I1211 08:13:35.127975 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 11 08:13:35 crc kubenswrapper[4860]: I1211 08:13:35.145211 4860 request.go:700] Waited for 1.005938071s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver-operator/secrets?fieldSelector=metadata.name%3Dkube-apiserver-operator-dockercfg-x57mr&limit=500&resourceVersion=0 Dec 11 08:13:35 crc kubenswrapper[4860]: I1211 08:13:35.148059 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 11 08:13:35 crc kubenswrapper[4860]: I1211 08:13:35.166499 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 11 08:13:35 crc kubenswrapper[4860]: I1211 08:13:35.186062 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 11 08:13:35 crc kubenswrapper[4860]: I1211 08:13:35.206475 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 11 08:13:35 crc kubenswrapper[4860]: I1211 08:13:35.208530 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:13:35 crc kubenswrapper[4860]: E1211 08:13:35.208696 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:13:35.708681404 +0000 UTC m=+148.437200459 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:13:35 crc kubenswrapper[4860]: I1211 08:13:35.209078 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zf572\" (UID: \"38005ee2-6635-498c-b156-ac57b4f2f4d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-zf572" Dec 11 08:13:35 crc kubenswrapper[4860]: E1211 08:13:35.209438 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:13:35.70942622 +0000 UTC m=+148.437945275 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zf572" (UID: "38005ee2-6635-498c-b156-ac57b4f2f4d9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:13:35 crc kubenswrapper[4860]: I1211 08:13:35.214698 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3e5d90ab-884d-420b-a284-74f4b578a864-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-5zqnk\" (UID: \"3e5d90ab-884d-420b-a284-74f4b578a864\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5zqnk" Dec 11 08:13:35 crc kubenswrapper[4860]: I1211 08:13:35.227496 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 11 08:13:35 crc kubenswrapper[4860]: I1211 08:13:35.246742 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 11 08:13:35 crc kubenswrapper[4860]: I1211 08:13:35.266229 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 11 08:13:35 crc kubenswrapper[4860]: I1211 08:13:35.286127 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 11 08:13:35 crc kubenswrapper[4860]: I1211 08:13:35.306791 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 11 08:13:35 crc kubenswrapper[4860]: I1211 08:13:35.310580 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:13:35 crc kubenswrapper[4860]: E1211 08:13:35.310774 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:13:35.810754485 +0000 UTC m=+148.539273540 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:13:35 crc kubenswrapper[4860]: I1211 08:13:35.310916 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zf572\" (UID: \"38005ee2-6635-498c-b156-ac57b4f2f4d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-zf572" Dec 11 08:13:35 crc kubenswrapper[4860]: E1211 08:13:35.311256 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:13:35.811248971 +0000 UTC m=+148.539768026 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zf572" (UID: "38005ee2-6635-498c-b156-ac57b4f2f4d9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:13:35 crc kubenswrapper[4860]: I1211 08:13:35.327543 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 11 08:13:35 crc kubenswrapper[4860]: I1211 08:13:35.380506 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 11 08:13:35 crc kubenswrapper[4860]: I1211 08:13:35.380767 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 11 08:13:35 crc kubenswrapper[4860]: E1211 08:13:35.385836 4860 configmap.go:193] Couldn't get configMap openshift-dns/dns-default: failed to sync configmap cache: timed out waiting for the condition Dec 11 08:13:35 crc kubenswrapper[4860]: E1211 08:13:35.385906 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/220b6958-d8ee-4733-a93a-35d2ced9f150-config-volume podName:220b6958-d8ee-4733-a93a-35d2ced9f150 nodeName:}" failed. No retries permitted until 2025-12-11 08:13:35.885886442 +0000 UTC m=+148.614405497 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-volume" (UniqueName: "kubernetes.io/configmap/220b6958-d8ee-4733-a93a-35d2ced9f150-config-volume") pod "dns-default-sbcxd" (UID: "220b6958-d8ee-4733-a93a-35d2ced9f150") : failed to sync configmap cache: timed out waiting for the condition Dec 11 08:13:35 crc kubenswrapper[4860]: I1211 08:13:35.386146 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 11 08:13:35 crc kubenswrapper[4860]: E1211 08:13:35.386194 4860 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/packageserver-service-cert: failed to sync secret cache: timed out waiting for the condition Dec 11 08:13:35 crc kubenswrapper[4860]: E1211 08:13:35.386226 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c78b27aa-097e-4308-ad5f-1d609b6b1b6c-webhook-cert podName:c78b27aa-097e-4308-ad5f-1d609b6b1b6c nodeName:}" failed. No retries permitted until 2025-12-11 08:13:35.886214723 +0000 UTC m=+148.614733778 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-cert" (UniqueName: "kubernetes.io/secret/c78b27aa-097e-4308-ad5f-1d609b6b1b6c-webhook-cert") pod "packageserver-d55dfcdfc-fsfs7" (UID: "c78b27aa-097e-4308-ad5f-1d609b6b1b6c") : failed to sync secret cache: timed out waiting for the condition Dec 11 08:13:35 crc kubenswrapper[4860]: I1211 08:13:35.387032 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"f72a933a571c0d4f1cfe37d533109ad4acb6fe7505448d749b6cf93cd7779473"} Dec 11 08:13:35 crc kubenswrapper[4860]: I1211 08:13:35.387119 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"b1fa827ec72cee475094627032831adb157ebc443697585fd357894188ba5a82"} Dec 11 08:13:35 crc kubenswrapper[4860]: I1211 08:13:35.387459 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:13:35 crc kubenswrapper[4860]: E1211 08:13:35.387733 4860 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/pprof-cert: failed to sync secret cache: timed out waiting for the condition Dec 11 08:13:35 crc kubenswrapper[4860]: E1211 08:13:35.387887 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/03384260-7d59-42d5-ab3c-b411e9ed88d1-profile-collector-cert podName:03384260-7d59-42d5-ab3c-b411e9ed88d1 nodeName:}" failed. No retries permitted until 2025-12-11 08:13:35.887860209 +0000 UTC m=+148.616379274 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "profile-collector-cert" (UniqueName: "kubernetes.io/secret/03384260-7d59-42d5-ab3c-b411e9ed88d1-profile-collector-cert") pod "olm-operator-6b444d44fb-svdpj" (UID: "03384260-7d59-42d5-ab3c-b411e9ed88d1") : failed to sync secret cache: timed out waiting for the condition Dec 11 08:13:35 crc kubenswrapper[4860]: E1211 08:13:35.387925 4860 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/pprof-cert: failed to sync secret cache: timed out waiting for the condition Dec 11 08:13:35 crc kubenswrapper[4860]: E1211 08:13:35.387937 4860 configmap.go:193] Couldn't get configMap openshift-operator-lifecycle-manager/collect-profiles-config: failed to sync configmap cache: timed out waiting for the condition Dec 11 08:13:35 crc kubenswrapper[4860]: E1211 08:13:35.387972 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0f6c4496-7751-49db-adc2-18099589c708-secret-volume podName:0f6c4496-7751-49db-adc2-18099589c708 nodeName:}" failed. No retries permitted until 2025-12-11 08:13:35.887960593 +0000 UTC m=+148.616479668 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "secret-volume" (UniqueName: "kubernetes.io/secret/0f6c4496-7751-49db-adc2-18099589c708-secret-volume") pod "collect-profiles-29424000-nf7zl" (UID: "0f6c4496-7751-49db-adc2-18099589c708") : failed to sync secret cache: timed out waiting for the condition Dec 11 08:13:35 crc kubenswrapper[4860]: E1211 08:13:35.388018 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/0f6c4496-7751-49db-adc2-18099589c708-config-volume podName:0f6c4496-7751-49db-adc2-18099589c708 nodeName:}" failed. No retries permitted until 2025-12-11 08:13:35.887992774 +0000 UTC m=+148.616511989 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-volume" (UniqueName: "kubernetes.io/configmap/0f6c4496-7751-49db-adc2-18099589c708-config-volume") pod "collect-profiles-29424000-nf7zl" (UID: "0f6c4496-7751-49db-adc2-18099589c708") : failed to sync configmap cache: timed out waiting for the condition Dec 11 08:13:35 crc kubenswrapper[4860]: E1211 08:13:35.388079 4860 secret.go:188] Couldn't get secret openshift-service-ca-operator/serving-cert: failed to sync secret cache: timed out waiting for the condition Dec 11 08:13:35 crc kubenswrapper[4860]: E1211 08:13:35.388153 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f7673917-9864-467b-b1f8-af4b2bac9a48-serving-cert podName:f7673917-9864-467b-b1f8-af4b2bac9a48 nodeName:}" failed. No retries permitted until 2025-12-11 08:13:35.888128948 +0000 UTC m=+148.616648013 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/f7673917-9864-467b-b1f8-af4b2bac9a48-serving-cert") pod "service-ca-operator-777779d784-6f4m5" (UID: "f7673917-9864-467b-b1f8-af4b2bac9a48") : failed to sync secret cache: timed out waiting for the condition Dec 11 08:13:35 crc kubenswrapper[4860]: E1211 08:13:35.388494 4860 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/packageserver-service-cert: failed to sync secret cache: timed out waiting for the condition Dec 11 08:13:35 crc kubenswrapper[4860]: E1211 08:13:35.388543 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c78b27aa-097e-4308-ad5f-1d609b6b1b6c-apiservice-cert podName:c78b27aa-097e-4308-ad5f-1d609b6b1b6c nodeName:}" failed. No retries permitted until 2025-12-11 08:13:35.888529902 +0000 UTC m=+148.617049127 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "apiservice-cert" (UniqueName: "kubernetes.io/secret/c78b27aa-097e-4308-ad5f-1d609b6b1b6c-apiservice-cert") pod "packageserver-d55dfcdfc-fsfs7" (UID: "c78b27aa-097e-4308-ad5f-1d609b6b1b6c") : failed to sync secret cache: timed out waiting for the condition Dec 11 08:13:35 crc kubenswrapper[4860]: E1211 08:13:35.388553 4860 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/pprof-cert: failed to sync secret cache: timed out waiting for the condition Dec 11 08:13:35 crc kubenswrapper[4860]: E1211 08:13:35.388570 4860 secret.go:188] Couldn't get secret openshift-multus/multus-admission-controller-secret: failed to sync secret cache: timed out waiting for the condition Dec 11 08:13:35 crc kubenswrapper[4860]: E1211 08:13:35.388577 4860 secret.go:188] Couldn't get secret openshift-machine-config-operator/node-bootstrapper-token: failed to sync secret cache: timed out waiting for the condition Dec 11 08:13:35 crc kubenswrapper[4860]: E1211 08:13:35.388602 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/64e3a9e4-d299-4349-b493-994630f747a5-webhook-certs podName:64e3a9e4-d299-4349-b493-994630f747a5 nodeName:}" failed. No retries permitted until 2025-12-11 08:13:35.888594545 +0000 UTC m=+148.617113800 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/64e3a9e4-d299-4349-b493-994630f747a5-webhook-certs") pod "multus-admission-controller-857f4d67dd-2z7ml" (UID: "64e3a9e4-d299-4349-b493-994630f747a5") : failed to sync secret cache: timed out waiting for the condition Dec 11 08:13:35 crc kubenswrapper[4860]: E1211 08:13:35.388630 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c4c320f7-3b40-4edd-8e25-40b477e1d374-profile-collector-cert podName:c4c320f7-3b40-4edd-8e25-40b477e1d374 nodeName:}" failed. No retries permitted until 2025-12-11 08:13:35.888612065 +0000 UTC m=+148.617131160 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "profile-collector-cert" (UniqueName: "kubernetes.io/secret/c4c320f7-3b40-4edd-8e25-40b477e1d374-profile-collector-cert") pod "catalog-operator-68c6474976-s6qsz" (UID: "c4c320f7-3b40-4edd-8e25-40b477e1d374") : failed to sync secret cache: timed out waiting for the condition Dec 11 08:13:35 crc kubenswrapper[4860]: E1211 08:13:35.388631 4860 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/catalog-operator-serving-cert: failed to sync secret cache: timed out waiting for the condition Dec 11 08:13:35 crc kubenswrapper[4860]: E1211 08:13:35.388658 4860 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/olm-operator-serving-cert: failed to sync secret cache: timed out waiting for the condition Dec 11 08:13:35 crc kubenswrapper[4860]: E1211 08:13:35.388677 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c4c320f7-3b40-4edd-8e25-40b477e1d374-srv-cert podName:c4c320f7-3b40-4edd-8e25-40b477e1d374 nodeName:}" failed. No retries permitted until 2025-12-11 08:13:35.888668237 +0000 UTC m=+148.617187492 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "srv-cert" (UniqueName: "kubernetes.io/secret/c4c320f7-3b40-4edd-8e25-40b477e1d374-srv-cert") pod "catalog-operator-68c6474976-s6qsz" (UID: "c4c320f7-3b40-4edd-8e25-40b477e1d374") : failed to sync secret cache: timed out waiting for the condition Dec 11 08:13:35 crc kubenswrapper[4860]: E1211 08:13:35.388698 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/93550904-9295-4c4a-98bb-7a228c2699bb-node-bootstrap-token podName:93550904-9295-4c4a-98bb-7a228c2699bb nodeName:}" failed. No retries permitted until 2025-12-11 08:13:35.888685358 +0000 UTC m=+148.617204453 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "node-bootstrap-token" (UniqueName: "kubernetes.io/secret/93550904-9295-4c4a-98bb-7a228c2699bb-node-bootstrap-token") pod "machine-config-server-9wb7z" (UID: "93550904-9295-4c4a-98bb-7a228c2699bb") : failed to sync secret cache: timed out waiting for the condition Dec 11 08:13:35 crc kubenswrapper[4860]: E1211 08:13:35.388727 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/03384260-7d59-42d5-ab3c-b411e9ed88d1-srv-cert podName:03384260-7d59-42d5-ab3c-b411e9ed88d1 nodeName:}" failed. No retries permitted until 2025-12-11 08:13:35.888715069 +0000 UTC m=+148.617234164 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "srv-cert" (UniqueName: "kubernetes.io/secret/03384260-7d59-42d5-ab3c-b411e9ed88d1-srv-cert") pod "olm-operator-6b444d44fb-svdpj" (UID: "03384260-7d59-42d5-ab3c-b411e9ed88d1") : failed to sync secret cache: timed out waiting for the condition Dec 11 08:13:35 crc kubenswrapper[4860]: E1211 08:13:35.389795 4860 secret.go:188] Couldn't get secret openshift-dns/dns-default-metrics-tls: failed to sync secret cache: timed out waiting for the condition Dec 11 08:13:35 crc kubenswrapper[4860]: E1211 08:13:35.389830 4860 configmap.go:193] Couldn't get configMap openshift-service-ca-operator/service-ca-operator-config: failed to sync configmap cache: timed out waiting for the condition Dec 11 08:13:35 crc kubenswrapper[4860]: E1211 08:13:35.389864 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/220b6958-d8ee-4733-a93a-35d2ced9f150-metrics-tls podName:220b6958-d8ee-4733-a93a-35d2ced9f150 nodeName:}" failed. No retries permitted until 2025-12-11 08:13:35.889846097 +0000 UTC m=+148.618365182 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/220b6958-d8ee-4733-a93a-35d2ced9f150-metrics-tls") pod "dns-default-sbcxd" (UID: "220b6958-d8ee-4733-a93a-35d2ced9f150") : failed to sync secret cache: timed out waiting for the condition Dec 11 08:13:35 crc kubenswrapper[4860]: E1211 08:13:35.389956 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/f7673917-9864-467b-b1f8-af4b2bac9a48-config podName:f7673917-9864-467b-b1f8-af4b2bac9a48 nodeName:}" failed. No retries permitted until 2025-12-11 08:13:35.889911509 +0000 UTC m=+148.618430564 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/f7673917-9864-467b-b1f8-af4b2bac9a48-config") pod "service-ca-operator-777779d784-6f4m5" (UID: "f7673917-9864-467b-b1f8-af4b2bac9a48") : failed to sync configmap cache: timed out waiting for the condition Dec 11 08:13:35 crc kubenswrapper[4860]: E1211 08:13:35.389996 4860 secret.go:188] Couldn't get secret openshift-machine-config-operator/machine-config-server-tls: failed to sync secret cache: timed out waiting for the condition Dec 11 08:13:35 crc kubenswrapper[4860]: E1211 08:13:35.390053 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/93550904-9295-4c4a-98bb-7a228c2699bb-certs podName:93550904-9295-4c4a-98bb-7a228c2699bb nodeName:}" failed. No retries permitted until 2025-12-11 08:13:35.890046064 +0000 UTC m=+148.618565119 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "certs" (UniqueName: "kubernetes.io/secret/93550904-9295-4c4a-98bb-7a228c2699bb-certs") pod "machine-config-server-9wb7z" (UID: "93550904-9295-4c4a-98bb-7a228c2699bb") : failed to sync secret cache: timed out waiting for the condition Dec 11 08:13:35 crc kubenswrapper[4860]: I1211 08:13:35.407388 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 11 08:13:35 crc kubenswrapper[4860]: I1211 08:13:35.412200 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:13:35 crc kubenswrapper[4860]: E1211 08:13:35.412819 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:13:35.912784754 +0000 UTC m=+148.641303849 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:13:35 crc kubenswrapper[4860]: I1211 08:13:35.425741 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 11 08:13:35 crc kubenswrapper[4860]: I1211 08:13:35.446286 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 11 08:13:35 crc kubenswrapper[4860]: I1211 08:13:35.466507 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 11 08:13:35 crc kubenswrapper[4860]: I1211 08:13:35.490104 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 11 08:13:35 crc kubenswrapper[4860]: I1211 08:13:35.506853 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 11 08:13:35 crc kubenswrapper[4860]: I1211 08:13:35.515033 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zf572\" (UID: \"38005ee2-6635-498c-b156-ac57b4f2f4d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-zf572" Dec 11 08:13:35 crc kubenswrapper[4860]: E1211 08:13:35.515470 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:13:36.015451165 +0000 UTC m=+148.743970230 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zf572" (UID: "38005ee2-6635-498c-b156-ac57b4f2f4d9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:13:35 crc kubenswrapper[4860]: I1211 08:13:35.526500 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 11 08:13:35 crc kubenswrapper[4860]: I1211 08:13:35.546941 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 11 08:13:35 crc kubenswrapper[4860]: I1211 08:13:35.568663 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 11 08:13:35 crc kubenswrapper[4860]: I1211 08:13:35.586314 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 11 08:13:35 crc kubenswrapper[4860]: I1211 08:13:35.606859 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 11 08:13:35 crc kubenswrapper[4860]: I1211 08:13:35.616468 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:13:35 crc kubenswrapper[4860]: E1211 08:13:35.616757 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:13:36.116723248 +0000 UTC m=+148.845242343 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:13:35 crc kubenswrapper[4860]: I1211 08:13:35.616902 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zf572\" (UID: \"38005ee2-6635-498c-b156-ac57b4f2f4d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-zf572" Dec 11 08:13:35 crc kubenswrapper[4860]: E1211 08:13:35.617446 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:13:36.117429863 +0000 UTC m=+148.845948958 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zf572" (UID: "38005ee2-6635-498c-b156-ac57b4f2f4d9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:13:35 crc kubenswrapper[4860]: I1211 08:13:35.626621 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 11 08:13:35 crc kubenswrapper[4860]: I1211 08:13:35.673342 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qww7r\" (UniqueName: \"kubernetes.io/projected/e55e20fb-bef2-4fee-b6a7-60e2d9d61faa-kube-api-access-qww7r\") pod \"cluster-samples-operator-665b6dd947-z5lg5\" (UID: \"e55e20fb-bef2-4fee-b6a7-60e2d9d61faa\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-z5lg5" Dec 11 08:13:35 crc kubenswrapper[4860]: I1211 08:13:35.695354 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hkdsn\" (UniqueName: \"kubernetes.io/projected/b7feb2dc-f1f2-4228-9974-8a771fe98f3c-kube-api-access-hkdsn\") pod \"controller-manager-879f6c89f-24dmg\" (UID: \"b7feb2dc-f1f2-4228-9974-8a771fe98f3c\") " pod="openshift-controller-manager/controller-manager-879f6c89f-24dmg" Dec 11 08:13:35 crc kubenswrapper[4860]: I1211 08:13:35.704774 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ngxkg\" (UniqueName: \"kubernetes.io/projected/2c04dbaf-f385-4cee-bccf-8506ecdeb680-kube-api-access-ngxkg\") pod \"authentication-operator-69f744f599-j5kxr\" (UID: \"2c04dbaf-f385-4cee-bccf-8506ecdeb680\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-j5kxr" Dec 11 08:13:35 crc kubenswrapper[4860]: I1211 08:13:35.718410 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:13:35 crc kubenswrapper[4860]: E1211 08:13:35.719384 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:13:36.219355859 +0000 UTC m=+148.947874934 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:13:35 crc kubenswrapper[4860]: I1211 08:13:35.719791 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zf572\" (UID: \"38005ee2-6635-498c-b156-ac57b4f2f4d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-zf572" Dec 11 08:13:35 crc kubenswrapper[4860]: E1211 08:13:35.720269 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:13:36.220259059 +0000 UTC m=+148.948778134 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zf572" (UID: "38005ee2-6635-498c-b156-ac57b4f2f4d9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:13:35 crc kubenswrapper[4860]: I1211 08:13:35.733737 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m96cx\" (UniqueName: \"kubernetes.io/projected/cb55e751-9913-4a3c-a0b4-3e681129d052-kube-api-access-m96cx\") pod \"machine-api-operator-5694c8668f-nc65d\" (UID: \"cb55e751-9913-4a3c-a0b4-3e681129d052\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-nc65d" Dec 11 08:13:35 crc kubenswrapper[4860]: I1211 08:13:35.753280 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6hs5b\" (UniqueName: \"kubernetes.io/projected/fb3441e4-e612-4cfe-9f12-6497cd9176a9-kube-api-access-6hs5b\") pod \"console-operator-58897d9998-b4q7q\" (UID: \"fb3441e4-e612-4cfe-9f12-6497cd9176a9\") " pod="openshift-console-operator/console-operator-58897d9998-b4q7q" Dec 11 08:13:35 crc kubenswrapper[4860]: I1211 08:13:35.765268 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c465q\" (UniqueName: \"kubernetes.io/projected/d7d1536c-ed04-4199-b839-b94e9482a63c-kube-api-access-c465q\") pod \"console-f9d7485db-rhl5v\" (UID: \"d7d1536c-ed04-4199-b839-b94e9482a63c\") " pod="openshift-console/console-f9d7485db-rhl5v" Dec 11 08:13:35 crc kubenswrapper[4860]: I1211 08:13:35.774020 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-nc65d" Dec 11 08:13:35 crc kubenswrapper[4860]: I1211 08:13:35.787398 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-24dmg" Dec 11 08:13:35 crc kubenswrapper[4860]: I1211 08:13:35.788156 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-67q4s\" (UniqueName: \"kubernetes.io/projected/e84a804f-75e1-46a2-b37a-f868de516379-kube-api-access-67q4s\") pod \"openshift-config-operator-7777fb866f-rvtb7\" (UID: \"e84a804f-75e1-46a2-b37a-f868de516379\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rvtb7" Dec 11 08:13:35 crc kubenswrapper[4860]: I1211 08:13:35.801222 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rvtb7" Dec 11 08:13:35 crc kubenswrapper[4860]: I1211 08:13:35.805225 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x9wpg\" (UniqueName: \"kubernetes.io/projected/09073ec8-36df-4ffb-a43a-10611999d8cb-kube-api-access-x9wpg\") pod \"openshift-apiserver-operator-796bbdcf4f-8x4xt\" (UID: \"09073ec8-36df-4ffb-a43a-10611999d8cb\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8x4xt" Dec 11 08:13:35 crc kubenswrapper[4860]: I1211 08:13:35.818098 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-rhl5v" Dec 11 08:13:35 crc kubenswrapper[4860]: I1211 08:13:35.821174 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:13:35 crc kubenswrapper[4860]: E1211 08:13:35.821418 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:13:36.321382738 +0000 UTC m=+149.049901823 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:13:35 crc kubenswrapper[4860]: I1211 08:13:35.822416 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zf572\" (UID: \"38005ee2-6635-498c-b156-ac57b4f2f4d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-zf572" Dec 11 08:13:35 crc kubenswrapper[4860]: E1211 08:13:35.823372 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:13:36.323347695 +0000 UTC m=+149.051866790 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zf572" (UID: "38005ee2-6635-498c-b156-ac57b4f2f4d9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:13:35 crc kubenswrapper[4860]: I1211 08:13:35.826188 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 11 08:13:35 crc kubenswrapper[4860]: I1211 08:13:35.829449 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9da229ee-dedc-417f-aec7-d2d1d9407a41-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-lbrj7\" (UID: \"9da229ee-dedc-417f-aec7-d2d1d9407a41\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lbrj7" Dec 11 08:13:35 crc kubenswrapper[4860]: I1211 08:13:35.848110 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 11 08:13:35 crc kubenswrapper[4860]: I1211 08:13:35.858133 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-b4q7q" Dec 11 08:13:35 crc kubenswrapper[4860]: I1211 08:13:35.868770 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 11 08:13:35 crc kubenswrapper[4860]: I1211 08:13:35.876065 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-z5lg5" Dec 11 08:13:35 crc kubenswrapper[4860]: I1211 08:13:35.887389 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 11 08:13:35 crc kubenswrapper[4860]: I1211 08:13:35.905432 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-j5kxr" Dec 11 08:13:35 crc kubenswrapper[4860]: I1211 08:13:35.906840 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 11 08:13:35 crc kubenswrapper[4860]: I1211 08:13:35.924990 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:13:35 crc kubenswrapper[4860]: I1211 08:13:35.925158 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/03384260-7d59-42d5-ab3c-b411e9ed88d1-profile-collector-cert\") pod \"olm-operator-6b444d44fb-svdpj\" (UID: \"03384260-7d59-42d5-ab3c-b411e9ed88d1\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-svdpj" Dec 11 08:13:35 crc kubenswrapper[4860]: I1211 08:13:35.925192 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0f6c4496-7751-49db-adc2-18099589c708-secret-volume\") pod \"collect-profiles-29424000-nf7zl\" (UID: \"0f6c4496-7751-49db-adc2-18099589c708\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424000-nf7zl" Dec 11 08:13:35 crc kubenswrapper[4860]: I1211 08:13:35.925219 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/64e3a9e4-d299-4349-b493-994630f747a5-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-2z7ml\" (UID: \"64e3a9e4-d299-4349-b493-994630f747a5\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-2z7ml" Dec 11 08:13:35 crc kubenswrapper[4860]: I1211 08:13:35.925244 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0f6c4496-7751-49db-adc2-18099589c708-config-volume\") pod \"collect-profiles-29424000-nf7zl\" (UID: \"0f6c4496-7751-49db-adc2-18099589c708\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424000-nf7zl" Dec 11 08:13:35 crc kubenswrapper[4860]: I1211 08:13:35.925268 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f7673917-9864-467b-b1f8-af4b2bac9a48-serving-cert\") pod \"service-ca-operator-777779d784-6f4m5\" (UID: \"f7673917-9864-467b-b1f8-af4b2bac9a48\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-6f4m5" Dec 11 08:13:35 crc kubenswrapper[4860]: I1211 08:13:35.925304 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c78b27aa-097e-4308-ad5f-1d609b6b1b6c-apiservice-cert\") pod \"packageserver-d55dfcdfc-fsfs7\" (UID: \"c78b27aa-097e-4308-ad5f-1d609b6b1b6c\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fsfs7" Dec 11 08:13:35 crc kubenswrapper[4860]: I1211 08:13:35.925328 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/c4c320f7-3b40-4edd-8e25-40b477e1d374-profile-collector-cert\") pod \"catalog-operator-68c6474976-s6qsz\" (UID: \"c4c320f7-3b40-4edd-8e25-40b477e1d374\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-s6qsz" Dec 11 08:13:35 crc kubenswrapper[4860]: I1211 08:13:35.925355 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/c4c320f7-3b40-4edd-8e25-40b477e1d374-srv-cert\") pod \"catalog-operator-68c6474976-s6qsz\" (UID: \"c4c320f7-3b40-4edd-8e25-40b477e1d374\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-s6qsz" Dec 11 08:13:35 crc kubenswrapper[4860]: I1211 08:13:35.925377 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/03384260-7d59-42d5-ab3c-b411e9ed88d1-srv-cert\") pod \"olm-operator-6b444d44fb-svdpj\" (UID: \"03384260-7d59-42d5-ab3c-b411e9ed88d1\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-svdpj" Dec 11 08:13:35 crc kubenswrapper[4860]: I1211 08:13:35.925417 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/93550904-9295-4c4a-98bb-7a228c2699bb-node-bootstrap-token\") pod \"machine-config-server-9wb7z\" (UID: \"93550904-9295-4c4a-98bb-7a228c2699bb\") " pod="openshift-machine-config-operator/machine-config-server-9wb7z" Dec 11 08:13:35 crc kubenswrapper[4860]: I1211 08:13:35.925510 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f7673917-9864-467b-b1f8-af4b2bac9a48-config\") pod \"service-ca-operator-777779d784-6f4m5\" (UID: \"f7673917-9864-467b-b1f8-af4b2bac9a48\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-6f4m5" Dec 11 08:13:35 crc kubenswrapper[4860]: I1211 08:13:35.925673 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/220b6958-d8ee-4733-a93a-35d2ced9f150-metrics-tls\") pod \"dns-default-sbcxd\" (UID: \"220b6958-d8ee-4733-a93a-35d2ced9f150\") " pod="openshift-dns/dns-default-sbcxd" Dec 11 08:13:35 crc kubenswrapper[4860]: I1211 08:13:35.925709 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/93550904-9295-4c4a-98bb-7a228c2699bb-certs\") pod \"machine-config-server-9wb7z\" (UID: \"93550904-9295-4c4a-98bb-7a228c2699bb\") " pod="openshift-machine-config-operator/machine-config-server-9wb7z" Dec 11 08:13:35 crc kubenswrapper[4860]: I1211 08:13:35.925775 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/220b6958-d8ee-4733-a93a-35d2ced9f150-config-volume\") pod \"dns-default-sbcxd\" (UID: \"220b6958-d8ee-4733-a93a-35d2ced9f150\") " pod="openshift-dns/dns-default-sbcxd" Dec 11 08:13:35 crc kubenswrapper[4860]: I1211 08:13:35.925804 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c78b27aa-097e-4308-ad5f-1d609b6b1b6c-webhook-cert\") pod \"packageserver-d55dfcdfc-fsfs7\" (UID: \"c78b27aa-097e-4308-ad5f-1d609b6b1b6c\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fsfs7" Dec 11 08:13:35 crc kubenswrapper[4860]: I1211 08:13:35.930457 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c78b27aa-097e-4308-ad5f-1d609b6b1b6c-webhook-cert\") pod \"packageserver-d55dfcdfc-fsfs7\" (UID: \"c78b27aa-097e-4308-ad5f-1d609b6b1b6c\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fsfs7" Dec 11 08:13:35 crc kubenswrapper[4860]: I1211 08:13:35.932429 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/c4c320f7-3b40-4edd-8e25-40b477e1d374-profile-collector-cert\") pod \"catalog-operator-68c6474976-s6qsz\" (UID: \"c4c320f7-3b40-4edd-8e25-40b477e1d374\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-s6qsz" Dec 11 08:13:35 crc kubenswrapper[4860]: I1211 08:13:35.933251 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0f6c4496-7751-49db-adc2-18099589c708-config-volume\") pod \"collect-profiles-29424000-nf7zl\" (UID: \"0f6c4496-7751-49db-adc2-18099589c708\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424000-nf7zl" Dec 11 08:13:35 crc kubenswrapper[4860]: I1211 08:13:35.933283 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lbrj7" Dec 11 08:13:35 crc kubenswrapper[4860]: E1211 08:13:35.933539 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:13:36.433517313 +0000 UTC m=+149.162036468 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:13:35 crc kubenswrapper[4860]: I1211 08:13:35.934223 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f7673917-9864-467b-b1f8-af4b2bac9a48-config\") pod \"service-ca-operator-777779d784-6f4m5\" (UID: \"f7673917-9864-467b-b1f8-af4b2bac9a48\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-6f4m5" Dec 11 08:13:35 crc kubenswrapper[4860]: I1211 08:13:35.939065 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/64e3a9e4-d299-4349-b493-994630f747a5-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-2z7ml\" (UID: \"64e3a9e4-d299-4349-b493-994630f747a5\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-2z7ml" Dec 11 08:13:35 crc kubenswrapper[4860]: I1211 08:13:35.940881 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/03384260-7d59-42d5-ab3c-b411e9ed88d1-profile-collector-cert\") pod \"olm-operator-6b444d44fb-svdpj\" (UID: \"03384260-7d59-42d5-ab3c-b411e9ed88d1\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-svdpj" Dec 11 08:13:35 crc kubenswrapper[4860]: I1211 08:13:35.940935 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/c4c320f7-3b40-4edd-8e25-40b477e1d374-srv-cert\") pod \"catalog-operator-68c6474976-s6qsz\" (UID: \"c4c320f7-3b40-4edd-8e25-40b477e1d374\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-s6qsz" Dec 11 08:13:35 crc kubenswrapper[4860]: I1211 08:13:35.942259 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0f6c4496-7751-49db-adc2-18099589c708-secret-volume\") pod \"collect-profiles-29424000-nf7zl\" (UID: \"0f6c4496-7751-49db-adc2-18099589c708\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424000-nf7zl" Dec 11 08:13:35 crc kubenswrapper[4860]: I1211 08:13:35.942767 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c78b27aa-097e-4308-ad5f-1d609b6b1b6c-apiservice-cert\") pod \"packageserver-d55dfcdfc-fsfs7\" (UID: \"c78b27aa-097e-4308-ad5f-1d609b6b1b6c\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fsfs7" Dec 11 08:13:35 crc kubenswrapper[4860]: I1211 08:13:35.943005 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f7673917-9864-467b-b1f8-af4b2bac9a48-serving-cert\") pod \"service-ca-operator-777779d784-6f4m5\" (UID: \"f7673917-9864-467b-b1f8-af4b2bac9a48\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-6f4m5" Dec 11 08:13:35 crc kubenswrapper[4860]: I1211 08:13:35.947808 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/03384260-7d59-42d5-ab3c-b411e9ed88d1-srv-cert\") pod \"olm-operator-6b444d44fb-svdpj\" (UID: \"03384260-7d59-42d5-ab3c-b411e9ed88d1\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-svdpj" Dec 11 08:13:35 crc kubenswrapper[4860]: I1211 08:13:35.957414 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8x4xt" Dec 11 08:13:35 crc kubenswrapper[4860]: I1211 08:13:35.960859 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mtwdk\" (UniqueName: \"kubernetes.io/projected/df36e993-8a70-42c5-ac22-508485572d35-kube-api-access-mtwdk\") pod \"apiserver-7bbb656c7d-cpflv\" (UID: \"df36e993-8a70-42c5-ac22-508485572d35\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cpflv" Dec 11 08:13:35 crc kubenswrapper[4860]: I1211 08:13:35.969224 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rkfd4\" (UniqueName: \"kubernetes.io/projected/c094602b-e1ec-4b9a-9640-cd6250a1fd94-kube-api-access-rkfd4\") pod \"openshift-controller-manager-operator-756b6f6bc6-gkg77\" (UID: \"c094602b-e1ec-4b9a-9640-cd6250a1fd94\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gkg77" Dec 11 08:13:35 crc kubenswrapper[4860]: I1211 08:13:35.983829 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x629f\" (UniqueName: \"kubernetes.io/projected/e86ba495-65e4-4539-928e-366376df64d8-kube-api-access-x629f\") pod \"apiserver-76f77b778f-jxft9\" (UID: \"e86ba495-65e4-4539-928e-366376df64d8\") " pod="openshift-apiserver/apiserver-76f77b778f-jxft9" Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.000204 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8k4gt\" (UniqueName: \"kubernetes.io/projected/ccbceeb2-58ab-4409-82df-5c99189569b8-kube-api-access-8k4gt\") pod \"oauth-openshift-558db77b4-dz7kr\" (UID: \"ccbceeb2-58ab-4409-82df-5c99189569b8\") " pod="openshift-authentication/oauth-openshift-558db77b4-dz7kr" Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.008879 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-24dmg"] Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.009002 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 11 08:13:36 crc kubenswrapper[4860]: W1211 08:13:36.024265 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb7feb2dc_f1f2_4228_9974_8a771fe98f3c.slice/crio-9ba7a684a61fa08b505109d09f08a0e2f9b26c3043f14821e92cbf51299f2cd0 WatchSource:0}: Error finding container 9ba7a684a61fa08b505109d09f08a0e2f9b26c3043f14821e92cbf51299f2cd0: Status 404 returned error can't find the container with id 9ba7a684a61fa08b505109d09f08a0e2f9b26c3043f14821e92cbf51299f2cd0 Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.028060 4860 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.029910 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zf572\" (UID: \"38005ee2-6635-498c-b156-ac57b4f2f4d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-zf572" Dec 11 08:13:36 crc kubenswrapper[4860]: E1211 08:13:36.030265 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:13:36.530254402 +0000 UTC m=+149.258773457 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zf572" (UID: "38005ee2-6635-498c-b156-ac57b4f2f4d9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.041546 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-jxft9" Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.047871 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.059291 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cpflv" Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.072539 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.075753 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/220b6958-d8ee-4733-a93a-35d2ced9f150-config-volume\") pod \"dns-default-sbcxd\" (UID: \"220b6958-d8ee-4733-a93a-35d2ced9f150\") " pod="openshift-dns/dns-default-sbcxd" Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.086442 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.103812 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-rvtb7"] Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.107179 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.120808 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/220b6958-d8ee-4733-a93a-35d2ced9f150-metrics-tls\") pod \"dns-default-sbcxd\" (UID: \"220b6958-d8ee-4733-a93a-35d2ced9f150\") " pod="openshift-dns/dns-default-sbcxd" Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.125837 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.130945 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:13:36 crc kubenswrapper[4860]: E1211 08:13:36.131046 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:13:36.631015818 +0000 UTC m=+149.359534873 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.131297 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zf572\" (UID: \"38005ee2-6635-498c-b156-ac57b4f2f4d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-zf572" Dec 11 08:13:36 crc kubenswrapper[4860]: E1211 08:13:36.132039 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:13:36.632030592 +0000 UTC m=+149.360549647 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zf572" (UID: "38005ee2-6635-498c-b156-ac57b4f2f4d9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.146759 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gkg77" Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.146764 4860 request.go:700] Waited for 1.92262551s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-ingress-canary/configmaps?fieldSelector=metadata.name%3Dopenshift-service-ca.crt&limit=500&resourceVersion=0 Dec 11 08:13:36 crc kubenswrapper[4860]: W1211 08:13:36.149068 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode84a804f_75e1_46a2_b37a_f868de516379.slice/crio-33810f8323c7532b28a7a20a6320e47b0ede4ce6f0dd24394a9a7e378c9c9288 WatchSource:0}: Error finding container 33810f8323c7532b28a7a20a6320e47b0ede4ce6f0dd24394a9a7e378c9c9288: Status 404 returned error can't find the container with id 33810f8323c7532b28a7a20a6320e47b0ede4ce6f0dd24394a9a7e378c9c9288 Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.149210 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.166307 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.169872 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-dz7kr" Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.193238 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.206761 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.227686 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.231513 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-nc65d"] Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.232797 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.233067 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-z5lg5"] Dec 11 08:13:36 crc kubenswrapper[4860]: E1211 08:13:36.233301 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:13:36.733278074 +0000 UTC m=+149.461797129 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.240245 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/93550904-9295-4c4a-98bb-7a228c2699bb-node-bootstrap-token\") pod \"machine-config-server-9wb7z\" (UID: \"93550904-9295-4c4a-98bb-7a228c2699bb\") " pod="openshift-machine-config-operator/machine-config-server-9wb7z" Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.247462 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.263501 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/93550904-9295-4c4a-98bb-7a228c2699bb-certs\") pod \"machine-config-server-9wb7z\" (UID: \"93550904-9295-4c4a-98bb-7a228c2699bb\") " pod="openshift-machine-config-operator/machine-config-server-9wb7z" Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.313293 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9xdkf\" (UniqueName: \"kubernetes.io/projected/e963dd44-250d-4467-980c-1b67ef1bdd93-kube-api-access-9xdkf\") pod \"machine-approver-56656f9798-wrqbs\" (UID: \"e963dd44-250d-4467-980c-1b67ef1bdd93\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wrqbs" Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.331975 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/38005ee2-6635-498c-b156-ac57b4f2f4d9-bound-sa-token\") pod \"image-registry-697d97f7c8-zf572\" (UID: \"38005ee2-6635-498c-b156-ac57b4f2f4d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-zf572" Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.334177 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-rhl5v"] Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.336396 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zf572\" (UID: \"38005ee2-6635-498c-b156-ac57b4f2f4d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-zf572" Dec 11 08:13:36 crc kubenswrapper[4860]: E1211 08:13:36.336691 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:13:36.836680061 +0000 UTC m=+149.565199116 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zf572" (UID: "38005ee2-6635-498c-b156-ac57b4f2f4d9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.352212 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/adb430a2-20a7-4e9d-8f92-446b56e36e4f-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-q6ck8\" (UID: \"adb430a2-20a7-4e9d-8f92-446b56e36e4f\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q6ck8" Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.370809 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fp9sr\" (UniqueName: \"kubernetes.io/projected/adb430a2-20a7-4e9d-8f92-446b56e36e4f-kube-api-access-fp9sr\") pod \"cluster-image-registry-operator-dc59b4c8b-q6ck8\" (UID: \"adb430a2-20a7-4e9d-8f92-446b56e36e4f\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q6ck8" Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.395804 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-24srs\" (UniqueName: \"kubernetes.io/projected/38005ee2-6635-498c-b156-ac57b4f2f4d9-kube-api-access-24srs\") pod \"image-registry-697d97f7c8-zf572\" (UID: \"38005ee2-6635-498c-b156-ac57b4f2f4d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-zf572" Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.395883 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-j5kxr"] Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.405410 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-cpflv"] Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.412718 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mk9kz\" (UniqueName: \"kubernetes.io/projected/cefdf543-6d3a-4e62-9875-ca26b060f437-kube-api-access-mk9kz\") pod \"machine-config-operator-74547568cd-6q4dc\" (UID: \"cefdf543-6d3a-4e62-9875-ca26b060f437\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6q4dc" Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.424493 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8dd6t\" (UniqueName: \"kubernetes.io/projected/30bec2ef-4635-4766-a34c-423c2e7f471b-kube-api-access-8dd6t\") pod \"router-default-5444994796-n8wrd\" (UID: \"30bec2ef-4635-4766-a34c-423c2e7f471b\") " pod="openshift-ingress/router-default-5444994796-n8wrd" Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.424833 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-nc65d" event={"ID":"cb55e751-9913-4a3c-a0b4-3e681129d052","Type":"ContainerStarted","Data":"ff807da7b3e45d43624cc59df44d33db403af359b40a4c64e4f94988e2ea1040"} Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.429328 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-jxft9"] Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.430163 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lbrj7"] Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.431560 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"346e6628ccd57d92117e71839c946bfbe198e406460cf2cfef9b6b5ce6c8e14f"} Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.431610 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"207c229ede0f0c62724a4b823285edee5849af2c3a70cf863896cf8f9f26dfb0"} Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.435555 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-z5lg5" event={"ID":"e55e20fb-bef2-4fee-b6a7-60e2d9d61faa","Type":"ContainerStarted","Data":"bc608386dcbb5418e571f25d5d6c6c070b20b822aa7741d65890456bb37fb563"} Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.436738 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"d12ad54af2804b1d62ab3c897bb136eaadcee7f3102e73f4f74e242338f38fa3"} Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.436775 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"aec9e8021486562cfa309ed4410b5d287169fc267a285faa7da4f20ba3e5911f"} Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.437750 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:13:36 crc kubenswrapper[4860]: E1211 08:13:36.437984 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:13:36.937958264 +0000 UTC m=+149.666477329 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.438369 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zf572\" (UID: \"38005ee2-6635-498c-b156-ac57b4f2f4d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-zf572" Dec 11 08:13:36 crc kubenswrapper[4860]: E1211 08:13:36.438838 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:13:36.938827325 +0000 UTC m=+149.667346380 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zf572" (UID: "38005ee2-6635-498c-b156-ac57b4f2f4d9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.445487 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2vgdm\" (UniqueName: \"kubernetes.io/projected/cf1bd6cb-c047-423c-9d36-7246a6108844-kube-api-access-2vgdm\") pod \"downloads-7954f5f757-mj2tr\" (UID: \"cf1bd6cb-c047-423c-9d36-7246a6108844\") " pod="openshift-console/downloads-7954f5f757-mj2tr" Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.445687 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-rhl5v" event={"ID":"d7d1536c-ed04-4199-b839-b94e9482a63c","Type":"ContainerStarted","Data":"f397ce5904d9e55432c6e8b65b71e54575de3faadf6cc259328986bc042780ce"} Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.452679 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rvtb7" event={"ID":"e84a804f-75e1-46a2-b37a-f868de516379","Type":"ContainerStarted","Data":"976b0d49bb1ffd80ce2033f3a3f58f15303bbb4376dc694d6c777648155bcd9a"} Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.452728 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rvtb7" event={"ID":"e84a804f-75e1-46a2-b37a-f868de516379","Type":"ContainerStarted","Data":"33810f8323c7532b28a7a20a6320e47b0ede4ce6f0dd24394a9a7e378c9c9288"} Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.455815 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-24dmg" event={"ID":"b7feb2dc-f1f2-4228-9974-8a771fe98f3c","Type":"ContainerStarted","Data":"5b3be03c9a9ede15a987ac40ff9f744b5f05266925533e6a30991b110196cf7e"} Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.455871 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-24dmg" event={"ID":"b7feb2dc-f1f2-4228-9974-8a771fe98f3c","Type":"ContainerStarted","Data":"9ba7a684a61fa08b505109d09f08a0e2f9b26c3043f14821e92cbf51299f2cd0"} Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.456495 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-24dmg" Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.457600 4860 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-24dmg container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.8:8443/healthz\": dial tcp 10.217.0.8:8443: connect: connection refused" start-of-body= Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.457763 4860 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-24dmg" podUID="b7feb2dc-f1f2-4228-9974-8a771fe98f3c" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.8:8443/healthz\": dial tcp 10.217.0.8:8443: connect: connection refused" Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.460795 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gspwb\" (UniqueName: \"kubernetes.io/projected/5b8b374c-b131-402c-8e3b-02d93d416489-kube-api-access-gspwb\") pod \"migrator-59844c95c7-l4m2j\" (UID: \"5b8b374c-b131-402c-8e3b-02d93d416489\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-l4m2j" Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.478083 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gkg77"] Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.478157 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8x4xt"] Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.482819 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rcfdr\" (UniqueName: \"kubernetes.io/projected/1fff60d5-55fd-41d8-9225-251277e8e3c2-kube-api-access-rcfdr\") pod \"dns-operator-744455d44c-g4bql\" (UID: \"1fff60d5-55fd-41d8-9225-251277e8e3c2\") " pod="openshift-dns-operator/dns-operator-744455d44c-g4bql" Dec 11 08:13:36 crc kubenswrapper[4860]: W1211 08:13:36.501820 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod09073ec8_36df_4ffb_a43a_10611999d8cb.slice/crio-950466ad812f36761fe49ab92d1febcdb89c6daf0632b22763fd519793f6159e WatchSource:0}: Error finding container 950466ad812f36761fe49ab92d1febcdb89c6daf0632b22763fd519793f6159e: Status 404 returned error can't find the container with id 950466ad812f36761fe49ab92d1febcdb89c6daf0632b22763fd519793f6159e Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.509823 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hckxn\" (UniqueName: \"kubernetes.io/projected/3e5d90ab-884d-420b-a284-74f4b578a864-kube-api-access-hckxn\") pod \"package-server-manager-789f6589d5-5zqnk\" (UID: \"3e5d90ab-884d-420b-a284-74f4b578a864\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5zqnk" Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.520318 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-b4q7q"] Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.528595 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-47grh\" (UniqueName: \"kubernetes.io/projected/03384260-7d59-42d5-ab3c-b411e9ed88d1-kube-api-access-47grh\") pod \"olm-operator-6b444d44fb-svdpj\" (UID: \"03384260-7d59-42d5-ab3c-b411e9ed88d1\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-svdpj" Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.530883 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-dz7kr"] Dec 11 08:13:36 crc kubenswrapper[4860]: W1211 08:13:36.531612 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode86ba495_65e4_4539_928e_366376df64d8.slice/crio-6d3d36b248589249a3ffa5f1e1cfa8fa5b8d57f8b6a1435c9052717801c68439 WatchSource:0}: Error finding container 6d3d36b248589249a3ffa5f1e1cfa8fa5b8d57f8b6a1435c9052717801c68439: Status 404 returned error can't find the container with id 6d3d36b248589249a3ffa5f1e1cfa8fa5b8d57f8b6a1435c9052717801c68439 Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.537812 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wrqbs" Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.539124 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:13:36 crc kubenswrapper[4860]: E1211 08:13:36.540190 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:13:37.04017017 +0000 UTC m=+149.768689245 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.540254 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zf572\" (UID: \"38005ee2-6635-498c-b156-ac57b4f2f4d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-zf572" Dec 11 08:13:36 crc kubenswrapper[4860]: E1211 08:13:36.543567 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:13:37.043557316 +0000 UTC m=+149.772076371 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zf572" (UID: "38005ee2-6635-498c-b156-ac57b4f2f4d9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.547165 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b08141e2-41b8-45f4-bb40-c319621719c8-bound-sa-token\") pod \"ingress-operator-5b745b69d9-h7zxk\" (UID: \"b08141e2-41b8-45f4-bb40-c319621719c8\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-h7zxk" Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.564208 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4bzjr\" (UniqueName: \"kubernetes.io/projected/c78b27aa-097e-4308-ad5f-1d609b6b1b6c-kube-api-access-4bzjr\") pod \"packageserver-d55dfcdfc-fsfs7\" (UID: \"c78b27aa-097e-4308-ad5f-1d609b6b1b6c\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fsfs7" Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.573092 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q6ck8" Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.581499 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-krqtd\" (UniqueName: \"kubernetes.io/projected/ad724960-18d1-4e83-944c-4cdeb033a436-kube-api-access-krqtd\") pod \"marketplace-operator-79b997595-ttrm2\" (UID: \"ad724960-18d1-4e83-944c-4cdeb033a436\") " pod="openshift-marketplace/marketplace-operator-79b997595-ttrm2" Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.607111 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-58dpf\" (UniqueName: \"kubernetes.io/projected/c4c320f7-3b40-4edd-8e25-40b477e1d374-kube-api-access-58dpf\") pod \"catalog-operator-68c6474976-s6qsz\" (UID: \"c4c320f7-3b40-4edd-8e25-40b477e1d374\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-s6qsz" Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.620958 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6q4dc" Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.624997 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xgz7t\" (UniqueName: \"kubernetes.io/projected/eceb74a4-c695-46e9-ba5f-b5145d20d968-kube-api-access-xgz7t\") pod \"csi-hostpathplugin-kbz2w\" (UID: \"eceb74a4-c695-46e9-ba5f-b5145d20d968\") " pod="hostpath-provisioner/csi-hostpathplugin-kbz2w" Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.626510 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-mj2tr" Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.639911 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-g4bql" Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.641021 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:13:36 crc kubenswrapper[4860]: E1211 08:13:36.641492 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:13:37.141473224 +0000 UTC m=+149.869992279 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.641837 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h8x5s\" (UniqueName: \"kubernetes.io/projected/64e3a9e4-d299-4349-b493-994630f747a5-kube-api-access-h8x5s\") pod \"multus-admission-controller-857f4d67dd-2z7ml\" (UID: \"64e3a9e4-d299-4349-b493-994630f747a5\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-2z7ml" Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.662572 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vmgfp\" (UniqueName: \"kubernetes.io/projected/73ab8adf-cf47-4482-bbdb-8fb6c3edc87a-kube-api-access-vmgfp\") pod \"route-controller-manager-6576b87f9c-87b7k\" (UID: \"73ab8adf-cf47-4482-bbdb-8fb6c3edc87a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-87b7k" Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.669979 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-n8wrd" Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.678755 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-ttrm2" Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.682949 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g6g9q\" (UniqueName: \"kubernetes.io/projected/b08141e2-41b8-45f4-bb40-c319621719c8-kube-api-access-g6g9q\") pod \"ingress-operator-5b745b69d9-h7zxk\" (UID: \"b08141e2-41b8-45f4-bb40-c319621719c8\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-h7zxk" Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.712771 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bmjqc\" (UniqueName: \"kubernetes.io/projected/5c0bfb74-4985-47a1-ade8-17afad50fabe-kube-api-access-bmjqc\") pod \"machine-config-controller-84d6567774-jgk9d\" (UID: \"5c0bfb74-4985-47a1-ade8-17afad50fabe\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jgk9d" Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.720579 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5zqnk" Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.729461 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-grc8b\" (UniqueName: \"kubernetes.io/projected/93550904-9295-4c4a-98bb-7a228c2699bb-kube-api-access-grc8b\") pod \"machine-config-server-9wb7z\" (UID: \"93550904-9295-4c4a-98bb-7a228c2699bb\") " pod="openshift-machine-config-operator/machine-config-server-9wb7z" Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.729916 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-l4m2j" Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.744098 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zf572\" (UID: \"38005ee2-6635-498c-b156-ac57b4f2f4d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-zf572" Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.744564 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-2z7ml" Dec 11 08:13:36 crc kubenswrapper[4860]: E1211 08:13:36.745411 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:13:37.245388218 +0000 UTC m=+149.973907353 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zf572" (UID: "38005ee2-6635-498c-b156-ac57b4f2f4d9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.754076 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fsfs7" Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.764347 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m8497\" (UniqueName: \"kubernetes.io/projected/0f6c4496-7751-49db-adc2-18099589c708-kube-api-access-m8497\") pod \"collect-profiles-29424000-nf7zl\" (UID: \"0f6c4496-7751-49db-adc2-18099589c708\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424000-nf7zl" Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.774092 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-svdpj" Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.775104 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9zrk8\" (UniqueName: \"kubernetes.io/projected/f7673917-9864-467b-b1f8-af4b2bac9a48-kube-api-access-9zrk8\") pod \"service-ca-operator-777779d784-6f4m5\" (UID: \"f7673917-9864-467b-b1f8-af4b2bac9a48\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-6f4m5" Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.795275 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-s6qsz" Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.799248 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h6gns\" (UniqueName: \"kubernetes.io/projected/c3ce3c34-9b72-4964-b59a-f6b3c0a0e5c8-kube-api-access-h6gns\") pod \"kube-storage-version-migrator-operator-b67b599dd-x28lp\" (UID: \"c3ce3c34-9b72-4964-b59a-f6b3c0a0e5c8\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-x28lp" Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.813065 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-kbz2w" Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.819617 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pt56k\" (UniqueName: \"kubernetes.io/projected/220b6958-d8ee-4733-a93a-35d2ced9f150-kube-api-access-pt56k\") pod \"dns-default-sbcxd\" (UID: \"220b6958-d8ee-4733-a93a-35d2ced9f150\") " pod="openshift-dns/dns-default-sbcxd" Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.834009 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-9wb7z" Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.845875 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.846184 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z885j\" (UniqueName: \"kubernetes.io/projected/89e3648e-71d3-4d85-aef0-aa60bc630c8f-kube-api-access-z885j\") pod \"service-ca-9c57cc56f-qx9pf\" (UID: \"89e3648e-71d3-4d85-aef0-aa60bc630c8f\") " pod="openshift-service-ca/service-ca-9c57cc56f-qx9pf" Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.846216 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/776e5c46-e988-4bb7-98e5-2d10c5fb5142-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-vc92z\" (UID: \"776e5c46-e988-4bb7-98e5-2d10c5fb5142\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vc92z" Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.846272 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eabe3a89-8c1c-41fa-a8e9-9a0afed00d4e-serving-cert\") pod \"etcd-operator-b45778765-rl2ff\" (UID: \"eabe3a89-8c1c-41fa-a8e9-9a0afed00d4e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rl2ff" Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.846295 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/eabe3a89-8c1c-41fa-a8e9-9a0afed00d4e-etcd-service-ca\") pod \"etcd-operator-b45778765-rl2ff\" (UID: \"eabe3a89-8c1c-41fa-a8e9-9a0afed00d4e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rl2ff" Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.846320 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/7a764867-5216-47fb-8211-cea78b79f707-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-76lnq\" (UID: \"7a764867-5216-47fb-8211-cea78b79f707\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-76lnq" Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.846382 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/776e5c46-e988-4bb7-98e5-2d10c5fb5142-config\") pod \"kube-apiserver-operator-766d6c64bb-vc92z\" (UID: \"776e5c46-e988-4bb7-98e5-2d10c5fb5142\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vc92z" Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.846408 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/67ae7457-ee23-4ae3-a118-fbfac6dd85bf-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-9tkth\" (UID: \"67ae7457-ee23-4ae3-a118-fbfac6dd85bf\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9tkth" Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.846428 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eabe3a89-8c1c-41fa-a8e9-9a0afed00d4e-config\") pod \"etcd-operator-b45778765-rl2ff\" (UID: \"eabe3a89-8c1c-41fa-a8e9-9a0afed00d4e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rl2ff" Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.846482 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67ae7457-ee23-4ae3-a118-fbfac6dd85bf-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-9tkth\" (UID: \"67ae7457-ee23-4ae3-a118-fbfac6dd85bf\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9tkth" Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.846506 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qfhss\" (UniqueName: \"kubernetes.io/projected/eabe3a89-8c1c-41fa-a8e9-9a0afed00d4e-kube-api-access-qfhss\") pod \"etcd-operator-b45778765-rl2ff\" (UID: \"eabe3a89-8c1c-41fa-a8e9-9a0afed00d4e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rl2ff" Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.846526 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bmqhg\" (UniqueName: \"kubernetes.io/projected/7a764867-5216-47fb-8211-cea78b79f707-kube-api-access-bmqhg\") pod \"control-plane-machine-set-operator-78cbb6b69f-76lnq\" (UID: \"7a764867-5216-47fb-8211-cea78b79f707\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-76lnq" Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.846545 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/89e3648e-71d3-4d85-aef0-aa60bc630c8f-signing-cabundle\") pod \"service-ca-9c57cc56f-qx9pf\" (UID: \"89e3648e-71d3-4d85-aef0-aa60bc630c8f\") " pod="openshift-service-ca/service-ca-9c57cc56f-qx9pf" Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.846559 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-x28lp" Dec 11 08:13:36 crc kubenswrapper[4860]: E1211 08:13:36.846848 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:13:37.346829578 +0000 UTC m=+150.075348643 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.846569 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/776e5c46-e988-4bb7-98e5-2d10c5fb5142-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-vc92z\" (UID: \"776e5c46-e988-4bb7-98e5-2d10c5fb5142\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vc92z" Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.847053 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/89e3648e-71d3-4d85-aef0-aa60bc630c8f-signing-key\") pod \"service-ca-9c57cc56f-qx9pf\" (UID: \"89e3648e-71d3-4d85-aef0-aa60bc630c8f\") " pod="openshift-service-ca/service-ca-9c57cc56f-qx9pf" Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.847106 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zf572\" (UID: \"38005ee2-6635-498c-b156-ac57b4f2f4d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-zf572" Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.847166 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/eabe3a89-8c1c-41fa-a8e9-9a0afed00d4e-etcd-client\") pod \"etcd-operator-b45778765-rl2ff\" (UID: \"eabe3a89-8c1c-41fa-a8e9-9a0afed00d4e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rl2ff" Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.847216 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5010fcd8-99df-473e-8ed3-9ae00612be20-cert\") pod \"ingress-canary-t8ffw\" (UID: \"5010fcd8-99df-473e-8ed3-9ae00612be20\") " pod="openshift-ingress-canary/ingress-canary-t8ffw" Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.847249 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/eabe3a89-8c1c-41fa-a8e9-9a0afed00d4e-etcd-ca\") pod \"etcd-operator-b45778765-rl2ff\" (UID: \"eabe3a89-8c1c-41fa-a8e9-9a0afed00d4e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rl2ff" Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.847272 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6klgl\" (UniqueName: \"kubernetes.io/projected/5010fcd8-99df-473e-8ed3-9ae00612be20-kube-api-access-6klgl\") pod \"ingress-canary-t8ffw\" (UID: \"5010fcd8-99df-473e-8ed3-9ae00612be20\") " pod="openshift-ingress-canary/ingress-canary-t8ffw" Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.847311 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/67ae7457-ee23-4ae3-a118-fbfac6dd85bf-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-9tkth\" (UID: \"67ae7457-ee23-4ae3-a118-fbfac6dd85bf\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9tkth" Dec 11 08:13:36 crc kubenswrapper[4860]: E1211 08:13:36.847650 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:13:37.347598444 +0000 UTC m=+150.076117589 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zf572" (UID: "38005ee2-6635-498c-b156-ac57b4f2f4d9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.935440 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-87b7k" Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.959263 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:13:36 crc kubenswrapper[4860]: E1211 08:13:36.959580 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:13:37.459553914 +0000 UTC m=+150.188072969 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.959698 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/eabe3a89-8c1c-41fa-a8e9-9a0afed00d4e-etcd-client\") pod \"etcd-operator-b45778765-rl2ff\" (UID: \"eabe3a89-8c1c-41fa-a8e9-9a0afed00d4e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rl2ff" Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.959933 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5010fcd8-99df-473e-8ed3-9ae00612be20-cert\") pod \"ingress-canary-t8ffw\" (UID: \"5010fcd8-99df-473e-8ed3-9ae00612be20\") " pod="openshift-ingress-canary/ingress-canary-t8ffw" Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.959990 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/eabe3a89-8c1c-41fa-a8e9-9a0afed00d4e-etcd-ca\") pod \"etcd-operator-b45778765-rl2ff\" (UID: \"eabe3a89-8c1c-41fa-a8e9-9a0afed00d4e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rl2ff" Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.960011 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6klgl\" (UniqueName: \"kubernetes.io/projected/5010fcd8-99df-473e-8ed3-9ae00612be20-kube-api-access-6klgl\") pod \"ingress-canary-t8ffw\" (UID: \"5010fcd8-99df-473e-8ed3-9ae00612be20\") " pod="openshift-ingress-canary/ingress-canary-t8ffw" Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.960035 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/67ae7457-ee23-4ae3-a118-fbfac6dd85bf-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-9tkth\" (UID: \"67ae7457-ee23-4ae3-a118-fbfac6dd85bf\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9tkth" Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.960106 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z885j\" (UniqueName: \"kubernetes.io/projected/89e3648e-71d3-4d85-aef0-aa60bc630c8f-kube-api-access-z885j\") pod \"service-ca-9c57cc56f-qx9pf\" (UID: \"89e3648e-71d3-4d85-aef0-aa60bc630c8f\") " pod="openshift-service-ca/service-ca-9c57cc56f-qx9pf" Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.960156 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/776e5c46-e988-4bb7-98e5-2d10c5fb5142-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-vc92z\" (UID: \"776e5c46-e988-4bb7-98e5-2d10c5fb5142\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vc92z" Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.960368 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eabe3a89-8c1c-41fa-a8e9-9a0afed00d4e-serving-cert\") pod \"etcd-operator-b45778765-rl2ff\" (UID: \"eabe3a89-8c1c-41fa-a8e9-9a0afed00d4e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rl2ff" Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.960408 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/eabe3a89-8c1c-41fa-a8e9-9a0afed00d4e-etcd-service-ca\") pod \"etcd-operator-b45778765-rl2ff\" (UID: \"eabe3a89-8c1c-41fa-a8e9-9a0afed00d4e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rl2ff" Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.960444 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/7a764867-5216-47fb-8211-cea78b79f707-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-76lnq\" (UID: \"7a764867-5216-47fb-8211-cea78b79f707\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-76lnq" Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.960543 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/776e5c46-e988-4bb7-98e5-2d10c5fb5142-config\") pod \"kube-apiserver-operator-766d6c64bb-vc92z\" (UID: \"776e5c46-e988-4bb7-98e5-2d10c5fb5142\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vc92z" Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.960603 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/67ae7457-ee23-4ae3-a118-fbfac6dd85bf-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-9tkth\" (UID: \"67ae7457-ee23-4ae3-a118-fbfac6dd85bf\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9tkth" Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.963067 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-h7zxk" Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.986498 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/eabe3a89-8c1c-41fa-a8e9-9a0afed00d4e-etcd-service-ca\") pod \"etcd-operator-b45778765-rl2ff\" (UID: \"eabe3a89-8c1c-41fa-a8e9-9a0afed00d4e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rl2ff" Dec 11 08:13:36 crc kubenswrapper[4860]: I1211 08:13:36.987505 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/eabe3a89-8c1c-41fa-a8e9-9a0afed00d4e-etcd-ca\") pod \"etcd-operator-b45778765-rl2ff\" (UID: \"eabe3a89-8c1c-41fa-a8e9-9a0afed00d4e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rl2ff" Dec 11 08:13:37 crc kubenswrapper[4860]: I1211 08:13:36.995271 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/776e5c46-e988-4bb7-98e5-2d10c5fb5142-config\") pod \"kube-apiserver-operator-766d6c64bb-vc92z\" (UID: \"776e5c46-e988-4bb7-98e5-2d10c5fb5142\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vc92z" Dec 11 08:13:37 crc kubenswrapper[4860]: I1211 08:13:36.998095 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eabe3a89-8c1c-41fa-a8e9-9a0afed00d4e-config\") pod \"etcd-operator-b45778765-rl2ff\" (UID: \"eabe3a89-8c1c-41fa-a8e9-9a0afed00d4e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rl2ff" Dec 11 08:13:37 crc kubenswrapper[4860]: I1211 08:13:36.998438 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67ae7457-ee23-4ae3-a118-fbfac6dd85bf-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-9tkth\" (UID: \"67ae7457-ee23-4ae3-a118-fbfac6dd85bf\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9tkth" Dec 11 08:13:37 crc kubenswrapper[4860]: I1211 08:13:37.000001 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eabe3a89-8c1c-41fa-a8e9-9a0afed00d4e-config\") pod \"etcd-operator-b45778765-rl2ff\" (UID: \"eabe3a89-8c1c-41fa-a8e9-9a0afed00d4e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rl2ff" Dec 11 08:13:37 crc kubenswrapper[4860]: I1211 08:13:37.000593 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qfhss\" (UniqueName: \"kubernetes.io/projected/eabe3a89-8c1c-41fa-a8e9-9a0afed00d4e-kube-api-access-qfhss\") pod \"etcd-operator-b45778765-rl2ff\" (UID: \"eabe3a89-8c1c-41fa-a8e9-9a0afed00d4e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rl2ff" Dec 11 08:13:37 crc kubenswrapper[4860]: I1211 08:13:37.000700 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bmqhg\" (UniqueName: \"kubernetes.io/projected/7a764867-5216-47fb-8211-cea78b79f707-kube-api-access-bmqhg\") pod \"control-plane-machine-set-operator-78cbb6b69f-76lnq\" (UID: \"7a764867-5216-47fb-8211-cea78b79f707\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-76lnq" Dec 11 08:13:37 crc kubenswrapper[4860]: I1211 08:13:37.000830 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/89e3648e-71d3-4d85-aef0-aa60bc630c8f-signing-cabundle\") pod \"service-ca-9c57cc56f-qx9pf\" (UID: \"89e3648e-71d3-4d85-aef0-aa60bc630c8f\") " pod="openshift-service-ca/service-ca-9c57cc56f-qx9pf" Dec 11 08:13:37 crc kubenswrapper[4860]: I1211 08:13:37.004093 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jgk9d" Dec 11 08:13:37 crc kubenswrapper[4860]: I1211 08:13:37.005170 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/89e3648e-71d3-4d85-aef0-aa60bc630c8f-signing-cabundle\") pod \"service-ca-9c57cc56f-qx9pf\" (UID: \"89e3648e-71d3-4d85-aef0-aa60bc630c8f\") " pod="openshift-service-ca/service-ca-9c57cc56f-qx9pf" Dec 11 08:13:37 crc kubenswrapper[4860]: I1211 08:13:37.006603 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/776e5c46-e988-4bb7-98e5-2d10c5fb5142-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-vc92z\" (UID: \"776e5c46-e988-4bb7-98e5-2d10c5fb5142\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vc92z" Dec 11 08:13:37 crc kubenswrapper[4860]: I1211 08:13:37.006657 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/89e3648e-71d3-4d85-aef0-aa60bc630c8f-signing-key\") pod \"service-ca-9c57cc56f-qx9pf\" (UID: \"89e3648e-71d3-4d85-aef0-aa60bc630c8f\") " pod="openshift-service-ca/service-ca-9c57cc56f-qx9pf" Dec 11 08:13:37 crc kubenswrapper[4860]: I1211 08:13:37.008750 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zf572\" (UID: \"38005ee2-6635-498c-b156-ac57b4f2f4d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-zf572" Dec 11 08:13:37 crc kubenswrapper[4860]: I1211 08:13:37.010190 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/67ae7457-ee23-4ae3-a118-fbfac6dd85bf-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-9tkth\" (UID: \"67ae7457-ee23-4ae3-a118-fbfac6dd85bf\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9tkth" Dec 11 08:13:37 crc kubenswrapper[4860]: E1211 08:13:37.013889 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:13:37.513868606 +0000 UTC m=+150.242387661 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zf572" (UID: "38005ee2-6635-498c-b156-ac57b4f2f4d9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:13:37 crc kubenswrapper[4860]: I1211 08:13:37.015427 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67ae7457-ee23-4ae3-a118-fbfac6dd85bf-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-9tkth\" (UID: \"67ae7457-ee23-4ae3-a118-fbfac6dd85bf\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9tkth" Dec 11 08:13:37 crc kubenswrapper[4860]: I1211 08:13:37.019689 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/eabe3a89-8c1c-41fa-a8e9-9a0afed00d4e-etcd-client\") pod \"etcd-operator-b45778765-rl2ff\" (UID: \"eabe3a89-8c1c-41fa-a8e9-9a0afed00d4e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rl2ff" Dec 11 08:13:37 crc kubenswrapper[4860]: I1211 08:13:37.021427 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/89e3648e-71d3-4d85-aef0-aa60bc630c8f-signing-key\") pod \"service-ca-9c57cc56f-qx9pf\" (UID: \"89e3648e-71d3-4d85-aef0-aa60bc630c8f\") " pod="openshift-service-ca/service-ca-9c57cc56f-qx9pf" Dec 11 08:13:37 crc kubenswrapper[4860]: I1211 08:13:37.025272 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/7a764867-5216-47fb-8211-cea78b79f707-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-76lnq\" (UID: \"7a764867-5216-47fb-8211-cea78b79f707\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-76lnq" Dec 11 08:13:37 crc kubenswrapper[4860]: I1211 08:13:37.026083 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/776e5c46-e988-4bb7-98e5-2d10c5fb5142-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-vc92z\" (UID: \"776e5c46-e988-4bb7-98e5-2d10c5fb5142\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vc92z" Dec 11 08:13:37 crc kubenswrapper[4860]: I1211 08:13:37.037220 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eabe3a89-8c1c-41fa-a8e9-9a0afed00d4e-serving-cert\") pod \"etcd-operator-b45778765-rl2ff\" (UID: \"eabe3a89-8c1c-41fa-a8e9-9a0afed00d4e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rl2ff" Dec 11 08:13:37 crc kubenswrapper[4860]: I1211 08:13:37.037838 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z885j\" (UniqueName: \"kubernetes.io/projected/89e3648e-71d3-4d85-aef0-aa60bc630c8f-kube-api-access-z885j\") pod \"service-ca-9c57cc56f-qx9pf\" (UID: \"89e3648e-71d3-4d85-aef0-aa60bc630c8f\") " pod="openshift-service-ca/service-ca-9c57cc56f-qx9pf" Dec 11 08:13:37 crc kubenswrapper[4860]: I1211 08:13:37.043157 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5010fcd8-99df-473e-8ed3-9ae00612be20-cert\") pod \"ingress-canary-t8ffw\" (UID: \"5010fcd8-99df-473e-8ed3-9ae00612be20\") " pod="openshift-ingress-canary/ingress-canary-t8ffw" Dec 11 08:13:37 crc kubenswrapper[4860]: I1211 08:13:37.050086 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6klgl\" (UniqueName: \"kubernetes.io/projected/5010fcd8-99df-473e-8ed3-9ae00612be20-kube-api-access-6klgl\") pod \"ingress-canary-t8ffw\" (UID: \"5010fcd8-99df-473e-8ed3-9ae00612be20\") " pod="openshift-ingress-canary/ingress-canary-t8ffw" Dec 11 08:13:37 crc kubenswrapper[4860]: I1211 08:13:37.058827 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29424000-nf7zl" Dec 11 08:13:37 crc kubenswrapper[4860]: I1211 08:13:37.068177 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/67ae7457-ee23-4ae3-a118-fbfac6dd85bf-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-9tkth\" (UID: \"67ae7457-ee23-4ae3-a118-fbfac6dd85bf\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9tkth" Dec 11 08:13:37 crc kubenswrapper[4860]: I1211 08:13:37.068596 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-6f4m5" Dec 11 08:13:37 crc kubenswrapper[4860]: I1211 08:13:37.088706 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bmqhg\" (UniqueName: \"kubernetes.io/projected/7a764867-5216-47fb-8211-cea78b79f707-kube-api-access-bmqhg\") pod \"control-plane-machine-set-operator-78cbb6b69f-76lnq\" (UID: \"7a764867-5216-47fb-8211-cea78b79f707\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-76lnq" Dec 11 08:13:37 crc kubenswrapper[4860]: I1211 08:13:37.094828 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-qx9pf" Dec 11 08:13:37 crc kubenswrapper[4860]: I1211 08:13:37.113478 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:13:37 crc kubenswrapper[4860]: E1211 08:13:37.114019 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:13:37.61400288 +0000 UTC m=+150.342521935 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:13:37 crc kubenswrapper[4860]: I1211 08:13:37.116181 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qfhss\" (UniqueName: \"kubernetes.io/projected/eabe3a89-8c1c-41fa-a8e9-9a0afed00d4e-kube-api-access-qfhss\") pod \"etcd-operator-b45778765-rl2ff\" (UID: \"eabe3a89-8c1c-41fa-a8e9-9a0afed00d4e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rl2ff" Dec 11 08:13:37 crc kubenswrapper[4860]: I1211 08:13:37.123589 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q6ck8"] Dec 11 08:13:37 crc kubenswrapper[4860]: I1211 08:13:37.129300 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-t8ffw" Dec 11 08:13:37 crc kubenswrapper[4860]: I1211 08:13:37.129949 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-sbcxd" Dec 11 08:13:37 crc kubenswrapper[4860]: I1211 08:13:37.135164 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/776e5c46-e988-4bb7-98e5-2d10c5fb5142-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-vc92z\" (UID: \"776e5c46-e988-4bb7-98e5-2d10c5fb5142\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vc92z" Dec 11 08:13:37 crc kubenswrapper[4860]: I1211 08:13:37.218561 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zf572\" (UID: \"38005ee2-6635-498c-b156-ac57b4f2f4d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-zf572" Dec 11 08:13:37 crc kubenswrapper[4860]: E1211 08:13:37.218850 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:13:37.718839526 +0000 UTC m=+150.447358581 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zf572" (UID: "38005ee2-6635-498c-b156-ac57b4f2f4d9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:13:37 crc kubenswrapper[4860]: I1211 08:13:37.267114 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-rl2ff" Dec 11 08:13:37 crc kubenswrapper[4860]: I1211 08:13:37.267703 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9tkth" Dec 11 08:13:37 crc kubenswrapper[4860]: W1211 08:13:37.280384 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podadb430a2_20a7_4e9d_8f92_446b56e36e4f.slice/crio-198e6dd61a9c2d3069f85232e914b4c45f8af7475867ace7ca2869bc6088c960 WatchSource:0}: Error finding container 198e6dd61a9c2d3069f85232e914b4c45f8af7475867ace7ca2869bc6088c960: Status 404 returned error can't find the container with id 198e6dd61a9c2d3069f85232e914b4c45f8af7475867ace7ca2869bc6088c960 Dec 11 08:13:37 crc kubenswrapper[4860]: I1211 08:13:37.311138 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vc92z" Dec 11 08:13:37 crc kubenswrapper[4860]: I1211 08:13:37.319333 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:13:37 crc kubenswrapper[4860]: E1211 08:13:37.319755 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:13:37.819741307 +0000 UTC m=+150.548260362 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:13:37 crc kubenswrapper[4860]: I1211 08:13:37.338183 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-76lnq" Dec 11 08:13:37 crc kubenswrapper[4860]: I1211 08:13:37.404248 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-6q4dc"] Dec 11 08:13:37 crc kubenswrapper[4860]: I1211 08:13:37.421006 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zf572\" (UID: \"38005ee2-6635-498c-b156-ac57b4f2f4d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-zf572" Dec 11 08:13:37 crc kubenswrapper[4860]: E1211 08:13:37.421411 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:13:37.921399054 +0000 UTC m=+150.649918099 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zf572" (UID: "38005ee2-6635-498c-b156-ac57b4f2f4d9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:13:37 crc kubenswrapper[4860]: I1211 08:13:37.459708 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-g4bql"] Dec 11 08:13:37 crc kubenswrapper[4860]: I1211 08:13:37.496259 4860 generic.go:334] "Generic (PLEG): container finished" podID="df36e993-8a70-42c5-ac22-508485572d35" containerID="736bade5ba0a799b475bb5f23e435a45842aff6a4209ce7350fb33d9511bfc51" exitCode=0 Dec 11 08:13:37 crc kubenswrapper[4860]: I1211 08:13:37.496346 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cpflv" event={"ID":"df36e993-8a70-42c5-ac22-508485572d35","Type":"ContainerDied","Data":"736bade5ba0a799b475bb5f23e435a45842aff6a4209ce7350fb33d9511bfc51"} Dec 11 08:13:37 crc kubenswrapper[4860]: I1211 08:13:37.496372 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cpflv" event={"ID":"df36e993-8a70-42c5-ac22-508485572d35","Type":"ContainerStarted","Data":"a07760e65a3679f599740dd48c7ce8ea2ea5e68e04399dbb4b38d97cf499e2ae"} Dec 11 08:13:37 crc kubenswrapper[4860]: I1211 08:13:37.503311 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-rhl5v" event={"ID":"d7d1536c-ed04-4199-b839-b94e9482a63c","Type":"ContainerStarted","Data":"212f84085c22f84a02e17d9434646d1d104baaecfabef26acde4f527d50cfcdf"} Dec 11 08:13:37 crc kubenswrapper[4860]: I1211 08:13:37.515212 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-n8wrd" event={"ID":"30bec2ef-4635-4766-a34c-423c2e7f471b","Type":"ContainerStarted","Data":"d690f220b69c2bb9800011b361275a25463468bacebc17703435cc95f1566692"} Dec 11 08:13:37 crc kubenswrapper[4860]: I1211 08:13:37.519230 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-jxft9" event={"ID":"e86ba495-65e4-4539-928e-366376df64d8","Type":"ContainerStarted","Data":"6d3d36b248589249a3ffa5f1e1cfa8fa5b8d57f8b6a1435c9052717801c68439"} Dec 11 08:13:37 crc kubenswrapper[4860]: I1211 08:13:37.521746 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:13:37 crc kubenswrapper[4860]: E1211 08:13:37.522005 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:13:38.021992364 +0000 UTC m=+150.750511419 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:13:37 crc kubenswrapper[4860]: I1211 08:13:37.524515 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lbrj7" event={"ID":"9da229ee-dedc-417f-aec7-d2d1d9407a41","Type":"ContainerStarted","Data":"ea4061ade1ed5fa35cbda8a945254ed8a3cf135fca480352af30396bf79f7f51"} Dec 11 08:13:37 crc kubenswrapper[4860]: I1211 08:13:37.524558 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lbrj7" event={"ID":"9da229ee-dedc-417f-aec7-d2d1d9407a41","Type":"ContainerStarted","Data":"61efca1944542bec2fbec35c549c1be1eff4bdd48eaac61ad6ea969706044ea5"} Dec 11 08:13:37 crc kubenswrapper[4860]: I1211 08:13:37.525882 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-nc65d" event={"ID":"cb55e751-9913-4a3c-a0b4-3e681129d052","Type":"ContainerStarted","Data":"94a9e61e83506d92e1738b4ebc4dd00ef387cbf798aa6f669e15dddb79b6276e"} Dec 11 08:13:37 crc kubenswrapper[4860]: I1211 08:13:37.525906 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-nc65d" event={"ID":"cb55e751-9913-4a3c-a0b4-3e681129d052","Type":"ContainerStarted","Data":"67af2d2046559bef0096c8f517dd0e7d4b4e877bb994217e903dde2fea1a87ec"} Dec 11 08:13:37 crc kubenswrapper[4860]: I1211 08:13:37.531997 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-b4q7q" event={"ID":"fb3441e4-e612-4cfe-9f12-6497cd9176a9","Type":"ContainerStarted","Data":"c2fbe9afd42f45598b48a912bb5c10310190bcfbc870710797635e66387522ff"} Dec 11 08:13:37 crc kubenswrapper[4860]: I1211 08:13:37.532030 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-b4q7q" event={"ID":"fb3441e4-e612-4cfe-9f12-6497cd9176a9","Type":"ContainerStarted","Data":"fcc2d6f807afcc65fb61186585f3f1af97fdc1929c1c7038d4b2d9840463c580"} Dec 11 08:13:37 crc kubenswrapper[4860]: I1211 08:13:37.533182 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-b4q7q" Dec 11 08:13:37 crc kubenswrapper[4860]: I1211 08:13:37.547784 4860 patch_prober.go:28] interesting pod/console-operator-58897d9998-b4q7q container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.17:8443/readyz\": dial tcp 10.217.0.17:8443: connect: connection refused" start-of-body= Dec 11 08:13:37 crc kubenswrapper[4860]: I1211 08:13:37.547847 4860 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-b4q7q" podUID="fb3441e4-e612-4cfe-9f12-6497cd9176a9" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.17:8443/readyz\": dial tcp 10.217.0.17:8443: connect: connection refused" Dec 11 08:13:37 crc kubenswrapper[4860]: I1211 08:13:37.548476 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-j5kxr" event={"ID":"2c04dbaf-f385-4cee-bccf-8506ecdeb680","Type":"ContainerStarted","Data":"e38e705073aeee97ce7accca1796126ad72d28d488161d7dc4a4a94d73a31aec"} Dec 11 08:13:37 crc kubenswrapper[4860]: I1211 08:13:37.548501 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-j5kxr" event={"ID":"2c04dbaf-f385-4cee-bccf-8506ecdeb680","Type":"ContainerStarted","Data":"1e05d40b3ed9ca37d9ffedd99849bb0154b0c2040f430a2626f44149a21b1742"} Dec 11 08:13:37 crc kubenswrapper[4860]: I1211 08:13:37.549907 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q6ck8" event={"ID":"adb430a2-20a7-4e9d-8f92-446b56e36e4f","Type":"ContainerStarted","Data":"198e6dd61a9c2d3069f85232e914b4c45f8af7475867ace7ca2869bc6088c960"} Dec 11 08:13:37 crc kubenswrapper[4860]: I1211 08:13:37.553471 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wrqbs" event={"ID":"e963dd44-250d-4467-980c-1b67ef1bdd93","Type":"ContainerStarted","Data":"cefd02edbaeb8302f91bd4b4c5156109e47199c8924b06eb6291e8aad094fd74"} Dec 11 08:13:37 crc kubenswrapper[4860]: I1211 08:13:37.553523 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wrqbs" event={"ID":"e963dd44-250d-4467-980c-1b67ef1bdd93","Type":"ContainerStarted","Data":"a2f5e06a49e8c56932985b714e97071b763b508bceb0fe9f33f286c8caeec901"} Dec 11 08:13:37 crc kubenswrapper[4860]: I1211 08:13:37.558018 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-dz7kr" event={"ID":"ccbceeb2-58ab-4409-82df-5c99189569b8","Type":"ContainerStarted","Data":"da81c95fa79e88485cc7ff78c8483d365da07d5e70bf98b6b224c2b59dea5747"} Dec 11 08:13:37 crc kubenswrapper[4860]: I1211 08:13:37.558065 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-dz7kr" event={"ID":"ccbceeb2-58ab-4409-82df-5c99189569b8","Type":"ContainerStarted","Data":"905c393d94be1cc4f1936a9844ee4416c64fad01ddc896a25f003c6e8777386b"} Dec 11 08:13:37 crc kubenswrapper[4860]: I1211 08:13:37.558232 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-dz7kr" Dec 11 08:13:37 crc kubenswrapper[4860]: I1211 08:13:37.563510 4860 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-dz7kr container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.21:6443/healthz\": dial tcp 10.217.0.21:6443: connect: connection refused" start-of-body= Dec 11 08:13:37 crc kubenswrapper[4860]: I1211 08:13:37.563558 4860 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-dz7kr" podUID="ccbceeb2-58ab-4409-82df-5c99189569b8" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.21:6443/healthz\": dial tcp 10.217.0.21:6443: connect: connection refused" Dec 11 08:13:37 crc kubenswrapper[4860]: I1211 08:13:37.563611 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gkg77" event={"ID":"c094602b-e1ec-4b9a-9640-cd6250a1fd94","Type":"ContainerStarted","Data":"0198be15e840d16765e529017b2ee87e0754c3e17ae00cd9b8cbdf2f49e36b7a"} Dec 11 08:13:37 crc kubenswrapper[4860]: I1211 08:13:37.563666 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gkg77" event={"ID":"c094602b-e1ec-4b9a-9640-cd6250a1fd94","Type":"ContainerStarted","Data":"8814ca3149242804426a048c45b6dbd5331dfca4f041533e848e7cebb9cf055e"} Dec 11 08:13:37 crc kubenswrapper[4860]: I1211 08:13:37.582458 4860 generic.go:334] "Generic (PLEG): container finished" podID="e84a804f-75e1-46a2-b37a-f868de516379" containerID="976b0d49bb1ffd80ce2033f3a3f58f15303bbb4376dc694d6c777648155bcd9a" exitCode=0 Dec 11 08:13:37 crc kubenswrapper[4860]: I1211 08:13:37.627302 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zf572\" (UID: \"38005ee2-6635-498c-b156-ac57b4f2f4d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-zf572" Dec 11 08:13:37 crc kubenswrapper[4860]: E1211 08:13:37.628598 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:13:38.12858639 +0000 UTC m=+150.857105445 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zf572" (UID: "38005ee2-6635-498c-b156-ac57b4f2f4d9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:13:37 crc kubenswrapper[4860]: I1211 08:13:37.663662 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-24dmg" Dec 11 08:13:37 crc kubenswrapper[4860]: I1211 08:13:37.663699 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rvtb7" event={"ID":"e84a804f-75e1-46a2-b37a-f868de516379","Type":"ContainerDied","Data":"976b0d49bb1ffd80ce2033f3a3f58f15303bbb4376dc694d6c777648155bcd9a"} Dec 11 08:13:37 crc kubenswrapper[4860]: I1211 08:13:37.663746 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-z5lg5" event={"ID":"e55e20fb-bef2-4fee-b6a7-60e2d9d61faa","Type":"ContainerStarted","Data":"048da4e3643018d33fe4a43924884f3c4afdd66d8b7e7f58e874b4f989ce1cc1"} Dec 11 08:13:37 crc kubenswrapper[4860]: I1211 08:13:37.663757 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-z5lg5" event={"ID":"e55e20fb-bef2-4fee-b6a7-60e2d9d61faa","Type":"ContainerStarted","Data":"b12763f3c0a6a7129f349eb62605bfb1740fa50f6c57964d0b1dacf27d134dc4"} Dec 11 08:13:37 crc kubenswrapper[4860]: I1211 08:13:37.663766 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8x4xt" event={"ID":"09073ec8-36df-4ffb-a43a-10611999d8cb","Type":"ContainerStarted","Data":"682ee1be2005484337649ddf0508af9f5cf7b28b661914ac34e8073dc0e145ae"} Dec 11 08:13:37 crc kubenswrapper[4860]: I1211 08:13:37.663776 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8x4xt" event={"ID":"09073ec8-36df-4ffb-a43a-10611999d8cb","Type":"ContainerStarted","Data":"950466ad812f36761fe49ab92d1febcdb89c6daf0632b22763fd519793f6159e"} Dec 11 08:13:37 crc kubenswrapper[4860]: I1211 08:13:37.729653 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:13:37 crc kubenswrapper[4860]: E1211 08:13:37.731043 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:13:38.231024093 +0000 UTC m=+150.959543158 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:13:37 crc kubenswrapper[4860]: I1211 08:13:37.777073 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-mj2tr"] Dec 11 08:13:37 crc kubenswrapper[4860]: I1211 08:13:37.805974 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-nc65d" podStartSLOduration=129.805951682 podStartE2EDuration="2m9.805951682s" podCreationTimestamp="2025-12-11 08:11:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:13:37.799479521 +0000 UTC m=+150.527998586" watchObservedRunningTime="2025-12-11 08:13:37.805951682 +0000 UTC m=+150.534470737" Dec 11 08:13:37 crc kubenswrapper[4860]: I1211 08:13:37.851549 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zf572\" (UID: \"38005ee2-6635-498c-b156-ac57b4f2f4d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-zf572" Dec 11 08:13:37 crc kubenswrapper[4860]: E1211 08:13:37.851859 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:13:38.351847847 +0000 UTC m=+151.080366902 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zf572" (UID: "38005ee2-6635-498c-b156-ac57b4f2f4d9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:13:37 crc kubenswrapper[4860]: I1211 08:13:37.912879 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-rhl5v" podStartSLOduration=129.912863179 podStartE2EDuration="2m9.912863179s" podCreationTimestamp="2025-12-11 08:11:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:13:37.911449991 +0000 UTC m=+150.639969046" watchObservedRunningTime="2025-12-11 08:13:37.912863179 +0000 UTC m=+150.641382224" Dec 11 08:13:37 crc kubenswrapper[4860]: I1211 08:13:37.954952 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:13:37 crc kubenswrapper[4860]: E1211 08:13:37.957720 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:13:38.457695837 +0000 UTC m=+151.186214892 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:13:37 crc kubenswrapper[4860]: I1211 08:13:37.958448 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zf572\" (UID: \"38005ee2-6635-498c-b156-ac57b4f2f4d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-zf572" Dec 11 08:13:37 crc kubenswrapper[4860]: E1211 08:13:37.958878 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:13:38.458866287 +0000 UTC m=+151.187385342 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zf572" (UID: "38005ee2-6635-498c-b156-ac57b4f2f4d9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:13:37 crc kubenswrapper[4860]: W1211 08:13:37.969212 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcf1bd6cb_c047_423c_9d36_7246a6108844.slice/crio-1c7807f1d0f5456cfacb210570ac3683fe3d4c3f9176a1041db83402354af3e7 WatchSource:0}: Error finding container 1c7807f1d0f5456cfacb210570ac3683fe3d4c3f9176a1041db83402354af3e7: Status 404 returned error can't find the container with id 1c7807f1d0f5456cfacb210570ac3683fe3d4c3f9176a1041db83402354af3e7 Dec 11 08:13:38 crc kubenswrapper[4860]: I1211 08:13:38.053744 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-2z7ml"] Dec 11 08:13:38 crc kubenswrapper[4860]: I1211 08:13:38.059172 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:13:38 crc kubenswrapper[4860]: E1211 08:13:38.059577 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:13:38.559562981 +0000 UTC m=+151.288082036 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:13:38 crc kubenswrapper[4860]: I1211 08:13:38.072108 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-b4q7q" podStartSLOduration=130.07209227 podStartE2EDuration="2m10.07209227s" podCreationTimestamp="2025-12-11 08:11:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:13:38.068018561 +0000 UTC m=+150.796537616" watchObservedRunningTime="2025-12-11 08:13:38.07209227 +0000 UTC m=+150.800611325" Dec 11 08:13:38 crc kubenswrapper[4860]: I1211 08:13:38.161171 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zf572\" (UID: \"38005ee2-6635-498c-b156-ac57b4f2f4d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-zf572" Dec 11 08:13:38 crc kubenswrapper[4860]: E1211 08:13:38.161966 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:13:38.661953792 +0000 UTC m=+151.390472847 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zf572" (UID: "38005ee2-6635-498c-b156-ac57b4f2f4d9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:13:38 crc kubenswrapper[4860]: I1211 08:13:38.171499 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fsfs7"] Dec 11 08:13:38 crc kubenswrapper[4860]: I1211 08:13:38.195175 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5zqnk"] Dec 11 08:13:38 crc kubenswrapper[4860]: I1211 08:13:38.215486 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-ttrm2"] Dec 11 08:13:38 crc kubenswrapper[4860]: I1211 08:13:38.262948 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:13:38 crc kubenswrapper[4860]: E1211 08:13:38.263311 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:13:38.763296368 +0000 UTC m=+151.491815413 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:13:38 crc kubenswrapper[4860]: I1211 08:13:38.307743 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-24dmg" podStartSLOduration=130.307727282 podStartE2EDuration="2m10.307727282s" podCreationTimestamp="2025-12-11 08:11:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:13:38.30593128 +0000 UTC m=+151.034450335" watchObservedRunningTime="2025-12-11 08:13:38.307727282 +0000 UTC m=+151.036246337" Dec 11 08:13:38 crc kubenswrapper[4860]: I1211 08:13:38.364958 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zf572\" (UID: \"38005ee2-6635-498c-b156-ac57b4f2f4d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-zf572" Dec 11 08:13:38 crc kubenswrapper[4860]: E1211 08:13:38.365367 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:13:38.865357018 +0000 UTC m=+151.593876073 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zf572" (UID: "38005ee2-6635-498c-b156-ac57b4f2f4d9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:13:38 crc kubenswrapper[4860]: I1211 08:13:38.450376 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-svdpj"] Dec 11 08:13:38 crc kubenswrapper[4860]: I1211 08:13:38.472983 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-s6qsz"] Dec 11 08:13:38 crc kubenswrapper[4860]: I1211 08:13:38.474000 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:13:38 crc kubenswrapper[4860]: E1211 08:13:38.474864 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:13:38.974846943 +0000 UTC m=+151.703365998 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:13:38 crc kubenswrapper[4860]: I1211 08:13:38.510109 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-l4m2j"] Dec 11 08:13:38 crc kubenswrapper[4860]: I1211 08:13:38.578900 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zf572\" (UID: \"38005ee2-6635-498c-b156-ac57b4f2f4d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-zf572" Dec 11 08:13:38 crc kubenswrapper[4860]: E1211 08:13:38.579224 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:13:39.079212993 +0000 UTC m=+151.807732048 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zf572" (UID: "38005ee2-6635-498c-b156-ac57b4f2f4d9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:13:38 crc kubenswrapper[4860]: I1211 08:13:38.664760 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-ttrm2" event={"ID":"ad724960-18d1-4e83-944c-4cdeb033a436","Type":"ContainerStarted","Data":"0e4f02a9720f30652b572acc71b8074196ee665d1a1bc67339fc4accc42ddee2"} Dec 11 08:13:38 crc kubenswrapper[4860]: I1211 08:13:38.679050 4860 generic.go:334] "Generic (PLEG): container finished" podID="e86ba495-65e4-4539-928e-366376df64d8" containerID="f37c341962b62eed3ffe117845b654663d22f37e344accf9c3b2ccabcd3a9955" exitCode=0 Dec 11 08:13:38 crc kubenswrapper[4860]: I1211 08:13:38.679392 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-jxft9" event={"ID":"e86ba495-65e4-4539-928e-366376df64d8","Type":"ContainerDied","Data":"f37c341962b62eed3ffe117845b654663d22f37e344accf9c3b2ccabcd3a9955"} Dec 11 08:13:38 crc kubenswrapper[4860]: I1211 08:13:38.679609 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:13:38 crc kubenswrapper[4860]: E1211 08:13:38.679868 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:13:39.179855464 +0000 UTC m=+151.908374519 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:13:38 crc kubenswrapper[4860]: I1211 08:13:38.703149 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q6ck8" event={"ID":"adb430a2-20a7-4e9d-8f92-446b56e36e4f","Type":"ContainerStarted","Data":"5495672f4545d23753daece544d2d9f5e44a8fee4ba8948e0178abd1eb5071a7"} Dec 11 08:13:38 crc kubenswrapper[4860]: I1211 08:13:38.733885 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-9wb7z" event={"ID":"93550904-9295-4c4a-98bb-7a228c2699bb","Type":"ContainerStarted","Data":"f57578cd8b833099fbd7dd6c0c342c73440d1429c624d10133d791afd7c4f54c"} Dec 11 08:13:38 crc kubenswrapper[4860]: I1211 08:13:38.748036 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5zqnk" event={"ID":"3e5d90ab-884d-420b-a284-74f4b578a864","Type":"ContainerStarted","Data":"8b8ceeaac107722480d2aac743d415afb8092b93fe3e6720061ab731e093ff73"} Dec 11 08:13:38 crc kubenswrapper[4860]: I1211 08:13:38.752608 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-mj2tr" event={"ID":"cf1bd6cb-c047-423c-9d36-7246a6108844","Type":"ContainerStarted","Data":"2fdf416b7be3e639dc20d799cdaf06b01d2768eae4517e337803fa664c577fb8"} Dec 11 08:13:38 crc kubenswrapper[4860]: I1211 08:13:38.752661 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-mj2tr" event={"ID":"cf1bd6cb-c047-423c-9d36-7246a6108844","Type":"ContainerStarted","Data":"1c7807f1d0f5456cfacb210570ac3683fe3d4c3f9176a1041db83402354af3e7"} Dec 11 08:13:38 crc kubenswrapper[4860]: I1211 08:13:38.753227 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-mj2tr" Dec 11 08:13:38 crc kubenswrapper[4860]: I1211 08:13:38.754582 4860 patch_prober.go:28] interesting pod/downloads-7954f5f757-mj2tr container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.25:8080/\": dial tcp 10.217.0.25:8080: connect: connection refused" start-of-body= Dec 11 08:13:38 crc kubenswrapper[4860]: I1211 08:13:38.754623 4860 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-mj2tr" podUID="cf1bd6cb-c047-423c-9d36-7246a6108844" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.25:8080/\": dial tcp 10.217.0.25:8080: connect: connection refused" Dec 11 08:13:38 crc kubenswrapper[4860]: I1211 08:13:38.780866 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zf572\" (UID: \"38005ee2-6635-498c-b156-ac57b4f2f4d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-zf572" Dec 11 08:13:38 crc kubenswrapper[4860]: E1211 08:13:38.782117 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:13:39.282105902 +0000 UTC m=+152.010624957 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zf572" (UID: "38005ee2-6635-498c-b156-ac57b4f2f4d9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:13:38 crc kubenswrapper[4860]: I1211 08:13:38.788601 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rvtb7" event={"ID":"e84a804f-75e1-46a2-b37a-f868de516379","Type":"ContainerStarted","Data":"bb657a34b6b3221db5714123a4a3a6196d8b63a7f0a0844f630f5bcd0642b7cc"} Dec 11 08:13:38 crc kubenswrapper[4860]: I1211 08:13:38.788864 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rvtb7" Dec 11 08:13:38 crc kubenswrapper[4860]: I1211 08:13:38.794853 4860 patch_prober.go:28] interesting pod/machine-config-daemon-99qgp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 08:13:38 crc kubenswrapper[4860]: I1211 08:13:38.794895 4860 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 08:13:38 crc kubenswrapper[4860]: I1211 08:13:38.800516 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lbrj7" podStartSLOduration=130.800498712 podStartE2EDuration="2m10.800498712s" podCreationTimestamp="2025-12-11 08:11:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:13:38.799056593 +0000 UTC m=+151.527575648" watchObservedRunningTime="2025-12-11 08:13:38.800498712 +0000 UTC m=+151.529017757" Dec 11 08:13:38 crc kubenswrapper[4860]: I1211 08:13:38.802426 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-n8wrd" event={"ID":"30bec2ef-4635-4766-a34c-423c2e7f471b","Type":"ContainerStarted","Data":"87fb7db0686dff55283e8bfa65a1ef4ff7d00d0795624d771482fc22e6f2b0f2"} Dec 11 08:13:38 crc kubenswrapper[4860]: I1211 08:13:38.824211 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-g4bql" event={"ID":"1fff60d5-55fd-41d8-9225-251277e8e3c2","Type":"ContainerStarted","Data":"1cca3c922bcf74f99e6c5511fa3abea560bafd418f71af5e8f8b4c0ae66242f1"} Dec 11 08:13:38 crc kubenswrapper[4860]: I1211 08:13:38.840167 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6q4dc" event={"ID":"cefdf543-6d3a-4e62-9875-ca26b060f437","Type":"ContainerStarted","Data":"7e4eb6e666ac5a2a9c0cf4ad6599eb78ab3aa3af100ca8867e8e8b3dc157a540"} Dec 11 08:13:38 crc kubenswrapper[4860]: I1211 08:13:38.840229 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6q4dc" event={"ID":"cefdf543-6d3a-4e62-9875-ca26b060f437","Type":"ContainerStarted","Data":"002d5449c6b4632a4eff46d883949de2bb236fab54374e75abc726a488f2c28f"} Dec 11 08:13:38 crc kubenswrapper[4860]: I1211 08:13:38.844910 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-2z7ml" event={"ID":"64e3a9e4-d299-4349-b493-994630f747a5","Type":"ContainerStarted","Data":"d96bbb020bd81732eada71378d69d2c917fe6c64b36bf4c5cc7c87f49392cf5a"} Dec 11 08:13:38 crc kubenswrapper[4860]: I1211 08:13:38.860416 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fsfs7" event={"ID":"c78b27aa-097e-4308-ad5f-1d609b6b1b6c","Type":"ContainerStarted","Data":"0f78a69c5b6b939cf1ab7e1b99a2487c94f8c2adc9dd0ba52304c3f721bb4cd2"} Dec 11 08:13:38 crc kubenswrapper[4860]: I1211 08:13:38.882077 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:13:38 crc kubenswrapper[4860]: E1211 08:13:38.883210 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:13:39.383194359 +0000 UTC m=+152.111713414 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:13:38 crc kubenswrapper[4860]: I1211 08:13:38.983919 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zf572\" (UID: \"38005ee2-6635-498c-b156-ac57b4f2f4d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-zf572" Dec 11 08:13:38 crc kubenswrapper[4860]: E1211 08:13:38.984242 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:13:39.484225383 +0000 UTC m=+152.212744438 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zf572" (UID: "38005ee2-6635-498c-b156-ac57b4f2f4d9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:13:39 crc kubenswrapper[4860]: I1211 08:13:39.089747 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:13:39 crc kubenswrapper[4860]: E1211 08:13:39.090403 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:13:39.590383305 +0000 UTC m=+152.318902360 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:13:39 crc kubenswrapper[4860]: I1211 08:13:39.166865 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-dz7kr" podStartSLOduration=132.166848407 podStartE2EDuration="2m12.166848407s" podCreationTimestamp="2025-12-11 08:11:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:13:39.166733813 +0000 UTC m=+151.895252868" watchObservedRunningTime="2025-12-11 08:13:39.166848407 +0000 UTC m=+151.895367462" Dec 11 08:13:39 crc kubenswrapper[4860]: I1211 08:13:39.190808 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zf572\" (UID: \"38005ee2-6635-498c-b156-ac57b4f2f4d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-zf572" Dec 11 08:13:39 crc kubenswrapper[4860]: E1211 08:13:39.191130 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:13:39.69112019 +0000 UTC m=+152.419639245 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zf572" (UID: "38005ee2-6635-498c-b156-ac57b4f2f4d9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:13:39 crc kubenswrapper[4860]: I1211 08:13:39.199399 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-9wb7z" podStartSLOduration=5.199368773 podStartE2EDuration="5.199368773s" podCreationTimestamp="2025-12-11 08:13:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:13:39.196876656 +0000 UTC m=+151.925395711" watchObservedRunningTime="2025-12-11 08:13:39.199368773 +0000 UTC m=+151.927887838" Dec 11 08:13:39 crc kubenswrapper[4860]: I1211 08:13:39.234497 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-z5lg5" podStartSLOduration=131.234480636 podStartE2EDuration="2m11.234480636s" podCreationTimestamp="2025-12-11 08:11:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:13:39.231821635 +0000 UTC m=+151.960340690" watchObservedRunningTime="2025-12-11 08:13:39.234480636 +0000 UTC m=+151.962999691" Dec 11 08:13:39 crc kubenswrapper[4860]: I1211 08:13:39.265749 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-j5kxr" podStartSLOduration=132.265723498 podStartE2EDuration="2m12.265723498s" podCreationTimestamp="2025-12-11 08:11:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:13:39.264230206 +0000 UTC m=+151.992749271" watchObservedRunningTime="2025-12-11 08:13:39.265723498 +0000 UTC m=+151.994242553" Dec 11 08:13:39 crc kubenswrapper[4860]: I1211 08:13:39.291167 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-b4q7q" Dec 11 08:13:39 crc kubenswrapper[4860]: I1211 08:13:39.291789 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:13:39 crc kubenswrapper[4860]: E1211 08:13:39.292104 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:13:39.792089802 +0000 UTC m=+152.520608857 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:13:39 crc kubenswrapper[4860]: I1211 08:13:39.390783 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rvtb7" podStartSLOduration=131.390763256 podStartE2EDuration="2m11.390763256s" podCreationTimestamp="2025-12-11 08:11:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:13:39.389215624 +0000 UTC m=+152.117734679" watchObservedRunningTime="2025-12-11 08:13:39.390763256 +0000 UTC m=+152.119282321" Dec 11 08:13:39 crc kubenswrapper[4860]: I1211 08:13:39.393295 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zf572\" (UID: \"38005ee2-6635-498c-b156-ac57b4f2f4d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-zf572" Dec 11 08:13:39 crc kubenswrapper[4860]: E1211 08:13:39.393608 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:13:39.893594443 +0000 UTC m=+152.622113498 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zf572" (UID: "38005ee2-6635-498c-b156-ac57b4f2f4d9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:13:39 crc kubenswrapper[4860]: I1211 08:13:39.467616 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-mj2tr" podStartSLOduration=131.467583651 podStartE2EDuration="2m11.467583651s" podCreationTimestamp="2025-12-11 08:11:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:13:39.407104497 +0000 UTC m=+152.135623582" watchObservedRunningTime="2025-12-11 08:13:39.467583651 +0000 UTC m=+152.196102706" Dec 11 08:13:39 crc kubenswrapper[4860]: I1211 08:13:39.496745 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:13:39 crc kubenswrapper[4860]: E1211 08:13:39.497269 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:13:39.997249769 +0000 UTC m=+152.725768824 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:13:39 crc kubenswrapper[4860]: I1211 08:13:39.510977 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-n8wrd" podStartSLOduration=131.510961529 podStartE2EDuration="2m11.510961529s" podCreationTimestamp="2025-12-11 08:11:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:13:39.468953668 +0000 UTC m=+152.197472713" watchObservedRunningTime="2025-12-11 08:13:39.510961529 +0000 UTC m=+152.239480584" Dec 11 08:13:39 crc kubenswrapper[4860]: I1211 08:13:39.528515 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-kbz2w"] Dec 11 08:13:39 crc kubenswrapper[4860]: I1211 08:13:39.572254 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gkg77" podStartSLOduration=131.57223327 podStartE2EDuration="2m11.57223327s" podCreationTimestamp="2025-12-11 08:11:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:13:39.572202179 +0000 UTC m=+152.300721234" watchObservedRunningTime="2025-12-11 08:13:39.57223327 +0000 UTC m=+152.300752315" Dec 11 08:13:39 crc kubenswrapper[4860]: I1211 08:13:39.598954 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zf572\" (UID: \"38005ee2-6635-498c-b156-ac57b4f2f4d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-zf572" Dec 11 08:13:39 crc kubenswrapper[4860]: E1211 08:13:39.599426 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:13:40.099410963 +0000 UTC m=+152.827930018 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zf572" (UID: "38005ee2-6635-498c-b156-ac57b4f2f4d9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:13:39 crc kubenswrapper[4860]: I1211 08:13:39.613794 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q6ck8" podStartSLOduration=131.613777526 podStartE2EDuration="2m11.613777526s" podCreationTimestamp="2025-12-11 08:11:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:13:39.613212176 +0000 UTC m=+152.341731231" watchObservedRunningTime="2025-12-11 08:13:39.613777526 +0000 UTC m=+152.342296581" Dec 11 08:13:39 crc kubenswrapper[4860]: I1211 08:13:39.627429 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-dz7kr" Dec 11 08:13:39 crc kubenswrapper[4860]: I1211 08:13:39.649780 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8x4xt" podStartSLOduration=132.64975981 podStartE2EDuration="2m12.64975981s" podCreationTimestamp="2025-12-11 08:11:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:13:39.640531303 +0000 UTC m=+152.369050358" watchObservedRunningTime="2025-12-11 08:13:39.64975981 +0000 UTC m=+152.378278865" Dec 11 08:13:39 crc kubenswrapper[4860]: I1211 08:13:39.671913 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-n8wrd" Dec 11 08:13:39 crc kubenswrapper[4860]: I1211 08:13:39.676777 4860 patch_prober.go:28] interesting pod/router-default-5444994796-n8wrd container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 11 08:13:39 crc kubenswrapper[4860]: [-]has-synced failed: reason withheld Dec 11 08:13:39 crc kubenswrapper[4860]: [+]process-running ok Dec 11 08:13:39 crc kubenswrapper[4860]: healthz check failed Dec 11 08:13:39 crc kubenswrapper[4860]: I1211 08:13:39.676830 4860 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-n8wrd" podUID="30bec2ef-4635-4766-a34c-423c2e7f471b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 11 08:13:39 crc kubenswrapper[4860]: I1211 08:13:39.706587 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:13:39 crc kubenswrapper[4860]: E1211 08:13:39.706928 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:13:40.206909479 +0000 UTC m=+152.935428534 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:13:39 crc kubenswrapper[4860]: I1211 08:13:39.807569 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zf572\" (UID: \"38005ee2-6635-498c-b156-ac57b4f2f4d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-zf572" Dec 11 08:13:39 crc kubenswrapper[4860]: E1211 08:13:39.808182 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:13:40.308166922 +0000 UTC m=+153.036685987 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zf572" (UID: "38005ee2-6635-498c-b156-ac57b4f2f4d9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:13:39 crc kubenswrapper[4860]: I1211 08:13:39.816486 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-87b7k"] Dec 11 08:13:39 crc kubenswrapper[4860]: I1211 08:13:39.843156 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-x28lp"] Dec 11 08:13:39 crc kubenswrapper[4860]: I1211 08:13:39.878329 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-76lnq"] Dec 11 08:13:39 crc kubenswrapper[4860]: I1211 08:13:39.887608 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-svdpj" event={"ID":"03384260-7d59-42d5-ab3c-b411e9ed88d1","Type":"ContainerStarted","Data":"0b28ce7f3ab642b66d96ddd0cf069910a572669c9ec4d66e65f27d5607921241"} Dec 11 08:13:39 crc kubenswrapper[4860]: I1211 08:13:39.887705 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-svdpj" event={"ID":"03384260-7d59-42d5-ab3c-b411e9ed88d1","Type":"ContainerStarted","Data":"2b9c594c4c84c6e4970b4f1f172a8107092a99a204a5f6e7f117f2ff0f85f99a"} Dec 11 08:13:39 crc kubenswrapper[4860]: I1211 08:13:39.893280 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-2z7ml" event={"ID":"64e3a9e4-d299-4349-b493-994630f747a5","Type":"ContainerStarted","Data":"8ee5af5324130f75c077da34dab90239f85e63a07dcb8d257629f67d9e28b51a"} Dec 11 08:13:39 crc kubenswrapper[4860]: I1211 08:13:39.894360 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-rl2ff"] Dec 11 08:13:39 crc kubenswrapper[4860]: I1211 08:13:39.895489 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cpflv" event={"ID":"df36e993-8a70-42c5-ac22-508485572d35","Type":"ContainerStarted","Data":"5c936da96f0de6878af977e095e67c1327b714c3643cae74d10c70104c17f5b4"} Dec 11 08:13:39 crc kubenswrapper[4860]: I1211 08:13:39.908398 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:13:39 crc kubenswrapper[4860]: E1211 08:13:39.909214 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:13:40.409199567 +0000 UTC m=+153.137718622 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:13:39 crc kubenswrapper[4860]: I1211 08:13:39.910114 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-jgk9d"] Dec 11 08:13:39 crc kubenswrapper[4860]: I1211 08:13:39.928211 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-87b7k" event={"ID":"73ab8adf-cf47-4482-bbdb-8fb6c3edc87a","Type":"ContainerStarted","Data":"8f5aadd283eb92b3ccba02c57fb38a8038de74ed2fe0d3b892fdfe51df4e0a69"} Dec 11 08:13:39 crc kubenswrapper[4860]: I1211 08:13:39.940486 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cpflv" podStartSLOduration=131.9404686 podStartE2EDuration="2m11.9404686s" podCreationTimestamp="2025-12-11 08:11:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:13:39.937868121 +0000 UTC m=+152.666387176" watchObservedRunningTime="2025-12-11 08:13:39.9404686 +0000 UTC m=+152.668987655" Dec 11 08:13:39 crc kubenswrapper[4860]: W1211 08:13:39.949809 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7a764867_5216_47fb_8211_cea78b79f707.slice/crio-c254b6625b6e40d25177edf1f2d70cba2fd9651c986b6affc6aa039745e994e4 WatchSource:0}: Error finding container c254b6625b6e40d25177edf1f2d70cba2fd9651c986b6affc6aa039745e994e4: Status 404 returned error can't find the container with id c254b6625b6e40d25177edf1f2d70cba2fd9651c986b6affc6aa039745e994e4 Dec 11 08:13:39 crc kubenswrapper[4860]: I1211 08:13:39.973463 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-l4m2j" event={"ID":"5b8b374c-b131-402c-8e3b-02d93d416489","Type":"ContainerStarted","Data":"2e9df079a0ecdb5ed53d65b00e211c621c5bd9c1c46cdfed96644536df234a67"} Dec 11 08:13:39 crc kubenswrapper[4860]: I1211 08:13:39.973518 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-l4m2j" event={"ID":"5b8b374c-b131-402c-8e3b-02d93d416489","Type":"ContainerStarted","Data":"ca3ff95ab312499f43c7054b5889cc637d6347e98dab3ce92e594f1b88165d8f"} Dec 11 08:13:39 crc kubenswrapper[4860]: I1211 08:13:39.988537 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fsfs7" Dec 11 08:13:40 crc kubenswrapper[4860]: I1211 08:13:39.995308 4860 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-fsfs7 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.35:5443/healthz\": dial tcp 10.217.0.35:5443: connect: connection refused" start-of-body= Dec 11 08:13:40 crc kubenswrapper[4860]: I1211 08:13:39.995573 4860 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fsfs7" podUID="c78b27aa-097e-4308-ad5f-1d609b6b1b6c" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.35:5443/healthz\": dial tcp 10.217.0.35:5443: connect: connection refused" Dec 11 08:13:40 crc kubenswrapper[4860]: I1211 08:13:40.009049 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29424000-nf7zl"] Dec 11 08:13:40 crc kubenswrapper[4860]: I1211 08:13:40.010018 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zf572\" (UID: \"38005ee2-6635-498c-b156-ac57b4f2f4d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-zf572" Dec 11 08:13:40 crc kubenswrapper[4860]: E1211 08:13:40.010290 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:13:40.510280264 +0000 UTC m=+153.238799309 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zf572" (UID: "38005ee2-6635-498c-b156-ac57b4f2f4d9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:13:40 crc kubenswrapper[4860]: I1211 08:13:40.017449 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-h7zxk"] Dec 11 08:13:40 crc kubenswrapper[4860]: I1211 08:13:40.019772 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-s6qsz" event={"ID":"c4c320f7-3b40-4edd-8e25-40b477e1d374","Type":"ContainerStarted","Data":"3c474c734899fc5f57a6ab828ac56e4d7aee8d765d4ea6af14f810e4fa6d2bc7"} Dec 11 08:13:40 crc kubenswrapper[4860]: I1211 08:13:40.020452 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-s6qsz" Dec 11 08:13:40 crc kubenswrapper[4860]: I1211 08:13:40.023555 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vc92z"] Dec 11 08:13:40 crc kubenswrapper[4860]: I1211 08:13:40.028819 4860 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-s6qsz container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.40:8443/healthz\": dial tcp 10.217.0.40:8443: connect: connection refused" start-of-body= Dec 11 08:13:40 crc kubenswrapper[4860]: I1211 08:13:40.028890 4860 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-s6qsz" podUID="c4c320f7-3b40-4edd-8e25-40b477e1d374" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.40:8443/healthz\": dial tcp 10.217.0.40:8443: connect: connection refused" Dec 11 08:13:40 crc kubenswrapper[4860]: I1211 08:13:40.037464 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fsfs7" podStartSLOduration=132.037439985 podStartE2EDuration="2m12.037439985s" podCreationTimestamp="2025-12-11 08:11:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:13:40.023131455 +0000 UTC m=+152.751650510" watchObservedRunningTime="2025-12-11 08:13:40.037439985 +0000 UTC m=+152.765959030" Dec 11 08:13:40 crc kubenswrapper[4860]: I1211 08:13:40.039281 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6q4dc" event={"ID":"cefdf543-6d3a-4e62-9875-ca26b060f437","Type":"ContainerStarted","Data":"c18e3ef18b50ad0c4a43181b9a9c166bf90bea1aaa41ad59eec837b23d8f22fa"} Dec 11 08:13:40 crc kubenswrapper[4860]: I1211 08:13:40.041298 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-kbz2w" event={"ID":"eceb74a4-c695-46e9-ba5f-b5145d20d968","Type":"ContainerStarted","Data":"b6948186890a7a373157ab3118bed892b16de49272ecf74d4a3d50c92622b515"} Dec 11 08:13:40 crc kubenswrapper[4860]: I1211 08:13:40.050347 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-s6qsz" podStartSLOduration=132.050328968 podStartE2EDuration="2m12.050328968s" podCreationTimestamp="2025-12-11 08:11:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:13:40.049560311 +0000 UTC m=+152.778079366" watchObservedRunningTime="2025-12-11 08:13:40.050328968 +0000 UTC m=+152.778848043" Dec 11 08:13:40 crc kubenswrapper[4860]: I1211 08:13:40.059574 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5zqnk" event={"ID":"3e5d90ab-884d-420b-a284-74f4b578a864","Type":"ContainerStarted","Data":"9cd74d8e1b6f6938761c9fbc0486ef86443c9b043145a14d0f5df81fecfd00bd"} Dec 11 08:13:40 crc kubenswrapper[4860]: I1211 08:13:40.077519 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9tkth"] Dec 11 08:13:40 crc kubenswrapper[4860]: I1211 08:13:40.077575 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-6f4m5"] Dec 11 08:13:40 crc kubenswrapper[4860]: I1211 08:13:40.085777 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-qx9pf"] Dec 11 08:13:40 crc kubenswrapper[4860]: W1211 08:13:40.089886 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb08141e2_41b8_45f4_bb40_c319621719c8.slice/crio-b559d0c25efb0429df1882fb7a408224ba7a61f1c5d6a87ba8b7486691aba246 WatchSource:0}: Error finding container b559d0c25efb0429df1882fb7a408224ba7a61f1c5d6a87ba8b7486691aba246: Status 404 returned error can't find the container with id b559d0c25efb0429df1882fb7a408224ba7a61f1c5d6a87ba8b7486691aba246 Dec 11 08:13:40 crc kubenswrapper[4860]: I1211 08:13:40.108957 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6q4dc" podStartSLOduration=132.108940758 podStartE2EDuration="2m12.108940758s" podCreationTimestamp="2025-12-11 08:11:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:13:40.097739784 +0000 UTC m=+152.826258859" watchObservedRunningTime="2025-12-11 08:13:40.108940758 +0000 UTC m=+152.837459813" Dec 11 08:13:40 crc kubenswrapper[4860]: I1211 08:13:40.111278 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-t8ffw"] Dec 11 08:13:40 crc kubenswrapper[4860]: I1211 08:13:40.111811 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:13:40 crc kubenswrapper[4860]: E1211 08:13:40.111888 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:13:40.611867889 +0000 UTC m=+153.340386944 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:13:40 crc kubenswrapper[4860]: I1211 08:13:40.112023 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zf572\" (UID: \"38005ee2-6635-498c-b156-ac57b4f2f4d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-zf572" Dec 11 08:13:40 crc kubenswrapper[4860]: E1211 08:13:40.112943 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:13:40.612931025 +0000 UTC m=+153.341450080 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zf572" (UID: "38005ee2-6635-498c-b156-ac57b4f2f4d9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:13:40 crc kubenswrapper[4860]: I1211 08:13:40.120318 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wrqbs" event={"ID":"e963dd44-250d-4467-980c-1b67ef1bdd93","Type":"ContainerStarted","Data":"d28da0c23f17aa3abffb1d39d2fe709de69fdabe2ed2818aef1b1a327dba9257"} Dec 11 08:13:40 crc kubenswrapper[4860]: I1211 08:13:40.130345 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-sbcxd"] Dec 11 08:13:40 crc kubenswrapper[4860]: I1211 08:13:40.147607 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wrqbs" podStartSLOduration=133.147588084 podStartE2EDuration="2m13.147588084s" podCreationTimestamp="2025-12-11 08:11:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:13:40.146725744 +0000 UTC m=+152.875244809" watchObservedRunningTime="2025-12-11 08:13:40.147588084 +0000 UTC m=+152.876107139" Dec 11 08:13:40 crc kubenswrapper[4860]: I1211 08:13:40.189065 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-9wb7z" event={"ID":"93550904-9295-4c4a-98bb-7a228c2699bb","Type":"ContainerStarted","Data":"4c72aa82d5bc436a5ea202707619766eb460acb7100763b319987fb78b2dee3f"} Dec 11 08:13:40 crc kubenswrapper[4860]: I1211 08:13:40.213441 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:13:40 crc kubenswrapper[4860]: E1211 08:13:40.214812 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:13:40.714795029 +0000 UTC m=+153.443314094 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:13:40 crc kubenswrapper[4860]: I1211 08:13:40.229655 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-ttrm2" event={"ID":"ad724960-18d1-4e83-944c-4cdeb033a436","Type":"ContainerStarted","Data":"cafda173280b05a3de720d8e7f8927fc27b4c53b7500ba9deb3878e2112d3d8c"} Dec 11 08:13:40 crc kubenswrapper[4860]: I1211 08:13:40.231134 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-ttrm2" Dec 11 08:13:40 crc kubenswrapper[4860]: I1211 08:13:40.231388 4860 patch_prober.go:28] interesting pod/downloads-7954f5f757-mj2tr container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.25:8080/\": dial tcp 10.217.0.25:8080: connect: connection refused" start-of-body= Dec 11 08:13:40 crc kubenswrapper[4860]: I1211 08:13:40.231429 4860 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-mj2tr" podUID="cf1bd6cb-c047-423c-9d36-7246a6108844" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.25:8080/\": dial tcp 10.217.0.25:8080: connect: connection refused" Dec 11 08:13:40 crc kubenswrapper[4860]: I1211 08:13:40.257014 4860 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-ttrm2 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.38:8080/healthz\": dial tcp 10.217.0.38:8080: connect: connection refused" start-of-body= Dec 11 08:13:40 crc kubenswrapper[4860]: I1211 08:13:40.257353 4860 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-ttrm2" podUID="ad724960-18d1-4e83-944c-4cdeb033a436" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.38:8080/healthz\": dial tcp 10.217.0.38:8080: connect: connection refused" Dec 11 08:13:40 crc kubenswrapper[4860]: I1211 08:13:40.321059 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zf572\" (UID: \"38005ee2-6635-498c-b156-ac57b4f2f4d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-zf572" Dec 11 08:13:40 crc kubenswrapper[4860]: E1211 08:13:40.323736 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:13:40.823722624 +0000 UTC m=+153.552241679 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zf572" (UID: "38005ee2-6635-498c-b156-ac57b4f2f4d9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:13:40 crc kubenswrapper[4860]: I1211 08:13:40.423858 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:13:40 crc kubenswrapper[4860]: E1211 08:13:40.424053 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:13:40.924019484 +0000 UTC m=+153.652538539 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:13:40 crc kubenswrapper[4860]: I1211 08:13:40.424164 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zf572\" (UID: \"38005ee2-6635-498c-b156-ac57b4f2f4d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-zf572" Dec 11 08:13:40 crc kubenswrapper[4860]: E1211 08:13:40.424947 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:13:40.924932276 +0000 UTC m=+153.653451331 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zf572" (UID: "38005ee2-6635-498c-b156-ac57b4f2f4d9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:13:40 crc kubenswrapper[4860]: I1211 08:13:40.524809 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:13:40 crc kubenswrapper[4860]: E1211 08:13:40.524954 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:13:41.024925745 +0000 UTC m=+153.753444800 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:13:40 crc kubenswrapper[4860]: I1211 08:13:40.525134 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zf572\" (UID: \"38005ee2-6635-498c-b156-ac57b4f2f4d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-zf572" Dec 11 08:13:40 crc kubenswrapper[4860]: E1211 08:13:40.525493 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:13:41.025484034 +0000 UTC m=+153.754003089 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zf572" (UID: "38005ee2-6635-498c-b156-ac57b4f2f4d9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:13:40 crc kubenswrapper[4860]: I1211 08:13:40.627266 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:13:40 crc kubenswrapper[4860]: E1211 08:13:40.627888 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:13:41.127868955 +0000 UTC m=+153.856388010 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:13:40 crc kubenswrapper[4860]: I1211 08:13:40.694989 4860 patch_prober.go:28] interesting pod/router-default-5444994796-n8wrd container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 11 08:13:40 crc kubenswrapper[4860]: [-]has-synced failed: reason withheld Dec 11 08:13:40 crc kubenswrapper[4860]: [+]process-running ok Dec 11 08:13:40 crc kubenswrapper[4860]: healthz check failed Dec 11 08:13:40 crc kubenswrapper[4860]: I1211 08:13:40.695037 4860 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-n8wrd" podUID="30bec2ef-4635-4766-a34c-423c2e7f471b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 11 08:13:40 crc kubenswrapper[4860]: I1211 08:13:40.729309 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zf572\" (UID: \"38005ee2-6635-498c-b156-ac57b4f2f4d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-zf572" Dec 11 08:13:40 crc kubenswrapper[4860]: E1211 08:13:40.729685 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:13:41.229669347 +0000 UTC m=+153.958188412 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zf572" (UID: "38005ee2-6635-498c-b156-ac57b4f2f4d9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:13:40 crc kubenswrapper[4860]: I1211 08:13:40.831203 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:13:40 crc kubenswrapper[4860]: E1211 08:13:40.831597 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:13:41.331582273 +0000 UTC m=+154.060101328 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:13:40 crc kubenswrapper[4860]: I1211 08:13:40.932043 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zf572\" (UID: \"38005ee2-6635-498c-b156-ac57b4f2f4d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-zf572" Dec 11 08:13:40 crc kubenswrapper[4860]: E1211 08:13:40.932415 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:13:41.432398511 +0000 UTC m=+154.160917566 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zf572" (UID: "38005ee2-6635-498c-b156-ac57b4f2f4d9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:13:41 crc kubenswrapper[4860]: I1211 08:13:41.032682 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:13:41 crc kubenswrapper[4860]: E1211 08:13:41.033100 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:13:41.533079744 +0000 UTC m=+154.261598799 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:13:41 crc kubenswrapper[4860]: I1211 08:13:41.059726 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cpflv" Dec 11 08:13:41 crc kubenswrapper[4860]: I1211 08:13:41.059905 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cpflv" Dec 11 08:13:41 crc kubenswrapper[4860]: I1211 08:13:41.134628 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zf572\" (UID: \"38005ee2-6635-498c-b156-ac57b4f2f4d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-zf572" Dec 11 08:13:41 crc kubenswrapper[4860]: E1211 08:13:41.135877 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:13:41.635864999 +0000 UTC m=+154.364384054 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zf572" (UID: "38005ee2-6635-498c-b156-ac57b4f2f4d9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:13:41 crc kubenswrapper[4860]: I1211 08:13:41.181562 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cpflv" Dec 11 08:13:41 crc kubenswrapper[4860]: I1211 08:13:41.213264 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-ttrm2" podStartSLOduration=133.213246663 podStartE2EDuration="2m13.213246663s" podCreationTimestamp="2025-12-11 08:11:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:13:40.253147074 +0000 UTC m=+152.981666129" watchObservedRunningTime="2025-12-11 08:13:41.213246663 +0000 UTC m=+153.941765708" Dec 11 08:13:41 crc kubenswrapper[4860]: I1211 08:13:41.236560 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:13:41 crc kubenswrapper[4860]: E1211 08:13:41.236695 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:13:41.736629515 +0000 UTC m=+154.465148570 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:13:41 crc kubenswrapper[4860]: I1211 08:13:41.236802 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zf572\" (UID: \"38005ee2-6635-498c-b156-ac57b4f2f4d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-zf572" Dec 11 08:13:41 crc kubenswrapper[4860]: E1211 08:13:41.237087 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:13:41.73708016 +0000 UTC m=+154.465599215 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zf572" (UID: "38005ee2-6635-498c-b156-ac57b4f2f4d9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:13:41 crc kubenswrapper[4860]: I1211 08:13:41.307168 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-jxft9" event={"ID":"e86ba495-65e4-4539-928e-366376df64d8","Type":"ContainerStarted","Data":"ae4c6266b76ec0993e80f8ac24be300a5e4a36156e46145c7946688205ea0dad"} Dec 11 08:13:41 crc kubenswrapper[4860]: I1211 08:13:41.307202 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-jxft9" event={"ID":"e86ba495-65e4-4539-928e-366376df64d8","Type":"ContainerStarted","Data":"aeb0804d73bbaa58a94986af8f045e9f0a7142fcea9c1438107950da42aace11"} Dec 11 08:13:41 crc kubenswrapper[4860]: I1211 08:13:41.315290 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-2z7ml" event={"ID":"64e3a9e4-d299-4349-b493-994630f747a5","Type":"ContainerStarted","Data":"c57107fc89bf26353e26cfc7ad54afef67a5a31454cd3e43fd6347e83a8b2147"} Dec 11 08:13:41 crc kubenswrapper[4860]: I1211 08:13:41.337930 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:13:41 crc kubenswrapper[4860]: E1211 08:13:41.338709 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:13:41.838692724 +0000 UTC m=+154.567211779 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:13:41 crc kubenswrapper[4860]: I1211 08:13:41.346717 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-jxft9" podStartSLOduration=134.34669839 podStartE2EDuration="2m14.34669839s" podCreationTimestamp="2025-12-11 08:11:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:13:41.344123181 +0000 UTC m=+154.072642236" watchObservedRunningTime="2025-12-11 08:13:41.34669839 +0000 UTC m=+154.075217445" Dec 11 08:13:41 crc kubenswrapper[4860]: I1211 08:13:41.357090 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-qx9pf" event={"ID":"89e3648e-71d3-4d85-aef0-aa60bc630c8f","Type":"ContainerStarted","Data":"d989ab1d0a2c317f409aee572737bee551255edea01bb2a654849f53fd383e75"} Dec 11 08:13:41 crc kubenswrapper[4860]: I1211 08:13:41.357135 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-qx9pf" event={"ID":"89e3648e-71d3-4d85-aef0-aa60bc630c8f","Type":"ContainerStarted","Data":"544b0fac2c423dbf4a48d558e048bc052536d6cd458961d137524ac6fd7884fd"} Dec 11 08:13:41 crc kubenswrapper[4860]: I1211 08:13:41.384623 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-rl2ff" event={"ID":"eabe3a89-8c1c-41fa-a8e9-9a0afed00d4e","Type":"ContainerStarted","Data":"f2ceaffb127bffaed3d82b8622423c7c63964fabfcd48cf2edec73831d9cf43a"} Dec 11 08:13:41 crc kubenswrapper[4860]: I1211 08:13:41.384702 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-rl2ff" event={"ID":"eabe3a89-8c1c-41fa-a8e9-9a0afed00d4e","Type":"ContainerStarted","Data":"d64c12210dda6f81ec1fa8f060f76b1330a16b2b985bb1cb5dfee94e9a2d011c"} Dec 11 08:13:41 crc kubenswrapper[4860]: I1211 08:13:41.411562 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-2z7ml" podStartSLOduration=133.411540973 podStartE2EDuration="2m13.411540973s" podCreationTimestamp="2025-12-11 08:11:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:13:41.399427098 +0000 UTC m=+154.127946143" watchObservedRunningTime="2025-12-11 08:13:41.411540973 +0000 UTC m=+154.140060028" Dec 11 08:13:41 crc kubenswrapper[4860]: I1211 08:13:41.415214 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-g4bql" event={"ID":"1fff60d5-55fd-41d8-9225-251277e8e3c2","Type":"ContainerStarted","Data":"22d7901de36461a9b5aa7afef9f76bd3accb1a4de9f1123fa656ac996753333a"} Dec 11 08:13:41 crc kubenswrapper[4860]: I1211 08:13:41.415273 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-g4bql" event={"ID":"1fff60d5-55fd-41d8-9225-251277e8e3c2","Type":"ContainerStarted","Data":"ae179239315e9e0f34620fc7e0d7ccb4a4e7ee925d00deb80c8108a38dac14a4"} Dec 11 08:13:41 crc kubenswrapper[4860]: I1211 08:13:41.437741 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-qx9pf" podStartSLOduration=133.437719801 podStartE2EDuration="2m13.437719801s" podCreationTimestamp="2025-12-11 08:11:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:13:41.436374976 +0000 UTC m=+154.164894031" watchObservedRunningTime="2025-12-11 08:13:41.437719801 +0000 UTC m=+154.166238856" Dec 11 08:13:41 crc kubenswrapper[4860]: I1211 08:13:41.439454 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zf572\" (UID: \"38005ee2-6635-498c-b156-ac57b4f2f4d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-zf572" Dec 11 08:13:41 crc kubenswrapper[4860]: I1211 08:13:41.439672 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29424000-nf7zl" event={"ID":"0f6c4496-7751-49db-adc2-18099589c708","Type":"ContainerStarted","Data":"f5f79575a837d49cfff4df714dbab8b7364ce4d99d9446a09a0108cb3027e312"} Dec 11 08:13:41 crc kubenswrapper[4860]: I1211 08:13:41.439743 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29424000-nf7zl" event={"ID":"0f6c4496-7751-49db-adc2-18099589c708","Type":"ContainerStarted","Data":"f7fdb07c94b081ce80b3ef7933886a6efe4a77ca6ad75126daf037599561e4ed"} Dec 11 08:13:41 crc kubenswrapper[4860]: E1211 08:13:41.442926 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:13:41.942916029 +0000 UTC m=+154.671435084 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zf572" (UID: "38005ee2-6635-498c-b156-ac57b4f2f4d9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:13:41 crc kubenswrapper[4860]: I1211 08:13:41.457808 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-6f4m5" event={"ID":"f7673917-9864-467b-b1f8-af4b2bac9a48","Type":"ContainerStarted","Data":"ed7dc45bb9ec20e12e90d3796d34b74e2548b76c247cb7833a70f81ebe87b508"} Dec 11 08:13:41 crc kubenswrapper[4860]: I1211 08:13:41.457858 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-6f4m5" event={"ID":"f7673917-9864-467b-b1f8-af4b2bac9a48","Type":"ContainerStarted","Data":"10c2c7640ab3636959699d6e2c59b19919bbe7bc141549901a64336f0483f5fa"} Dec 11 08:13:41 crc kubenswrapper[4860]: I1211 08:13:41.473401 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-rl2ff" podStartSLOduration=133.473385464 podStartE2EDuration="2m13.473385464s" podCreationTimestamp="2025-12-11 08:11:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:13:41.472924258 +0000 UTC m=+154.201443313" watchObservedRunningTime="2025-12-11 08:13:41.473385464 +0000 UTC m=+154.201904519" Dec 11 08:13:41 crc kubenswrapper[4860]: I1211 08:13:41.483391 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-x28lp" event={"ID":"c3ce3c34-9b72-4964-b59a-f6b3c0a0e5c8","Type":"ContainerStarted","Data":"bb07110b41230d551fd6fac6fc0777001ac39c0e03a0b1edb7545bd0fa0f3ee8"} Dec 11 08:13:41 crc kubenswrapper[4860]: I1211 08:13:41.483441 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-x28lp" event={"ID":"c3ce3c34-9b72-4964-b59a-f6b3c0a0e5c8","Type":"ContainerStarted","Data":"1626fc10921886e4ed46782dac868f9c1345b8567320f1b1f38136c2724435e9"} Dec 11 08:13:41 crc kubenswrapper[4860]: I1211 08:13:41.492055 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-6f4m5" podStartSLOduration=133.492024844 podStartE2EDuration="2m13.492024844s" podCreationTimestamp="2025-12-11 08:11:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:13:41.491415413 +0000 UTC m=+154.219934478" watchObservedRunningTime="2025-12-11 08:13:41.492024844 +0000 UTC m=+154.220543899" Dec 11 08:13:41 crc kubenswrapper[4860]: I1211 08:13:41.505857 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5zqnk" event={"ID":"3e5d90ab-884d-420b-a284-74f4b578a864","Type":"ContainerStarted","Data":"a1e0c41aad6939ea3b0e5b1fb1508453f51f8b70ef73fb2543d243f0ec87df16"} Dec 11 08:13:41 crc kubenswrapper[4860]: I1211 08:13:41.506692 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5zqnk" Dec 11 08:13:41 crc kubenswrapper[4860]: I1211 08:13:41.510288 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vc92z" event={"ID":"776e5c46-e988-4bb7-98e5-2d10c5fb5142","Type":"ContainerStarted","Data":"ae0c42883793cdd143dc949b9f29af82e26e2a1b2b705d4d323050720d8ebd87"} Dec 11 08:13:41 crc kubenswrapper[4860]: I1211 08:13:41.510320 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vc92z" event={"ID":"776e5c46-e988-4bb7-98e5-2d10c5fb5142","Type":"ContainerStarted","Data":"98640cb61ced88bfc2e347b2a32b8663190e400d236330e1d2ded313a8ccbf6c"} Dec 11 08:13:41 crc kubenswrapper[4860]: I1211 08:13:41.513201 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9tkth" event={"ID":"67ae7457-ee23-4ae3-a118-fbfac6dd85bf","Type":"ContainerStarted","Data":"864026ee3eb22cba0d655e8fce6dc62e45fbad7f89eb19f96924ea921f82851b"} Dec 11 08:13:41 crc kubenswrapper[4860]: I1211 08:13:41.521172 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-g4bql" podStartSLOduration=133.521154313 podStartE2EDuration="2m13.521154313s" podCreationTimestamp="2025-12-11 08:11:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:13:41.515157977 +0000 UTC m=+154.243677032" watchObservedRunningTime="2025-12-11 08:13:41.521154313 +0000 UTC m=+154.249673368" Dec 11 08:13:41 crc kubenswrapper[4860]: I1211 08:13:41.527899 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fsfs7" event={"ID":"c78b27aa-097e-4308-ad5f-1d609b6b1b6c","Type":"ContainerStarted","Data":"271c15461a0b784107e9ae37933bb276005db3a6ddcbd4d61873fc58218583d1"} Dec 11 08:13:41 crc kubenswrapper[4860]: I1211 08:13:41.540713 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:13:41 crc kubenswrapper[4860]: E1211 08:13:41.542724 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:13:42.042666421 +0000 UTC m=+154.771185476 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:13:41 crc kubenswrapper[4860]: I1211 08:13:41.545523 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29424000-nf7zl" podStartSLOduration=134.545507689 podStartE2EDuration="2m14.545507689s" podCreationTimestamp="2025-12-11 08:11:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:13:41.544009247 +0000 UTC m=+154.272528312" watchObservedRunningTime="2025-12-11 08:13:41.545507689 +0000 UTC m=+154.274026744" Dec 11 08:13:41 crc kubenswrapper[4860]: I1211 08:13:41.549001 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-kbz2w" event={"ID":"eceb74a4-c695-46e9-ba5f-b5145d20d968","Type":"ContainerStarted","Data":"bc2cadf42829d35f81fbb46350e9f906bf2db6a96f45f7010c9ecffc073926fe"} Dec 11 08:13:41 crc kubenswrapper[4860]: I1211 08:13:41.551900 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-sbcxd" event={"ID":"220b6958-d8ee-4733-a93a-35d2ced9f150","Type":"ContainerStarted","Data":"cda5414d71c3105fce4591eb55b3e1cf73b6c51d49268790f7bf1d91fe6e87e5"} Dec 11 08:13:41 crc kubenswrapper[4860]: I1211 08:13:41.564479 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5zqnk" podStartSLOduration=133.564463028 podStartE2EDuration="2m13.564463028s" podCreationTimestamp="2025-12-11 08:11:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:13:41.563012578 +0000 UTC m=+154.291531623" watchObservedRunningTime="2025-12-11 08:13:41.564463028 +0000 UTC m=+154.292982073" Dec 11 08:13:41 crc kubenswrapper[4860]: I1211 08:13:41.586082 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vc92z" podStartSLOduration=133.586066929 podStartE2EDuration="2m13.586066929s" podCreationTimestamp="2025-12-11 08:11:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:13:41.584863608 +0000 UTC m=+154.313382673" watchObservedRunningTime="2025-12-11 08:13:41.586066929 +0000 UTC m=+154.314585984" Dec 11 08:13:41 crc kubenswrapper[4860]: I1211 08:13:41.592991 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jgk9d" event={"ID":"5c0bfb74-4985-47a1-ade8-17afad50fabe","Type":"ContainerStarted","Data":"91faa457af7497f7edbebbd3da6680c12bb6e3c99013d7e1290f646c0e5692e2"} Dec 11 08:13:41 crc kubenswrapper[4860]: I1211 08:13:41.593031 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jgk9d" event={"ID":"5c0bfb74-4985-47a1-ade8-17afad50fabe","Type":"ContainerStarted","Data":"081ecdd547a1d267dde6bf0c7be01fe47e497acafccaed0a882ae1eccbeb2d40"} Dec 11 08:13:41 crc kubenswrapper[4860]: I1211 08:13:41.593041 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jgk9d" event={"ID":"5c0bfb74-4985-47a1-ade8-17afad50fabe","Type":"ContainerStarted","Data":"25f9c25eac580301864d37f520a1cc172f468ba69d7e33fefca7704c399c9d36"} Dec 11 08:13:41 crc kubenswrapper[4860]: I1211 08:13:41.597054 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-s6qsz" event={"ID":"c4c320f7-3b40-4edd-8e25-40b477e1d374","Type":"ContainerStarted","Data":"053ff0910647fa69e568e5c56556bbf647a5926cd6a802a783cb69e43ce6c4fd"} Dec 11 08:13:41 crc kubenswrapper[4860]: I1211 08:13:41.608876 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-x28lp" podStartSLOduration=133.608858881 podStartE2EDuration="2m13.608858881s" podCreationTimestamp="2025-12-11 08:11:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:13:41.608078374 +0000 UTC m=+154.336597429" watchObservedRunningTime="2025-12-11 08:13:41.608858881 +0000 UTC m=+154.337377936" Dec 11 08:13:41 crc kubenswrapper[4860]: I1211 08:13:41.609150 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-s6qsz" Dec 11 08:13:41 crc kubenswrapper[4860]: I1211 08:13:41.610156 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-h7zxk" event={"ID":"b08141e2-41b8-45f4-bb40-c319621719c8","Type":"ContainerStarted","Data":"83a562b692199e0e9655fc6e127276a9e7cd6d182e58fb73cd4f83d67448ab25"} Dec 11 08:13:41 crc kubenswrapper[4860]: I1211 08:13:41.610185 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-h7zxk" event={"ID":"b08141e2-41b8-45f4-bb40-c319621719c8","Type":"ContainerStarted","Data":"b559d0c25efb0429df1882fb7a408224ba7a61f1c5d6a87ba8b7486691aba246"} Dec 11 08:13:41 crc kubenswrapper[4860]: I1211 08:13:41.618884 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-87b7k" event={"ID":"73ab8adf-cf47-4482-bbdb-8fb6c3edc87a","Type":"ContainerStarted","Data":"ed658c6c2c4c4a661b4f95ad3cdd1ddf32c8e29375acb39b8f4467ba79fb6113"} Dec 11 08:13:41 crc kubenswrapper[4860]: I1211 08:13:41.619264 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-87b7k" Dec 11 08:13:41 crc kubenswrapper[4860]: I1211 08:13:41.621579 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-l4m2j" event={"ID":"5b8b374c-b131-402c-8e3b-02d93d416489","Type":"ContainerStarted","Data":"4fb596376e680b427384e7f52cb68be49005abd4e62dcb52005797dee95ac9df"} Dec 11 08:13:41 crc kubenswrapper[4860]: I1211 08:13:41.623954 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-76lnq" event={"ID":"7a764867-5216-47fb-8211-cea78b79f707","Type":"ContainerStarted","Data":"2dd94e32c8aaeab1d6b9ba881c97eb877197e7ace7cf79aa5a6cef6a37db6f63"} Dec 11 08:13:41 crc kubenswrapper[4860]: I1211 08:13:41.623984 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-76lnq" event={"ID":"7a764867-5216-47fb-8211-cea78b79f707","Type":"ContainerStarted","Data":"c254b6625b6e40d25177edf1f2d70cba2fd9651c986b6affc6aa039745e994e4"} Dec 11 08:13:41 crc kubenswrapper[4860]: I1211 08:13:41.626700 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-t8ffw" event={"ID":"5010fcd8-99df-473e-8ed3-9ae00612be20","Type":"ContainerStarted","Data":"27fa6a44096083dee1c0dfc49ad645c04b2d1bd101cdc439be1d7b4b6c04a726"} Dec 11 08:13:41 crc kubenswrapper[4860]: I1211 08:13:41.626724 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-t8ffw" event={"ID":"5010fcd8-99df-473e-8ed3-9ae00612be20","Type":"ContainerStarted","Data":"70baccbec0af21a58dbc9282e4281bd1423661629a5a128eaeb8e4a3099348a2"} Dec 11 08:13:41 crc kubenswrapper[4860]: I1211 08:13:41.627756 4860 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-ttrm2 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.38:8080/healthz\": dial tcp 10.217.0.38:8080: connect: connection refused" start-of-body= Dec 11 08:13:41 crc kubenswrapper[4860]: I1211 08:13:41.627791 4860 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-ttrm2" podUID="ad724960-18d1-4e83-944c-4cdeb033a436" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.38:8080/healthz\": dial tcp 10.217.0.38:8080: connect: connection refused" Dec 11 08:13:41 crc kubenswrapper[4860]: I1211 08:13:41.629446 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jgk9d" podStartSLOduration=133.629432097 podStartE2EDuration="2m13.629432097s" podCreationTimestamp="2025-12-11 08:11:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:13:41.626096272 +0000 UTC m=+154.354615337" watchObservedRunningTime="2025-12-11 08:13:41.629432097 +0000 UTC m=+154.357951152" Dec 11 08:13:41 crc kubenswrapper[4860]: I1211 08:13:41.633945 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cpflv" Dec 11 08:13:41 crc kubenswrapper[4860]: I1211 08:13:41.641987 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zf572\" (UID: \"38005ee2-6635-498c-b156-ac57b4f2f4d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-zf572" Dec 11 08:13:41 crc kubenswrapper[4860]: E1211 08:13:41.642761 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:13:42.142745704 +0000 UTC m=+154.871264819 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zf572" (UID: "38005ee2-6635-498c-b156-ac57b4f2f4d9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:13:41 crc kubenswrapper[4860]: I1211 08:13:41.653447 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-87b7k" podStartSLOduration=133.653421399 podStartE2EDuration="2m13.653421399s" podCreationTimestamp="2025-12-11 08:11:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:13:41.650175748 +0000 UTC m=+154.378694813" watchObservedRunningTime="2025-12-11 08:13:41.653421399 +0000 UTC m=+154.381940454" Dec 11 08:13:41 crc kubenswrapper[4860]: I1211 08:13:41.682325 4860 patch_prober.go:28] interesting pod/router-default-5444994796-n8wrd container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 11 08:13:41 crc kubenswrapper[4860]: [-]has-synced failed: reason withheld Dec 11 08:13:41 crc kubenswrapper[4860]: [+]process-running ok Dec 11 08:13:41 crc kubenswrapper[4860]: healthz check failed Dec 11 08:13:41 crc kubenswrapper[4860]: I1211 08:13:41.682389 4860 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-n8wrd" podUID="30bec2ef-4635-4766-a34c-423c2e7f471b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 11 08:13:41 crc kubenswrapper[4860]: I1211 08:13:41.683813 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-h7zxk" podStartSLOduration=133.683792081 podStartE2EDuration="2m13.683792081s" podCreationTimestamp="2025-12-11 08:11:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:13:41.683303044 +0000 UTC m=+154.411822099" watchObservedRunningTime="2025-12-11 08:13:41.683792081 +0000 UTC m=+154.412311136" Dec 11 08:13:41 crc kubenswrapper[4860]: I1211 08:13:41.743410 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:13:41 crc kubenswrapper[4860]: I1211 08:13:41.743494 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-l4m2j" podStartSLOduration=133.743485369 podStartE2EDuration="2m13.743485369s" podCreationTimestamp="2025-12-11 08:11:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:13:41.742258626 +0000 UTC m=+154.470777691" watchObservedRunningTime="2025-12-11 08:13:41.743485369 +0000 UTC m=+154.472004424" Dec 11 08:13:41 crc kubenswrapper[4860]: E1211 08:13:41.745497 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:13:42.245456566 +0000 UTC m=+154.973975621 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:13:41 crc kubenswrapper[4860]: I1211 08:13:41.763940 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-87b7k" Dec 11 08:13:41 crc kubenswrapper[4860]: I1211 08:13:41.803700 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-76lnq" podStartSLOduration=133.803682853 podStartE2EDuration="2m13.803682853s" podCreationTimestamp="2025-12-11 08:11:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:13:41.801061003 +0000 UTC m=+154.529580058" watchObservedRunningTime="2025-12-11 08:13:41.803682853 +0000 UTC m=+154.532201908" Dec 11 08:13:41 crc kubenswrapper[4860]: I1211 08:13:41.847463 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zf572\" (UID: \"38005ee2-6635-498c-b156-ac57b4f2f4d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-zf572" Dec 11 08:13:41 crc kubenswrapper[4860]: E1211 08:13:41.848198 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:13:42.348183869 +0000 UTC m=+155.076702924 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zf572" (UID: "38005ee2-6635-498c-b156-ac57b4f2f4d9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:13:41 crc kubenswrapper[4860]: I1211 08:13:41.850357 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rvtb7" Dec 11 08:13:41 crc kubenswrapper[4860]: I1211 08:13:41.950632 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:13:41 crc kubenswrapper[4860]: E1211 08:13:41.951562 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:13:42.451546145 +0000 UTC m=+155.180065200 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:13:42 crc kubenswrapper[4860]: I1211 08:13:42.039490 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-t8ffw" podStartSLOduration=8.03947414 podStartE2EDuration="8.03947414s" podCreationTimestamp="2025-12-11 08:13:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:13:42.036548279 +0000 UTC m=+154.765067334" watchObservedRunningTime="2025-12-11 08:13:42.03947414 +0000 UTC m=+154.767993195" Dec 11 08:13:42 crc kubenswrapper[4860]: I1211 08:13:42.040369 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-svdpj" podStartSLOduration=134.040362461 podStartE2EDuration="2m14.040362461s" podCreationTimestamp="2025-12-11 08:11:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:13:42.016712279 +0000 UTC m=+154.745231364" watchObservedRunningTime="2025-12-11 08:13:42.040362461 +0000 UTC m=+154.768881516" Dec 11 08:13:42 crc kubenswrapper[4860]: I1211 08:13:42.052042 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zf572\" (UID: \"38005ee2-6635-498c-b156-ac57b4f2f4d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-zf572" Dec 11 08:13:42 crc kubenswrapper[4860]: E1211 08:13:42.061015 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:13:42.560990977 +0000 UTC m=+155.289510022 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zf572" (UID: "38005ee2-6635-498c-b156-ac57b4f2f4d9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:13:42 crc kubenswrapper[4860]: E1211 08:13:42.152957 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:13:42.652914631 +0000 UTC m=+155.381433696 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:13:42 crc kubenswrapper[4860]: I1211 08:13:42.153020 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:13:42 crc kubenswrapper[4860]: I1211 08:13:42.153383 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zf572\" (UID: \"38005ee2-6635-498c-b156-ac57b4f2f4d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-zf572" Dec 11 08:13:42 crc kubenswrapper[4860]: E1211 08:13:42.153732 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:13:42.653724008 +0000 UTC m=+155.382243063 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zf572" (UID: "38005ee2-6635-498c-b156-ac57b4f2f4d9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:13:42 crc kubenswrapper[4860]: I1211 08:13:42.163086 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fsfs7" Dec 11 08:13:42 crc kubenswrapper[4860]: I1211 08:13:42.200936 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-kqwgm"] Dec 11 08:13:42 crc kubenswrapper[4860]: I1211 08:13:42.201887 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kqwgm" Dec 11 08:13:42 crc kubenswrapper[4860]: I1211 08:13:42.209399 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 11 08:13:42 crc kubenswrapper[4860]: I1211 08:13:42.218631 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kqwgm"] Dec 11 08:13:42 crc kubenswrapper[4860]: I1211 08:13:42.256051 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:13:42 crc kubenswrapper[4860]: E1211 08:13:42.256631 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:13:42.756614968 +0000 UTC m=+155.485134023 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:13:42 crc kubenswrapper[4860]: I1211 08:13:42.358431 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zf572\" (UID: \"38005ee2-6635-498c-b156-ac57b4f2f4d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-zf572" Dec 11 08:13:42 crc kubenswrapper[4860]: I1211 08:13:42.358522 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f92fq\" (UniqueName: \"kubernetes.io/projected/093a9903-7041-410b-bfb2-c74ef741e0be-kube-api-access-f92fq\") pod \"certified-operators-kqwgm\" (UID: \"093a9903-7041-410b-bfb2-c74ef741e0be\") " pod="openshift-marketplace/certified-operators-kqwgm" Dec 11 08:13:42 crc kubenswrapper[4860]: I1211 08:13:42.358544 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/093a9903-7041-410b-bfb2-c74ef741e0be-utilities\") pod \"certified-operators-kqwgm\" (UID: \"093a9903-7041-410b-bfb2-c74ef741e0be\") " pod="openshift-marketplace/certified-operators-kqwgm" Dec 11 08:13:42 crc kubenswrapper[4860]: I1211 08:13:42.358570 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/093a9903-7041-410b-bfb2-c74ef741e0be-catalog-content\") pod \"certified-operators-kqwgm\" (UID: \"093a9903-7041-410b-bfb2-c74ef741e0be\") " pod="openshift-marketplace/certified-operators-kqwgm" Dec 11 08:13:42 crc kubenswrapper[4860]: E1211 08:13:42.358850 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:13:42.858838283 +0000 UTC m=+155.587357338 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zf572" (UID: "38005ee2-6635-498c-b156-ac57b4f2f4d9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:13:42 crc kubenswrapper[4860]: I1211 08:13:42.383528 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-pqtvk"] Dec 11 08:13:42 crc kubenswrapper[4860]: I1211 08:13:42.384425 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pqtvk" Dec 11 08:13:42 crc kubenswrapper[4860]: I1211 08:13:42.389274 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 11 08:13:42 crc kubenswrapper[4860]: I1211 08:13:42.397328 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pqtvk"] Dec 11 08:13:42 crc kubenswrapper[4860]: I1211 08:13:42.463093 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:13:42 crc kubenswrapper[4860]: I1211 08:13:42.463433 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/093a9903-7041-410b-bfb2-c74ef741e0be-catalog-content\") pod \"certified-operators-kqwgm\" (UID: \"093a9903-7041-410b-bfb2-c74ef741e0be\") " pod="openshift-marketplace/certified-operators-kqwgm" Dec 11 08:13:42 crc kubenswrapper[4860]: I1211 08:13:42.463494 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ac89442-1754-45d3-b67a-aa6cc31a1235-utilities\") pod \"community-operators-pqtvk\" (UID: \"7ac89442-1754-45d3-b67a-aa6cc31a1235\") " pod="openshift-marketplace/community-operators-pqtvk" Dec 11 08:13:42 crc kubenswrapper[4860]: I1211 08:13:42.463616 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tf7hl\" (UniqueName: \"kubernetes.io/projected/7ac89442-1754-45d3-b67a-aa6cc31a1235-kube-api-access-tf7hl\") pod \"community-operators-pqtvk\" (UID: \"7ac89442-1754-45d3-b67a-aa6cc31a1235\") " pod="openshift-marketplace/community-operators-pqtvk" Dec 11 08:13:42 crc kubenswrapper[4860]: I1211 08:13:42.463665 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ac89442-1754-45d3-b67a-aa6cc31a1235-catalog-content\") pod \"community-operators-pqtvk\" (UID: \"7ac89442-1754-45d3-b67a-aa6cc31a1235\") " pod="openshift-marketplace/community-operators-pqtvk" Dec 11 08:13:42 crc kubenswrapper[4860]: I1211 08:13:42.463723 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f92fq\" (UniqueName: \"kubernetes.io/projected/093a9903-7041-410b-bfb2-c74ef741e0be-kube-api-access-f92fq\") pod \"certified-operators-kqwgm\" (UID: \"093a9903-7041-410b-bfb2-c74ef741e0be\") " pod="openshift-marketplace/certified-operators-kqwgm" Dec 11 08:13:42 crc kubenswrapper[4860]: E1211 08:13:42.463823 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:13:42.963776633 +0000 UTC m=+155.692295688 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:13:42 crc kubenswrapper[4860]: I1211 08:13:42.463953 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/093a9903-7041-410b-bfb2-c74ef741e0be-utilities\") pod \"certified-operators-kqwgm\" (UID: \"093a9903-7041-410b-bfb2-c74ef741e0be\") " pod="openshift-marketplace/certified-operators-kqwgm" Dec 11 08:13:42 crc kubenswrapper[4860]: I1211 08:13:42.464087 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/093a9903-7041-410b-bfb2-c74ef741e0be-catalog-content\") pod \"certified-operators-kqwgm\" (UID: \"093a9903-7041-410b-bfb2-c74ef741e0be\") " pod="openshift-marketplace/certified-operators-kqwgm" Dec 11 08:13:42 crc kubenswrapper[4860]: I1211 08:13:42.464375 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/093a9903-7041-410b-bfb2-c74ef741e0be-utilities\") pod \"certified-operators-kqwgm\" (UID: \"093a9903-7041-410b-bfb2-c74ef741e0be\") " pod="openshift-marketplace/certified-operators-kqwgm" Dec 11 08:13:42 crc kubenswrapper[4860]: I1211 08:13:42.492557 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f92fq\" (UniqueName: \"kubernetes.io/projected/093a9903-7041-410b-bfb2-c74ef741e0be-kube-api-access-f92fq\") pod \"certified-operators-kqwgm\" (UID: \"093a9903-7041-410b-bfb2-c74ef741e0be\") " pod="openshift-marketplace/certified-operators-kqwgm" Dec 11 08:13:42 crc kubenswrapper[4860]: I1211 08:13:42.527572 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kqwgm" Dec 11 08:13:42 crc kubenswrapper[4860]: I1211 08:13:42.565210 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ac89442-1754-45d3-b67a-aa6cc31a1235-utilities\") pod \"community-operators-pqtvk\" (UID: \"7ac89442-1754-45d3-b67a-aa6cc31a1235\") " pod="openshift-marketplace/community-operators-pqtvk" Dec 11 08:13:42 crc kubenswrapper[4860]: I1211 08:13:42.565250 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zf572\" (UID: \"38005ee2-6635-498c-b156-ac57b4f2f4d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-zf572" Dec 11 08:13:42 crc kubenswrapper[4860]: I1211 08:13:42.565304 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tf7hl\" (UniqueName: \"kubernetes.io/projected/7ac89442-1754-45d3-b67a-aa6cc31a1235-kube-api-access-tf7hl\") pod \"community-operators-pqtvk\" (UID: \"7ac89442-1754-45d3-b67a-aa6cc31a1235\") " pod="openshift-marketplace/community-operators-pqtvk" Dec 11 08:13:42 crc kubenswrapper[4860]: I1211 08:13:42.565326 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ac89442-1754-45d3-b67a-aa6cc31a1235-catalog-content\") pod \"community-operators-pqtvk\" (UID: \"7ac89442-1754-45d3-b67a-aa6cc31a1235\") " pod="openshift-marketplace/community-operators-pqtvk" Dec 11 08:13:42 crc kubenswrapper[4860]: E1211 08:13:42.565723 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:13:43.065706888 +0000 UTC m=+155.794225933 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zf572" (UID: "38005ee2-6635-498c-b156-ac57b4f2f4d9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:13:42 crc kubenswrapper[4860]: I1211 08:13:42.566110 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ac89442-1754-45d3-b67a-aa6cc31a1235-utilities\") pod \"community-operators-pqtvk\" (UID: \"7ac89442-1754-45d3-b67a-aa6cc31a1235\") " pod="openshift-marketplace/community-operators-pqtvk" Dec 11 08:13:42 crc kubenswrapper[4860]: I1211 08:13:42.566132 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ac89442-1754-45d3-b67a-aa6cc31a1235-catalog-content\") pod \"community-operators-pqtvk\" (UID: \"7ac89442-1754-45d3-b67a-aa6cc31a1235\") " pod="openshift-marketplace/community-operators-pqtvk" Dec 11 08:13:42 crc kubenswrapper[4860]: I1211 08:13:42.603111 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-55jtz"] Dec 11 08:13:42 crc kubenswrapper[4860]: I1211 08:13:42.604342 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-55jtz" Dec 11 08:13:42 crc kubenswrapper[4860]: I1211 08:13:42.609318 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tf7hl\" (UniqueName: \"kubernetes.io/projected/7ac89442-1754-45d3-b67a-aa6cc31a1235-kube-api-access-tf7hl\") pod \"community-operators-pqtvk\" (UID: \"7ac89442-1754-45d3-b67a-aa6cc31a1235\") " pod="openshift-marketplace/community-operators-pqtvk" Dec 11 08:13:42 crc kubenswrapper[4860]: I1211 08:13:42.633904 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-55jtz"] Dec 11 08:13:42 crc kubenswrapper[4860]: I1211 08:13:42.640468 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9tkth" event={"ID":"67ae7457-ee23-4ae3-a118-fbfac6dd85bf","Type":"ContainerStarted","Data":"c2982640958203e882a934435f8ed59a1783edd9c44f57a6c9307ff84052ba6b"} Dec 11 08:13:42 crc kubenswrapper[4860]: I1211 08:13:42.650149 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-kbz2w" event={"ID":"eceb74a4-c695-46e9-ba5f-b5145d20d968","Type":"ContainerStarted","Data":"f8f226cb888a2837af91bf204e3aa3f8f136c8a17e56c273e3fe3453e80a26d0"} Dec 11 08:13:42 crc kubenswrapper[4860]: I1211 08:13:42.667443 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:13:42 crc kubenswrapper[4860]: I1211 08:13:42.667734 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gf8wg\" (UniqueName: \"kubernetes.io/projected/b498b689-633b-4fea-9981-159362f44053-kube-api-access-gf8wg\") pod \"certified-operators-55jtz\" (UID: \"b498b689-633b-4fea-9981-159362f44053\") " pod="openshift-marketplace/certified-operators-55jtz" Dec 11 08:13:42 crc kubenswrapper[4860]: I1211 08:13:42.667778 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b498b689-633b-4fea-9981-159362f44053-utilities\") pod \"certified-operators-55jtz\" (UID: \"b498b689-633b-4fea-9981-159362f44053\") " pod="openshift-marketplace/certified-operators-55jtz" Dec 11 08:13:42 crc kubenswrapper[4860]: I1211 08:13:42.667796 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b498b689-633b-4fea-9981-159362f44053-catalog-content\") pod \"certified-operators-55jtz\" (UID: \"b498b689-633b-4fea-9981-159362f44053\") " pod="openshift-marketplace/certified-operators-55jtz" Dec 11 08:13:42 crc kubenswrapper[4860]: E1211 08:13:42.667914 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-11 08:13:43.167899213 +0000 UTC m=+155.896418268 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:13:42 crc kubenswrapper[4860]: I1211 08:13:42.680305 4860 patch_prober.go:28] interesting pod/router-default-5444994796-n8wrd container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 11 08:13:42 crc kubenswrapper[4860]: [-]has-synced failed: reason withheld Dec 11 08:13:42 crc kubenswrapper[4860]: [+]process-running ok Dec 11 08:13:42 crc kubenswrapper[4860]: healthz check failed Dec 11 08:13:42 crc kubenswrapper[4860]: I1211 08:13:42.680369 4860 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-n8wrd" podUID="30bec2ef-4635-4766-a34c-423c2e7f471b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 11 08:13:42 crc kubenswrapper[4860]: I1211 08:13:42.690254 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-sbcxd" event={"ID":"220b6958-d8ee-4733-a93a-35d2ced9f150","Type":"ContainerStarted","Data":"f565c4c117cdb7627f3fc2e147d64f44c29c96103fa4453df17e5ffbfa2794ef"} Dec 11 08:13:42 crc kubenswrapper[4860]: I1211 08:13:42.690306 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-sbcxd" event={"ID":"220b6958-d8ee-4733-a93a-35d2ced9f150","Type":"ContainerStarted","Data":"78a309d8066a399d3740038a831bfb66e69cf340df99879557736a7191ae24a3"} Dec 11 08:13:42 crc kubenswrapper[4860]: I1211 08:13:42.690800 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-sbcxd" Dec 11 08:13:42 crc kubenswrapper[4860]: I1211 08:13:42.703025 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pqtvk" Dec 11 08:13:42 crc kubenswrapper[4860]: I1211 08:13:42.710673 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-h7zxk" event={"ID":"b08141e2-41b8-45f4-bb40-c319621719c8","Type":"ContainerStarted","Data":"784a5706b95687423cce83e45d4c2b8ff6d4d7df5a6c31b8818812c23e619a5f"} Dec 11 08:13:42 crc kubenswrapper[4860]: I1211 08:13:42.713350 4860 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-ttrm2 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.38:8080/healthz\": dial tcp 10.217.0.38:8080: connect: connection refused" start-of-body= Dec 11 08:13:42 crc kubenswrapper[4860]: I1211 08:13:42.713382 4860 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-ttrm2" podUID="ad724960-18d1-4e83-944c-4cdeb033a436" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.38:8080/healthz\": dial tcp 10.217.0.38:8080: connect: connection refused" Dec 11 08:13:42 crc kubenswrapper[4860]: I1211 08:13:42.715414 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-svdpj" Dec 11 08:13:42 crc kubenswrapper[4860]: I1211 08:13:42.722580 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9tkth" podStartSLOduration=134.722565878 podStartE2EDuration="2m14.722565878s" podCreationTimestamp="2025-12-11 08:11:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:13:42.667063064 +0000 UTC m=+155.395582119" watchObservedRunningTime="2025-12-11 08:13:42.722565878 +0000 UTC m=+155.451084933" Dec 11 08:13:42 crc kubenswrapper[4860]: I1211 08:13:42.723859 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-sbcxd" podStartSLOduration=8.723853692 podStartE2EDuration="8.723853692s" podCreationTimestamp="2025-12-11 08:13:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:13:42.721568833 +0000 UTC m=+155.450087898" watchObservedRunningTime="2025-12-11 08:13:42.723853692 +0000 UTC m=+155.452372747" Dec 11 08:13:42 crc kubenswrapper[4860]: I1211 08:13:42.734610 4860 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Dec 11 08:13:42 crc kubenswrapper[4860]: I1211 08:13:42.738257 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-svdpj" Dec 11 08:13:42 crc kubenswrapper[4860]: I1211 08:13:42.770302 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zf572\" (UID: \"38005ee2-6635-498c-b156-ac57b4f2f4d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-zf572" Dec 11 08:13:42 crc kubenswrapper[4860]: I1211 08:13:42.770368 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b498b689-633b-4fea-9981-159362f44053-utilities\") pod \"certified-operators-55jtz\" (UID: \"b498b689-633b-4fea-9981-159362f44053\") " pod="openshift-marketplace/certified-operators-55jtz" Dec 11 08:13:42 crc kubenswrapper[4860]: I1211 08:13:42.770392 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b498b689-633b-4fea-9981-159362f44053-catalog-content\") pod \"certified-operators-55jtz\" (UID: \"b498b689-633b-4fea-9981-159362f44053\") " pod="openshift-marketplace/certified-operators-55jtz" Dec 11 08:13:42 crc kubenswrapper[4860]: I1211 08:13:42.770545 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gf8wg\" (UniqueName: \"kubernetes.io/projected/b498b689-633b-4fea-9981-159362f44053-kube-api-access-gf8wg\") pod \"certified-operators-55jtz\" (UID: \"b498b689-633b-4fea-9981-159362f44053\") " pod="openshift-marketplace/certified-operators-55jtz" Dec 11 08:13:42 crc kubenswrapper[4860]: E1211 08:13:42.771139 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-11 08:13:43.271123273 +0000 UTC m=+155.999642328 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zf572" (UID: "38005ee2-6635-498c-b156-ac57b4f2f4d9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 08:13:42 crc kubenswrapper[4860]: I1211 08:13:42.772486 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b498b689-633b-4fea-9981-159362f44053-utilities\") pod \"certified-operators-55jtz\" (UID: \"b498b689-633b-4fea-9981-159362f44053\") " pod="openshift-marketplace/certified-operators-55jtz" Dec 11 08:13:42 crc kubenswrapper[4860]: I1211 08:13:42.772946 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b498b689-633b-4fea-9981-159362f44053-catalog-content\") pod \"certified-operators-55jtz\" (UID: \"b498b689-633b-4fea-9981-159362f44053\") " pod="openshift-marketplace/certified-operators-55jtz" Dec 11 08:13:42 crc kubenswrapper[4860]: I1211 08:13:42.801773 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-p6f75"] Dec 11 08:13:42 crc kubenswrapper[4860]: I1211 08:13:42.803269 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p6f75" Dec 11 08:13:42 crc kubenswrapper[4860]: I1211 08:13:42.804794 4860 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-12-11T08:13:42.734685384Z","Handler":null,"Name":""} Dec 11 08:13:42 crc kubenswrapper[4860]: I1211 08:13:42.819600 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gf8wg\" (UniqueName: \"kubernetes.io/projected/b498b689-633b-4fea-9981-159362f44053-kube-api-access-gf8wg\") pod \"certified-operators-55jtz\" (UID: \"b498b689-633b-4fea-9981-159362f44053\") " pod="openshift-marketplace/certified-operators-55jtz" Dec 11 08:13:42 crc kubenswrapper[4860]: I1211 08:13:42.842006 4860 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Dec 11 08:13:42 crc kubenswrapper[4860]: I1211 08:13:42.842042 4860 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Dec 11 08:13:42 crc kubenswrapper[4860]: I1211 08:13:42.844065 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-p6f75"] Dec 11 08:13:42 crc kubenswrapper[4860]: I1211 08:13:42.882954 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 11 08:13:42 crc kubenswrapper[4860]: I1211 08:13:42.883679 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-whc24\" (UniqueName: \"kubernetes.io/projected/f9b23a65-98b8-4d83-a71f-2316fc1d46a0-kube-api-access-whc24\") pod \"community-operators-p6f75\" (UID: \"f9b23a65-98b8-4d83-a71f-2316fc1d46a0\") " pod="openshift-marketplace/community-operators-p6f75" Dec 11 08:13:42 crc kubenswrapper[4860]: I1211 08:13:42.926532 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 11 08:13:42 crc kubenswrapper[4860]: I1211 08:13:42.928286 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f9b23a65-98b8-4d83-a71f-2316fc1d46a0-catalog-content\") pod \"community-operators-p6f75\" (UID: \"f9b23a65-98b8-4d83-a71f-2316fc1d46a0\") " pod="openshift-marketplace/community-operators-p6f75" Dec 11 08:13:42 crc kubenswrapper[4860]: I1211 08:13:42.928572 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zf572\" (UID: \"38005ee2-6635-498c-b156-ac57b4f2f4d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-zf572" Dec 11 08:13:42 crc kubenswrapper[4860]: I1211 08:13:42.947555 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 11 08:13:42 crc kubenswrapper[4860]: I1211 08:13:42.955927 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 11 08:13:42 crc kubenswrapper[4860]: I1211 08:13:42.956714 4860 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 11 08:13:42 crc kubenswrapper[4860]: I1211 08:13:42.956748 4860 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zf572\" (UID: \"38005ee2-6635-498c-b156-ac57b4f2f4d9\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-zf572" Dec 11 08:13:42 crc kubenswrapper[4860]: I1211 08:13:42.959469 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-55jtz" Dec 11 08:13:42 crc kubenswrapper[4860]: I1211 08:13:42.961139 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Dec 11 08:13:42 crc kubenswrapper[4860]: I1211 08:13:42.961243 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Dec 11 08:13:43 crc kubenswrapper[4860]: I1211 08:13:42.998523 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f9b23a65-98b8-4d83-a71f-2316fc1d46a0-utilities\") pod \"community-operators-p6f75\" (UID: \"f9b23a65-98b8-4d83-a71f-2316fc1d46a0\") " pod="openshift-marketplace/community-operators-p6f75" Dec 11 08:13:43 crc kubenswrapper[4860]: I1211 08:13:43.010761 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 11 08:13:43 crc kubenswrapper[4860]: I1211 08:13:43.104350 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f9b23a65-98b8-4d83-a71f-2316fc1d46a0-utilities\") pod \"community-operators-p6f75\" (UID: \"f9b23a65-98b8-4d83-a71f-2316fc1d46a0\") " pod="openshift-marketplace/community-operators-p6f75" Dec 11 08:13:43 crc kubenswrapper[4860]: I1211 08:13:43.104414 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-whc24\" (UniqueName: \"kubernetes.io/projected/f9b23a65-98b8-4d83-a71f-2316fc1d46a0-kube-api-access-whc24\") pod \"community-operators-p6f75\" (UID: \"f9b23a65-98b8-4d83-a71f-2316fc1d46a0\") " pod="openshift-marketplace/community-operators-p6f75" Dec 11 08:13:43 crc kubenswrapper[4860]: I1211 08:13:43.104454 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/50e857f4-cdd5-4fb0-8b5d-812fdca09adf-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"50e857f4-cdd5-4fb0-8b5d-812fdca09adf\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 11 08:13:43 crc kubenswrapper[4860]: I1211 08:13:43.104483 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f9b23a65-98b8-4d83-a71f-2316fc1d46a0-catalog-content\") pod \"community-operators-p6f75\" (UID: \"f9b23a65-98b8-4d83-a71f-2316fc1d46a0\") " pod="openshift-marketplace/community-operators-p6f75" Dec 11 08:13:43 crc kubenswrapper[4860]: I1211 08:13:43.104521 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/50e857f4-cdd5-4fb0-8b5d-812fdca09adf-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"50e857f4-cdd5-4fb0-8b5d-812fdca09adf\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 11 08:13:43 crc kubenswrapper[4860]: I1211 08:13:43.104894 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f9b23a65-98b8-4d83-a71f-2316fc1d46a0-utilities\") pod \"community-operators-p6f75\" (UID: \"f9b23a65-98b8-4d83-a71f-2316fc1d46a0\") " pod="openshift-marketplace/community-operators-p6f75" Dec 11 08:13:43 crc kubenswrapper[4860]: I1211 08:13:43.105141 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f9b23a65-98b8-4d83-a71f-2316fc1d46a0-catalog-content\") pod \"community-operators-p6f75\" (UID: \"f9b23a65-98b8-4d83-a71f-2316fc1d46a0\") " pod="openshift-marketplace/community-operators-p6f75" Dec 11 08:13:43 crc kubenswrapper[4860]: I1211 08:13:43.151135 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-whc24\" (UniqueName: \"kubernetes.io/projected/f9b23a65-98b8-4d83-a71f-2316fc1d46a0-kube-api-access-whc24\") pod \"community-operators-p6f75\" (UID: \"f9b23a65-98b8-4d83-a71f-2316fc1d46a0\") " pod="openshift-marketplace/community-operators-p6f75" Dec 11 08:13:43 crc kubenswrapper[4860]: I1211 08:13:43.176982 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p6f75" Dec 11 08:13:43 crc kubenswrapper[4860]: I1211 08:13:43.209547 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/50e857f4-cdd5-4fb0-8b5d-812fdca09adf-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"50e857f4-cdd5-4fb0-8b5d-812fdca09adf\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 11 08:13:43 crc kubenswrapper[4860]: I1211 08:13:43.209596 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/50e857f4-cdd5-4fb0-8b5d-812fdca09adf-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"50e857f4-cdd5-4fb0-8b5d-812fdca09adf\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 11 08:13:43 crc kubenswrapper[4860]: I1211 08:13:43.209700 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/50e857f4-cdd5-4fb0-8b5d-812fdca09adf-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"50e857f4-cdd5-4fb0-8b5d-812fdca09adf\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 11 08:13:43 crc kubenswrapper[4860]: I1211 08:13:43.237100 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kqwgm"] Dec 11 08:13:43 crc kubenswrapper[4860]: I1211 08:13:43.240608 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/50e857f4-cdd5-4fb0-8b5d-812fdca09adf-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"50e857f4-cdd5-4fb0-8b5d-812fdca09adf\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 11 08:13:43 crc kubenswrapper[4860]: I1211 08:13:43.269469 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zf572\" (UID: \"38005ee2-6635-498c-b156-ac57b4f2f4d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-zf572" Dec 11 08:13:43 crc kubenswrapper[4860]: I1211 08:13:43.330045 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 11 08:13:43 crc kubenswrapper[4860]: I1211 08:13:43.388318 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pqtvk"] Dec 11 08:13:43 crc kubenswrapper[4860]: I1211 08:13:43.462985 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-zf572" Dec 11 08:13:43 crc kubenswrapper[4860]: I1211 08:13:43.558587 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-55jtz"] Dec 11 08:13:43 crc kubenswrapper[4860]: I1211 08:13:43.617029 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Dec 11 08:13:43 crc kubenswrapper[4860]: I1211 08:13:43.720954 4860 patch_prober.go:28] interesting pod/router-default-5444994796-n8wrd container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 11 08:13:43 crc kubenswrapper[4860]: [-]has-synced failed: reason withheld Dec 11 08:13:43 crc kubenswrapper[4860]: [+]process-running ok Dec 11 08:13:43 crc kubenswrapper[4860]: healthz check failed Dec 11 08:13:43 crc kubenswrapper[4860]: I1211 08:13:43.721014 4860 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-n8wrd" podUID="30bec2ef-4635-4766-a34c-423c2e7f471b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 11 08:13:43 crc kubenswrapper[4860]: I1211 08:13:43.722282 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-kbz2w" event={"ID":"eceb74a4-c695-46e9-ba5f-b5145d20d968","Type":"ContainerStarted","Data":"abc3d422478a53dabcc32c4e26d975f2fe95da38c851b377cbad3658d8456892"} Dec 11 08:13:43 crc kubenswrapper[4860]: I1211 08:13:43.726774 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pqtvk" event={"ID":"7ac89442-1754-45d3-b67a-aa6cc31a1235","Type":"ContainerStarted","Data":"bf503ca2ebe64d9958c793d4c050d3487e5fb82e8e8c12d45bb1796ee873902f"} Dec 11 08:13:43 crc kubenswrapper[4860]: I1211 08:13:43.735828 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-55jtz" event={"ID":"b498b689-633b-4fea-9981-159362f44053","Type":"ContainerStarted","Data":"c2f4f4cd1623e2a47b32453685b7ff3d774fe09f1d80bb0181e5ee581550c48d"} Dec 11 08:13:43 crc kubenswrapper[4860]: I1211 08:13:43.746174 4860 generic.go:334] "Generic (PLEG): container finished" podID="093a9903-7041-410b-bfb2-c74ef741e0be" containerID="1c218b10241590e53b4f67b40eb1342404ed52f9336ace9cde6ab0c86e632357" exitCode=0 Dec 11 08:13:43 crc kubenswrapper[4860]: I1211 08:13:43.747760 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kqwgm" event={"ID":"093a9903-7041-410b-bfb2-c74ef741e0be","Type":"ContainerDied","Data":"1c218b10241590e53b4f67b40eb1342404ed52f9336ace9cde6ab0c86e632357"} Dec 11 08:13:43 crc kubenswrapper[4860]: I1211 08:13:43.747795 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kqwgm" event={"ID":"093a9903-7041-410b-bfb2-c74ef741e0be","Type":"ContainerStarted","Data":"687246172b9ae710d66e39478fe0475ab20c8db92fe23f13ad0072894b373ba6"} Dec 11 08:13:43 crc kubenswrapper[4860]: I1211 08:13:43.748957 4860 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 11 08:13:43 crc kubenswrapper[4860]: I1211 08:13:43.888901 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-p6f75"] Dec 11 08:13:43 crc kubenswrapper[4860]: I1211 08:13:43.928550 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 11 08:13:44 crc kubenswrapper[4860]: I1211 08:13:44.211517 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-dzg4x"] Dec 11 08:13:44 crc kubenswrapper[4860]: I1211 08:13:44.215478 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dzg4x" Dec 11 08:13:44 crc kubenswrapper[4860]: I1211 08:13:44.223049 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 11 08:13:44 crc kubenswrapper[4860]: I1211 08:13:44.259646 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-zf572"] Dec 11 08:13:44 crc kubenswrapper[4860]: I1211 08:13:44.268471 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dzg4x"] Dec 11 08:13:44 crc kubenswrapper[4860]: I1211 08:13:44.343598 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/294792dd-5104-4308-a10f-f3ab0c4504e9-catalog-content\") pod \"redhat-marketplace-dzg4x\" (UID: \"294792dd-5104-4308-a10f-f3ab0c4504e9\") " pod="openshift-marketplace/redhat-marketplace-dzg4x" Dec 11 08:13:44 crc kubenswrapper[4860]: I1211 08:13:44.343650 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/294792dd-5104-4308-a10f-f3ab0c4504e9-utilities\") pod \"redhat-marketplace-dzg4x\" (UID: \"294792dd-5104-4308-a10f-f3ab0c4504e9\") " pod="openshift-marketplace/redhat-marketplace-dzg4x" Dec 11 08:13:44 crc kubenswrapper[4860]: I1211 08:13:44.343702 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ftgqm\" (UniqueName: \"kubernetes.io/projected/294792dd-5104-4308-a10f-f3ab0c4504e9-kube-api-access-ftgqm\") pod \"redhat-marketplace-dzg4x\" (UID: \"294792dd-5104-4308-a10f-f3ab0c4504e9\") " pod="openshift-marketplace/redhat-marketplace-dzg4x" Dec 11 08:13:44 crc kubenswrapper[4860]: I1211 08:13:44.444787 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ftgqm\" (UniqueName: \"kubernetes.io/projected/294792dd-5104-4308-a10f-f3ab0c4504e9-kube-api-access-ftgqm\") pod \"redhat-marketplace-dzg4x\" (UID: \"294792dd-5104-4308-a10f-f3ab0c4504e9\") " pod="openshift-marketplace/redhat-marketplace-dzg4x" Dec 11 08:13:44 crc kubenswrapper[4860]: I1211 08:13:44.444863 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/294792dd-5104-4308-a10f-f3ab0c4504e9-catalog-content\") pod \"redhat-marketplace-dzg4x\" (UID: \"294792dd-5104-4308-a10f-f3ab0c4504e9\") " pod="openshift-marketplace/redhat-marketplace-dzg4x" Dec 11 08:13:44 crc kubenswrapper[4860]: I1211 08:13:44.444894 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/294792dd-5104-4308-a10f-f3ab0c4504e9-utilities\") pod \"redhat-marketplace-dzg4x\" (UID: \"294792dd-5104-4308-a10f-f3ab0c4504e9\") " pod="openshift-marketplace/redhat-marketplace-dzg4x" Dec 11 08:13:44 crc kubenswrapper[4860]: I1211 08:13:44.445330 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/294792dd-5104-4308-a10f-f3ab0c4504e9-utilities\") pod \"redhat-marketplace-dzg4x\" (UID: \"294792dd-5104-4308-a10f-f3ab0c4504e9\") " pod="openshift-marketplace/redhat-marketplace-dzg4x" Dec 11 08:13:44 crc kubenswrapper[4860]: I1211 08:13:44.445817 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/294792dd-5104-4308-a10f-f3ab0c4504e9-catalog-content\") pod \"redhat-marketplace-dzg4x\" (UID: \"294792dd-5104-4308-a10f-f3ab0c4504e9\") " pod="openshift-marketplace/redhat-marketplace-dzg4x" Dec 11 08:13:44 crc kubenswrapper[4860]: I1211 08:13:44.470456 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ftgqm\" (UniqueName: \"kubernetes.io/projected/294792dd-5104-4308-a10f-f3ab0c4504e9-kube-api-access-ftgqm\") pod \"redhat-marketplace-dzg4x\" (UID: \"294792dd-5104-4308-a10f-f3ab0c4504e9\") " pod="openshift-marketplace/redhat-marketplace-dzg4x" Dec 11 08:13:44 crc kubenswrapper[4860]: I1211 08:13:44.549480 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dzg4x" Dec 11 08:13:44 crc kubenswrapper[4860]: I1211 08:13:44.588314 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-dcccn"] Dec 11 08:13:44 crc kubenswrapper[4860]: I1211 08:13:44.589337 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dcccn" Dec 11 08:13:44 crc kubenswrapper[4860]: I1211 08:13:44.617411 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dcccn"] Dec 11 08:13:44 crc kubenswrapper[4860]: I1211 08:13:44.682013 4860 patch_prober.go:28] interesting pod/router-default-5444994796-n8wrd container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 11 08:13:44 crc kubenswrapper[4860]: [-]has-synced failed: reason withheld Dec 11 08:13:44 crc kubenswrapper[4860]: [+]process-running ok Dec 11 08:13:44 crc kubenswrapper[4860]: healthz check failed Dec 11 08:13:44 crc kubenswrapper[4860]: I1211 08:13:44.682394 4860 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-n8wrd" podUID="30bec2ef-4635-4766-a34c-423c2e7f471b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 11 08:13:44 crc kubenswrapper[4860]: I1211 08:13:44.747725 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e052d7c-59d1-43a4-ac7a-e9daf2a88c65-utilities\") pod \"redhat-marketplace-dcccn\" (UID: \"8e052d7c-59d1-43a4-ac7a-e9daf2a88c65\") " pod="openshift-marketplace/redhat-marketplace-dcccn" Dec 11 08:13:44 crc kubenswrapper[4860]: I1211 08:13:44.747780 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x55qf\" (UniqueName: \"kubernetes.io/projected/8e052d7c-59d1-43a4-ac7a-e9daf2a88c65-kube-api-access-x55qf\") pod \"redhat-marketplace-dcccn\" (UID: \"8e052d7c-59d1-43a4-ac7a-e9daf2a88c65\") " pod="openshift-marketplace/redhat-marketplace-dcccn" Dec 11 08:13:44 crc kubenswrapper[4860]: I1211 08:13:44.747845 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e052d7c-59d1-43a4-ac7a-e9daf2a88c65-catalog-content\") pod \"redhat-marketplace-dcccn\" (UID: \"8e052d7c-59d1-43a4-ac7a-e9daf2a88c65\") " pod="openshift-marketplace/redhat-marketplace-dcccn" Dec 11 08:13:44 crc kubenswrapper[4860]: I1211 08:13:44.752579 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p6f75" event={"ID":"f9b23a65-98b8-4d83-a71f-2316fc1d46a0","Type":"ContainerStarted","Data":"a8c240a8856f6d5e175157582fa03875d3ba8ea02a4fa6a5f8fdeefc96eea465"} Dec 11 08:13:44 crc kubenswrapper[4860]: I1211 08:13:44.756129 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-kbz2w" event={"ID":"eceb74a4-c695-46e9-ba5f-b5145d20d968","Type":"ContainerStarted","Data":"99313afc683b921c832275806ac2226bc84e8157bb303e24c22dae8bc965f9b8"} Dec 11 08:13:44 crc kubenswrapper[4860]: I1211 08:13:44.758498 4860 generic.go:334] "Generic (PLEG): container finished" podID="7ac89442-1754-45d3-b67a-aa6cc31a1235" containerID="684bf006e94de3d89bc7b8ef53686ca15a7aa10ba602e2881e0cee03ab001acb" exitCode=0 Dec 11 08:13:44 crc kubenswrapper[4860]: I1211 08:13:44.758579 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pqtvk" event={"ID":"7ac89442-1754-45d3-b67a-aa6cc31a1235","Type":"ContainerDied","Data":"684bf006e94de3d89bc7b8ef53686ca15a7aa10ba602e2881e0cee03ab001acb"} Dec 11 08:13:44 crc kubenswrapper[4860]: I1211 08:13:44.760746 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-zf572" event={"ID":"38005ee2-6635-498c-b156-ac57b4f2f4d9","Type":"ContainerStarted","Data":"292d1c897abc2bdf07e4b29e9ad59768345c805bda79fc0adb56dabb549b08ca"} Dec 11 08:13:44 crc kubenswrapper[4860]: I1211 08:13:44.769315 4860 generic.go:334] "Generic (PLEG): container finished" podID="b498b689-633b-4fea-9981-159362f44053" containerID="0d6492f503343d818f1c697b3a863bfc2e6cf553475503ec2395cae61ed2e3ea" exitCode=0 Dec 11 08:13:44 crc kubenswrapper[4860]: I1211 08:13:44.769445 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-55jtz" event={"ID":"b498b689-633b-4fea-9981-159362f44053","Type":"ContainerDied","Data":"0d6492f503343d818f1c697b3a863bfc2e6cf553475503ec2395cae61ed2e3ea"} Dec 11 08:13:44 crc kubenswrapper[4860]: I1211 08:13:44.773358 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"50e857f4-cdd5-4fb0-8b5d-812fdca09adf","Type":"ContainerStarted","Data":"7ae2db137ba0e7547bfed907a1edcf26510cc342229d5f1897bcf3c6fb4da540"} Dec 11 08:13:44 crc kubenswrapper[4860]: I1211 08:13:44.791352 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-kbz2w" podStartSLOduration=11.79132967 podStartE2EDuration="11.79132967s" podCreationTimestamp="2025-12-11 08:13:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:13:44.790916016 +0000 UTC m=+157.519435071" watchObservedRunningTime="2025-12-11 08:13:44.79132967 +0000 UTC m=+157.519848725" Dec 11 08:13:44 crc kubenswrapper[4860]: I1211 08:13:44.848588 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e052d7c-59d1-43a4-ac7a-e9daf2a88c65-utilities\") pod \"redhat-marketplace-dcccn\" (UID: \"8e052d7c-59d1-43a4-ac7a-e9daf2a88c65\") " pod="openshift-marketplace/redhat-marketplace-dcccn" Dec 11 08:13:44 crc kubenswrapper[4860]: I1211 08:13:44.848627 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x55qf\" (UniqueName: \"kubernetes.io/projected/8e052d7c-59d1-43a4-ac7a-e9daf2a88c65-kube-api-access-x55qf\") pod \"redhat-marketplace-dcccn\" (UID: \"8e052d7c-59d1-43a4-ac7a-e9daf2a88c65\") " pod="openshift-marketplace/redhat-marketplace-dcccn" Dec 11 08:13:44 crc kubenswrapper[4860]: I1211 08:13:44.848742 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e052d7c-59d1-43a4-ac7a-e9daf2a88c65-catalog-content\") pod \"redhat-marketplace-dcccn\" (UID: \"8e052d7c-59d1-43a4-ac7a-e9daf2a88c65\") " pod="openshift-marketplace/redhat-marketplace-dcccn" Dec 11 08:13:44 crc kubenswrapper[4860]: I1211 08:13:44.849867 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e052d7c-59d1-43a4-ac7a-e9daf2a88c65-utilities\") pod \"redhat-marketplace-dcccn\" (UID: \"8e052d7c-59d1-43a4-ac7a-e9daf2a88c65\") " pod="openshift-marketplace/redhat-marketplace-dcccn" Dec 11 08:13:44 crc kubenswrapper[4860]: I1211 08:13:44.852426 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e052d7c-59d1-43a4-ac7a-e9daf2a88c65-catalog-content\") pod \"redhat-marketplace-dcccn\" (UID: \"8e052d7c-59d1-43a4-ac7a-e9daf2a88c65\") " pod="openshift-marketplace/redhat-marketplace-dcccn" Dec 11 08:13:44 crc kubenswrapper[4860]: I1211 08:13:44.871840 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x55qf\" (UniqueName: \"kubernetes.io/projected/8e052d7c-59d1-43a4-ac7a-e9daf2a88c65-kube-api-access-x55qf\") pod \"redhat-marketplace-dcccn\" (UID: \"8e052d7c-59d1-43a4-ac7a-e9daf2a88c65\") " pod="openshift-marketplace/redhat-marketplace-dcccn" Dec 11 08:13:44 crc kubenswrapper[4860]: I1211 08:13:44.937385 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dcccn" Dec 11 08:13:45 crc kubenswrapper[4860]: I1211 08:13:45.030780 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" Dec 11 08:13:45 crc kubenswrapper[4860]: I1211 08:13:45.088131 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dzg4x"] Dec 11 08:13:45 crc kubenswrapper[4860]: I1211 08:13:45.380628 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-8p4mb"] Dec 11 08:13:45 crc kubenswrapper[4860]: I1211 08:13:45.382277 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8p4mb" Dec 11 08:13:45 crc kubenswrapper[4860]: I1211 08:13:45.388550 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 11 08:13:45 crc kubenswrapper[4860]: I1211 08:13:45.393943 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dcccn"] Dec 11 08:13:45 crc kubenswrapper[4860]: I1211 08:13:45.403974 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8p4mb"] Dec 11 08:13:45 crc kubenswrapper[4860]: I1211 08:13:45.467078 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/01481957-a690-40f4-8c89-352b7350d327-utilities\") pod \"redhat-operators-8p4mb\" (UID: \"01481957-a690-40f4-8c89-352b7350d327\") " pod="openshift-marketplace/redhat-operators-8p4mb" Dec 11 08:13:45 crc kubenswrapper[4860]: I1211 08:13:45.467231 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4n7mg\" (UniqueName: \"kubernetes.io/projected/01481957-a690-40f4-8c89-352b7350d327-kube-api-access-4n7mg\") pod \"redhat-operators-8p4mb\" (UID: \"01481957-a690-40f4-8c89-352b7350d327\") " pod="openshift-marketplace/redhat-operators-8p4mb" Dec 11 08:13:45 crc kubenswrapper[4860]: I1211 08:13:45.467258 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/01481957-a690-40f4-8c89-352b7350d327-catalog-content\") pod \"redhat-operators-8p4mb\" (UID: \"01481957-a690-40f4-8c89-352b7350d327\") " pod="openshift-marketplace/redhat-operators-8p4mb" Dec 11 08:13:45 crc kubenswrapper[4860]: I1211 08:13:45.568350 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/01481957-a690-40f4-8c89-352b7350d327-utilities\") pod \"redhat-operators-8p4mb\" (UID: \"01481957-a690-40f4-8c89-352b7350d327\") " pod="openshift-marketplace/redhat-operators-8p4mb" Dec 11 08:13:45 crc kubenswrapper[4860]: I1211 08:13:45.568478 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4n7mg\" (UniqueName: \"kubernetes.io/projected/01481957-a690-40f4-8c89-352b7350d327-kube-api-access-4n7mg\") pod \"redhat-operators-8p4mb\" (UID: \"01481957-a690-40f4-8c89-352b7350d327\") " pod="openshift-marketplace/redhat-operators-8p4mb" Dec 11 08:13:45 crc kubenswrapper[4860]: I1211 08:13:45.568507 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/01481957-a690-40f4-8c89-352b7350d327-catalog-content\") pod \"redhat-operators-8p4mb\" (UID: \"01481957-a690-40f4-8c89-352b7350d327\") " pod="openshift-marketplace/redhat-operators-8p4mb" Dec 11 08:13:45 crc kubenswrapper[4860]: I1211 08:13:45.569078 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/01481957-a690-40f4-8c89-352b7350d327-catalog-content\") pod \"redhat-operators-8p4mb\" (UID: \"01481957-a690-40f4-8c89-352b7350d327\") " pod="openshift-marketplace/redhat-operators-8p4mb" Dec 11 08:13:45 crc kubenswrapper[4860]: I1211 08:13:45.569401 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/01481957-a690-40f4-8c89-352b7350d327-utilities\") pod \"redhat-operators-8p4mb\" (UID: \"01481957-a690-40f4-8c89-352b7350d327\") " pod="openshift-marketplace/redhat-operators-8p4mb" Dec 11 08:13:45 crc kubenswrapper[4860]: I1211 08:13:45.599685 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4n7mg\" (UniqueName: \"kubernetes.io/projected/01481957-a690-40f4-8c89-352b7350d327-kube-api-access-4n7mg\") pod \"redhat-operators-8p4mb\" (UID: \"01481957-a690-40f4-8c89-352b7350d327\") " pod="openshift-marketplace/redhat-operators-8p4mb" Dec 11 08:13:45 crc kubenswrapper[4860]: I1211 08:13:45.681136 4860 patch_prober.go:28] interesting pod/router-default-5444994796-n8wrd container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 11 08:13:45 crc kubenswrapper[4860]: [-]has-synced failed: reason withheld Dec 11 08:13:45 crc kubenswrapper[4860]: [+]process-running ok Dec 11 08:13:45 crc kubenswrapper[4860]: healthz check failed Dec 11 08:13:45 crc kubenswrapper[4860]: I1211 08:13:45.681216 4860 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-n8wrd" podUID="30bec2ef-4635-4766-a34c-423c2e7f471b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 11 08:13:45 crc kubenswrapper[4860]: I1211 08:13:45.759565 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8p4mb" Dec 11 08:13:45 crc kubenswrapper[4860]: I1211 08:13:45.792621 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-mgfr4"] Dec 11 08:13:45 crc kubenswrapper[4860]: I1211 08:13:45.794007 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mgfr4" Dec 11 08:13:45 crc kubenswrapper[4860]: I1211 08:13:45.798252 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mgfr4"] Dec 11 08:13:45 crc kubenswrapper[4860]: I1211 08:13:45.801348 4860 generic.go:334] "Generic (PLEG): container finished" podID="294792dd-5104-4308-a10f-f3ab0c4504e9" containerID="1f93e20350117e6aec17bfb51b14de14dc965a6d4a235499361eed82a2fe74d7" exitCode=0 Dec 11 08:13:45 crc kubenswrapper[4860]: I1211 08:13:45.801408 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dzg4x" event={"ID":"294792dd-5104-4308-a10f-f3ab0c4504e9","Type":"ContainerDied","Data":"1f93e20350117e6aec17bfb51b14de14dc965a6d4a235499361eed82a2fe74d7"} Dec 11 08:13:45 crc kubenswrapper[4860]: I1211 08:13:45.801433 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dzg4x" event={"ID":"294792dd-5104-4308-a10f-f3ab0c4504e9","Type":"ContainerStarted","Data":"c989cbc39fe1f5dd125136d91c8ece373ef71dc48fb73cff33382e4d60655abc"} Dec 11 08:13:45 crc kubenswrapper[4860]: I1211 08:13:45.818581 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-rhl5v" Dec 11 08:13:45 crc kubenswrapper[4860]: I1211 08:13:45.818624 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-rhl5v" Dec 11 08:13:45 crc kubenswrapper[4860]: I1211 08:13:45.818927 4860 generic.go:334] "Generic (PLEG): container finished" podID="8e052d7c-59d1-43a4-ac7a-e9daf2a88c65" containerID="3d49cb2b0cb5b10ea530e20f72a5bce70bd0b2bf06405ab7b75dc3260495cba3" exitCode=0 Dec 11 08:13:45 crc kubenswrapper[4860]: I1211 08:13:45.818989 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dcccn" event={"ID":"8e052d7c-59d1-43a4-ac7a-e9daf2a88c65","Type":"ContainerDied","Data":"3d49cb2b0cb5b10ea530e20f72a5bce70bd0b2bf06405ab7b75dc3260495cba3"} Dec 11 08:13:45 crc kubenswrapper[4860]: I1211 08:13:45.819013 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dcccn" event={"ID":"8e052d7c-59d1-43a4-ac7a-e9daf2a88c65","Type":"ContainerStarted","Data":"e978765edc94f035346011ecb76709f38a004e2a46dfcfa9c3b24c676dbdcbad"} Dec 11 08:13:45 crc kubenswrapper[4860]: I1211 08:13:45.824008 4860 patch_prober.go:28] interesting pod/console-f9d7485db-rhl5v container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.11:8443/health\": dial tcp 10.217.0.11:8443: connect: connection refused" start-of-body= Dec 11 08:13:45 crc kubenswrapper[4860]: I1211 08:13:45.824052 4860 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-rhl5v" podUID="d7d1536c-ed04-4199-b839-b94e9482a63c" containerName="console" probeResult="failure" output="Get \"https://10.217.0.11:8443/health\": dial tcp 10.217.0.11:8443: connect: connection refused" Dec 11 08:13:45 crc kubenswrapper[4860]: I1211 08:13:45.842866 4860 generic.go:334] "Generic (PLEG): container finished" podID="50e857f4-cdd5-4fb0-8b5d-812fdca09adf" containerID="41566f2a0973cb69bf97fff9c06bf6e664b534704ec291f1c50e0da2aa931445" exitCode=0 Dec 11 08:13:45 crc kubenswrapper[4860]: I1211 08:13:45.842934 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"50e857f4-cdd5-4fb0-8b5d-812fdca09adf","Type":"ContainerDied","Data":"41566f2a0973cb69bf97fff9c06bf6e664b534704ec291f1c50e0da2aa931445"} Dec 11 08:13:45 crc kubenswrapper[4860]: I1211 08:13:45.848155 4860 generic.go:334] "Generic (PLEG): container finished" podID="f9b23a65-98b8-4d83-a71f-2316fc1d46a0" containerID="e7891ca8b7cb19f5e3732a1f072123261d58a224b8793c6a7c0923fe1959b620" exitCode=0 Dec 11 08:13:45 crc kubenswrapper[4860]: I1211 08:13:45.848687 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p6f75" event={"ID":"f9b23a65-98b8-4d83-a71f-2316fc1d46a0","Type":"ContainerDied","Data":"e7891ca8b7cb19f5e3732a1f072123261d58a224b8793c6a7c0923fe1959b620"} Dec 11 08:13:45 crc kubenswrapper[4860]: I1211 08:13:45.859443 4860 generic.go:334] "Generic (PLEG): container finished" podID="0f6c4496-7751-49db-adc2-18099589c708" containerID="f5f79575a837d49cfff4df714dbab8b7364ce4d99d9446a09a0108cb3027e312" exitCode=0 Dec 11 08:13:45 crc kubenswrapper[4860]: I1211 08:13:45.859573 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29424000-nf7zl" event={"ID":"0f6c4496-7751-49db-adc2-18099589c708","Type":"ContainerDied","Data":"f5f79575a837d49cfff4df714dbab8b7364ce4d99d9446a09a0108cb3027e312"} Dec 11 08:13:45 crc kubenswrapper[4860]: I1211 08:13:45.870046 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-zf572" event={"ID":"38005ee2-6635-498c-b156-ac57b4f2f4d9","Type":"ContainerStarted","Data":"027d8ed039343e3ce7d35642642a40ab245c222261386a8d009a083692ad1f2b"} Dec 11 08:13:45 crc kubenswrapper[4860]: I1211 08:13:45.870101 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-zf572" Dec 11 08:13:45 crc kubenswrapper[4860]: I1211 08:13:45.879332 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fnfjs\" (UniqueName: \"kubernetes.io/projected/30107273-766a-4575-8cb1-3de387248437-kube-api-access-fnfjs\") pod \"redhat-operators-mgfr4\" (UID: \"30107273-766a-4575-8cb1-3de387248437\") " pod="openshift-marketplace/redhat-operators-mgfr4" Dec 11 08:13:45 crc kubenswrapper[4860]: I1211 08:13:45.879421 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30107273-766a-4575-8cb1-3de387248437-catalog-content\") pod \"redhat-operators-mgfr4\" (UID: \"30107273-766a-4575-8cb1-3de387248437\") " pod="openshift-marketplace/redhat-operators-mgfr4" Dec 11 08:13:45 crc kubenswrapper[4860]: I1211 08:13:45.879542 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30107273-766a-4575-8cb1-3de387248437-utilities\") pod \"redhat-operators-mgfr4\" (UID: \"30107273-766a-4575-8cb1-3de387248437\") " pod="openshift-marketplace/redhat-operators-mgfr4" Dec 11 08:13:45 crc kubenswrapper[4860]: I1211 08:13:45.935042 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-zf572" podStartSLOduration=137.935024906 podStartE2EDuration="2m17.935024906s" podCreationTimestamp="2025-12-11 08:11:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:13:45.932909543 +0000 UTC m=+158.661428598" watchObservedRunningTime="2025-12-11 08:13:45.935024906 +0000 UTC m=+158.663543961" Dec 11 08:13:45 crc kubenswrapper[4860]: I1211 08:13:45.981350 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30107273-766a-4575-8cb1-3de387248437-utilities\") pod \"redhat-operators-mgfr4\" (UID: \"30107273-766a-4575-8cb1-3de387248437\") " pod="openshift-marketplace/redhat-operators-mgfr4" Dec 11 08:13:45 crc kubenswrapper[4860]: I1211 08:13:45.981443 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fnfjs\" (UniqueName: \"kubernetes.io/projected/30107273-766a-4575-8cb1-3de387248437-kube-api-access-fnfjs\") pod \"redhat-operators-mgfr4\" (UID: \"30107273-766a-4575-8cb1-3de387248437\") " pod="openshift-marketplace/redhat-operators-mgfr4" Dec 11 08:13:45 crc kubenswrapper[4860]: I1211 08:13:45.981519 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30107273-766a-4575-8cb1-3de387248437-catalog-content\") pod \"redhat-operators-mgfr4\" (UID: \"30107273-766a-4575-8cb1-3de387248437\") " pod="openshift-marketplace/redhat-operators-mgfr4" Dec 11 08:13:45 crc kubenswrapper[4860]: I1211 08:13:45.983116 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30107273-766a-4575-8cb1-3de387248437-catalog-content\") pod \"redhat-operators-mgfr4\" (UID: \"30107273-766a-4575-8cb1-3de387248437\") " pod="openshift-marketplace/redhat-operators-mgfr4" Dec 11 08:13:45 crc kubenswrapper[4860]: I1211 08:13:45.983407 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30107273-766a-4575-8cb1-3de387248437-utilities\") pod \"redhat-operators-mgfr4\" (UID: \"30107273-766a-4575-8cb1-3de387248437\") " pod="openshift-marketplace/redhat-operators-mgfr4" Dec 11 08:13:46 crc kubenswrapper[4860]: I1211 08:13:46.009763 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fnfjs\" (UniqueName: \"kubernetes.io/projected/30107273-766a-4575-8cb1-3de387248437-kube-api-access-fnfjs\") pod \"redhat-operators-mgfr4\" (UID: \"30107273-766a-4575-8cb1-3de387248437\") " pod="openshift-marketplace/redhat-operators-mgfr4" Dec 11 08:13:46 crc kubenswrapper[4860]: I1211 08:13:46.047487 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-jxft9" Dec 11 08:13:46 crc kubenswrapper[4860]: I1211 08:13:46.047526 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-jxft9" Dec 11 08:13:46 crc kubenswrapper[4860]: I1211 08:13:46.111752 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-jxft9" Dec 11 08:13:46 crc kubenswrapper[4860]: I1211 08:13:46.154288 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mgfr4" Dec 11 08:13:46 crc kubenswrapper[4860]: I1211 08:13:46.390958 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8p4mb"] Dec 11 08:13:46 crc kubenswrapper[4860]: W1211 08:13:46.397546 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod01481957_a690_40f4_8c89_352b7350d327.slice/crio-09109e706dbeaf7ff53ad518750d66808c1f71b08f2a704acbd51462026a8a49 WatchSource:0}: Error finding container 09109e706dbeaf7ff53ad518750d66808c1f71b08f2a704acbd51462026a8a49: Status 404 returned error can't find the container with id 09109e706dbeaf7ff53ad518750d66808c1f71b08f2a704acbd51462026a8a49 Dec 11 08:13:46 crc kubenswrapper[4860]: I1211 08:13:46.627823 4860 patch_prober.go:28] interesting pod/downloads-7954f5f757-mj2tr container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.25:8080/\": dial tcp 10.217.0.25:8080: connect: connection refused" start-of-body= Dec 11 08:13:46 crc kubenswrapper[4860]: I1211 08:13:46.628160 4860 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-mj2tr" podUID="cf1bd6cb-c047-423c-9d36-7246a6108844" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.25:8080/\": dial tcp 10.217.0.25:8080: connect: connection refused" Dec 11 08:13:46 crc kubenswrapper[4860]: I1211 08:13:46.628105 4860 patch_prober.go:28] interesting pod/downloads-7954f5f757-mj2tr container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.25:8080/\": dial tcp 10.217.0.25:8080: connect: connection refused" start-of-body= Dec 11 08:13:46 crc kubenswrapper[4860]: I1211 08:13:46.628468 4860 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-mj2tr" podUID="cf1bd6cb-c047-423c-9d36-7246a6108844" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.25:8080/\": dial tcp 10.217.0.25:8080: connect: connection refused" Dec 11 08:13:46 crc kubenswrapper[4860]: I1211 08:13:46.671073 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-n8wrd" Dec 11 08:13:46 crc kubenswrapper[4860]: I1211 08:13:46.675827 4860 patch_prober.go:28] interesting pod/router-default-5444994796-n8wrd container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 11 08:13:46 crc kubenswrapper[4860]: [-]has-synced failed: reason withheld Dec 11 08:13:46 crc kubenswrapper[4860]: [+]process-running ok Dec 11 08:13:46 crc kubenswrapper[4860]: healthz check failed Dec 11 08:13:46 crc kubenswrapper[4860]: I1211 08:13:46.675888 4860 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-n8wrd" podUID="30bec2ef-4635-4766-a34c-423c2e7f471b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 11 08:13:46 crc kubenswrapper[4860]: I1211 08:13:46.689846 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-ttrm2" Dec 11 08:13:46 crc kubenswrapper[4860]: I1211 08:13:46.738069 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mgfr4"] Dec 11 08:13:46 crc kubenswrapper[4860]: W1211 08:13:46.842243 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod30107273_766a_4575_8cb1_3de387248437.slice/crio-df3f1a03bb7d67e0dd6af440ad08068302ae497e58fefd9296b73bfb75c129cc WatchSource:0}: Error finding container df3f1a03bb7d67e0dd6af440ad08068302ae497e58fefd9296b73bfb75c129cc: Status 404 returned error can't find the container with id df3f1a03bb7d67e0dd6af440ad08068302ae497e58fefd9296b73bfb75c129cc Dec 11 08:13:46 crc kubenswrapper[4860]: I1211 08:13:46.910226 4860 generic.go:334] "Generic (PLEG): container finished" podID="01481957-a690-40f4-8c89-352b7350d327" containerID="a9587c21a079451102e0829c093f0967af046ffa260e000b16b6bcfc2d8a3dbe" exitCode=0 Dec 11 08:13:46 crc kubenswrapper[4860]: I1211 08:13:46.910394 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8p4mb" event={"ID":"01481957-a690-40f4-8c89-352b7350d327","Type":"ContainerDied","Data":"a9587c21a079451102e0829c093f0967af046ffa260e000b16b6bcfc2d8a3dbe"} Dec 11 08:13:46 crc kubenswrapper[4860]: I1211 08:13:46.910432 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8p4mb" event={"ID":"01481957-a690-40f4-8c89-352b7350d327","Type":"ContainerStarted","Data":"09109e706dbeaf7ff53ad518750d66808c1f71b08f2a704acbd51462026a8a49"} Dec 11 08:13:46 crc kubenswrapper[4860]: I1211 08:13:46.915926 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-jxft9" Dec 11 08:13:47 crc kubenswrapper[4860]: I1211 08:13:47.674034 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29424000-nf7zl" Dec 11 08:13:47 crc kubenswrapper[4860]: I1211 08:13:47.678262 4860 patch_prober.go:28] interesting pod/router-default-5444994796-n8wrd container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 11 08:13:47 crc kubenswrapper[4860]: [-]has-synced failed: reason withheld Dec 11 08:13:47 crc kubenswrapper[4860]: [+]process-running ok Dec 11 08:13:47 crc kubenswrapper[4860]: healthz check failed Dec 11 08:13:47 crc kubenswrapper[4860]: I1211 08:13:47.678290 4860 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-n8wrd" podUID="30bec2ef-4635-4766-a34c-423c2e7f471b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 11 08:13:47 crc kubenswrapper[4860]: I1211 08:13:47.734798 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 11 08:13:47 crc kubenswrapper[4860]: I1211 08:13:47.761100 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m8497\" (UniqueName: \"kubernetes.io/projected/0f6c4496-7751-49db-adc2-18099589c708-kube-api-access-m8497\") pod \"0f6c4496-7751-49db-adc2-18099589c708\" (UID: \"0f6c4496-7751-49db-adc2-18099589c708\") " Dec 11 08:13:47 crc kubenswrapper[4860]: I1211 08:13:47.761450 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0f6c4496-7751-49db-adc2-18099589c708-config-volume\") pod \"0f6c4496-7751-49db-adc2-18099589c708\" (UID: \"0f6c4496-7751-49db-adc2-18099589c708\") " Dec 11 08:13:47 crc kubenswrapper[4860]: I1211 08:13:47.761538 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0f6c4496-7751-49db-adc2-18099589c708-secret-volume\") pod \"0f6c4496-7751-49db-adc2-18099589c708\" (UID: \"0f6c4496-7751-49db-adc2-18099589c708\") " Dec 11 08:13:47 crc kubenswrapper[4860]: I1211 08:13:47.765231 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0f6c4496-7751-49db-adc2-18099589c708-config-volume" (OuterVolumeSpecName: "config-volume") pod "0f6c4496-7751-49db-adc2-18099589c708" (UID: "0f6c4496-7751-49db-adc2-18099589c708"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:13:47 crc kubenswrapper[4860]: I1211 08:13:47.791814 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f6c4496-7751-49db-adc2-18099589c708-kube-api-access-m8497" (OuterVolumeSpecName: "kube-api-access-m8497") pod "0f6c4496-7751-49db-adc2-18099589c708" (UID: "0f6c4496-7751-49db-adc2-18099589c708"). InnerVolumeSpecName "kube-api-access-m8497". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:13:47 crc kubenswrapper[4860]: I1211 08:13:47.792001 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f6c4496-7751-49db-adc2-18099589c708-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "0f6c4496-7751-49db-adc2-18099589c708" (UID: "0f6c4496-7751-49db-adc2-18099589c708"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:13:47 crc kubenswrapper[4860]: I1211 08:13:47.863286 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/50e857f4-cdd5-4fb0-8b5d-812fdca09adf-kube-api-access\") pod \"50e857f4-cdd5-4fb0-8b5d-812fdca09adf\" (UID: \"50e857f4-cdd5-4fb0-8b5d-812fdca09adf\") " Dec 11 08:13:47 crc kubenswrapper[4860]: I1211 08:13:47.863381 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/50e857f4-cdd5-4fb0-8b5d-812fdca09adf-kubelet-dir\") pod \"50e857f4-cdd5-4fb0-8b5d-812fdca09adf\" (UID: \"50e857f4-cdd5-4fb0-8b5d-812fdca09adf\") " Dec 11 08:13:47 crc kubenswrapper[4860]: I1211 08:13:47.863536 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/50e857f4-cdd5-4fb0-8b5d-812fdca09adf-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "50e857f4-cdd5-4fb0-8b5d-812fdca09adf" (UID: "50e857f4-cdd5-4fb0-8b5d-812fdca09adf"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 08:13:47 crc kubenswrapper[4860]: I1211 08:13:47.863779 4860 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/50e857f4-cdd5-4fb0-8b5d-812fdca09adf-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 11 08:13:47 crc kubenswrapper[4860]: I1211 08:13:47.863794 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m8497\" (UniqueName: \"kubernetes.io/projected/0f6c4496-7751-49db-adc2-18099589c708-kube-api-access-m8497\") on node \"crc\" DevicePath \"\"" Dec 11 08:13:47 crc kubenswrapper[4860]: I1211 08:13:47.863805 4860 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0f6c4496-7751-49db-adc2-18099589c708-config-volume\") on node \"crc\" DevicePath \"\"" Dec 11 08:13:47 crc kubenswrapper[4860]: I1211 08:13:47.863813 4860 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0f6c4496-7751-49db-adc2-18099589c708-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 11 08:13:47 crc kubenswrapper[4860]: I1211 08:13:47.884896 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/50e857f4-cdd5-4fb0-8b5d-812fdca09adf-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "50e857f4-cdd5-4fb0-8b5d-812fdca09adf" (UID: "50e857f4-cdd5-4fb0-8b5d-812fdca09adf"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:13:47 crc kubenswrapper[4860]: I1211 08:13:47.917987 4860 generic.go:334] "Generic (PLEG): container finished" podID="30107273-766a-4575-8cb1-3de387248437" containerID="5fd620a265f48ef3dbf7fbe3d0076723092d08f33ac19e073d2931edfd8efaa9" exitCode=0 Dec 11 08:13:47 crc kubenswrapper[4860]: I1211 08:13:47.918067 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mgfr4" event={"ID":"30107273-766a-4575-8cb1-3de387248437","Type":"ContainerDied","Data":"5fd620a265f48ef3dbf7fbe3d0076723092d08f33ac19e073d2931edfd8efaa9"} Dec 11 08:13:47 crc kubenswrapper[4860]: I1211 08:13:47.918093 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mgfr4" event={"ID":"30107273-766a-4575-8cb1-3de387248437","Type":"ContainerStarted","Data":"df3f1a03bb7d67e0dd6af440ad08068302ae497e58fefd9296b73bfb75c129cc"} Dec 11 08:13:47 crc kubenswrapper[4860]: I1211 08:13:47.926478 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"50e857f4-cdd5-4fb0-8b5d-812fdca09adf","Type":"ContainerDied","Data":"7ae2db137ba0e7547bfed907a1edcf26510cc342229d5f1897bcf3c6fb4da540"} Dec 11 08:13:47 crc kubenswrapper[4860]: I1211 08:13:47.926511 4860 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7ae2db137ba0e7547bfed907a1edcf26510cc342229d5f1897bcf3c6fb4da540" Dec 11 08:13:47 crc kubenswrapper[4860]: I1211 08:13:47.926562 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 11 08:13:47 crc kubenswrapper[4860]: I1211 08:13:47.972406 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/50e857f4-cdd5-4fb0-8b5d-812fdca09adf-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 11 08:13:47 crc kubenswrapper[4860]: I1211 08:13:47.983343 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29424000-nf7zl" event={"ID":"0f6c4496-7751-49db-adc2-18099589c708","Type":"ContainerDied","Data":"f7fdb07c94b081ce80b3ef7933886a6efe4a77ca6ad75126daf037599561e4ed"} Dec 11 08:13:47 crc kubenswrapper[4860]: I1211 08:13:47.983426 4860 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f7fdb07c94b081ce80b3ef7933886a6efe4a77ca6ad75126daf037599561e4ed" Dec 11 08:13:47 crc kubenswrapper[4860]: I1211 08:13:47.983390 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29424000-nf7zl" Dec 11 08:13:48 crc kubenswrapper[4860]: I1211 08:13:48.678269 4860 patch_prober.go:28] interesting pod/router-default-5444994796-n8wrd container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 11 08:13:48 crc kubenswrapper[4860]: [-]has-synced failed: reason withheld Dec 11 08:13:48 crc kubenswrapper[4860]: [+]process-running ok Dec 11 08:13:48 crc kubenswrapper[4860]: healthz check failed Dec 11 08:13:48 crc kubenswrapper[4860]: I1211 08:13:48.678335 4860 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-n8wrd" podUID="30bec2ef-4635-4766-a34c-423c2e7f471b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 11 08:13:49 crc kubenswrapper[4860]: I1211 08:13:49.673284 4860 patch_prober.go:28] interesting pod/router-default-5444994796-n8wrd container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 11 08:13:49 crc kubenswrapper[4860]: [-]has-synced failed: reason withheld Dec 11 08:13:49 crc kubenswrapper[4860]: [+]process-running ok Dec 11 08:13:49 crc kubenswrapper[4860]: healthz check failed Dec 11 08:13:49 crc kubenswrapper[4860]: I1211 08:13:49.673562 4860 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-n8wrd" podUID="30bec2ef-4635-4766-a34c-423c2e7f471b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 11 08:13:50 crc kubenswrapper[4860]: I1211 08:13:50.169510 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 11 08:13:50 crc kubenswrapper[4860]: E1211 08:13:50.169764 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50e857f4-cdd5-4fb0-8b5d-812fdca09adf" containerName="pruner" Dec 11 08:13:50 crc kubenswrapper[4860]: I1211 08:13:50.169775 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="50e857f4-cdd5-4fb0-8b5d-812fdca09adf" containerName="pruner" Dec 11 08:13:50 crc kubenswrapper[4860]: E1211 08:13:50.169786 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f6c4496-7751-49db-adc2-18099589c708" containerName="collect-profiles" Dec 11 08:13:50 crc kubenswrapper[4860]: I1211 08:13:50.169793 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f6c4496-7751-49db-adc2-18099589c708" containerName="collect-profiles" Dec 11 08:13:50 crc kubenswrapper[4860]: I1211 08:13:50.169901 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="50e857f4-cdd5-4fb0-8b5d-812fdca09adf" containerName="pruner" Dec 11 08:13:50 crc kubenswrapper[4860]: I1211 08:13:50.169911 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f6c4496-7751-49db-adc2-18099589c708" containerName="collect-profiles" Dec 11 08:13:50 crc kubenswrapper[4860]: I1211 08:13:50.170320 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 11 08:13:50 crc kubenswrapper[4860]: I1211 08:13:50.180834 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 11 08:13:50 crc kubenswrapper[4860]: I1211 08:13:50.182294 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 11 08:13:50 crc kubenswrapper[4860]: I1211 08:13:50.191329 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 11 08:13:50 crc kubenswrapper[4860]: I1211 08:13:50.241899 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1d2c560f-1648-48c1-847a-54da5b58a777-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"1d2c560f-1648-48c1-847a-54da5b58a777\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 11 08:13:50 crc kubenswrapper[4860]: I1211 08:13:50.242016 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1d2c560f-1648-48c1-847a-54da5b58a777-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"1d2c560f-1648-48c1-847a-54da5b58a777\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 11 08:13:50 crc kubenswrapper[4860]: I1211 08:13:50.345403 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1d2c560f-1648-48c1-847a-54da5b58a777-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"1d2c560f-1648-48c1-847a-54da5b58a777\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 11 08:13:50 crc kubenswrapper[4860]: I1211 08:13:50.345713 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1d2c560f-1648-48c1-847a-54da5b58a777-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"1d2c560f-1648-48c1-847a-54da5b58a777\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 11 08:13:50 crc kubenswrapper[4860]: I1211 08:13:50.345925 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1d2c560f-1648-48c1-847a-54da5b58a777-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"1d2c560f-1648-48c1-847a-54da5b58a777\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 11 08:13:50 crc kubenswrapper[4860]: I1211 08:13:50.394526 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1d2c560f-1648-48c1-847a-54da5b58a777-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"1d2c560f-1648-48c1-847a-54da5b58a777\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 11 08:13:50 crc kubenswrapper[4860]: I1211 08:13:50.521292 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 11 08:13:50 crc kubenswrapper[4860]: I1211 08:13:50.549426 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5cfd640a-f3d2-4b28-87e1-5f49f6341e21-metrics-certs\") pod \"network-metrics-daemon-gjhlj\" (UID: \"5cfd640a-f3d2-4b28-87e1-5f49f6341e21\") " pod="openshift-multus/network-metrics-daemon-gjhlj" Dec 11 08:13:50 crc kubenswrapper[4860]: I1211 08:13:50.560363 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5cfd640a-f3d2-4b28-87e1-5f49f6341e21-metrics-certs\") pod \"network-metrics-daemon-gjhlj\" (UID: \"5cfd640a-f3d2-4b28-87e1-5f49f6341e21\") " pod="openshift-multus/network-metrics-daemon-gjhlj" Dec 11 08:13:50 crc kubenswrapper[4860]: I1211 08:13:50.602789 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gjhlj" Dec 11 08:13:50 crc kubenswrapper[4860]: I1211 08:13:50.676550 4860 patch_prober.go:28] interesting pod/router-default-5444994796-n8wrd container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 11 08:13:50 crc kubenswrapper[4860]: [-]has-synced failed: reason withheld Dec 11 08:13:50 crc kubenswrapper[4860]: [+]process-running ok Dec 11 08:13:50 crc kubenswrapper[4860]: healthz check failed Dec 11 08:13:50 crc kubenswrapper[4860]: I1211 08:13:50.676621 4860 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-n8wrd" podUID="30bec2ef-4635-4766-a34c-423c2e7f471b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 11 08:13:51 crc kubenswrapper[4860]: I1211 08:13:51.365673 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-gjhlj"] Dec 11 08:13:51 crc kubenswrapper[4860]: W1211 08:13:51.383637 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5cfd640a_f3d2_4b28_87e1_5f49f6341e21.slice/crio-aed46246c42defc8c4e6dba58af9d932518c090af349c16b6aca13d0a6a1086e WatchSource:0}: Error finding container aed46246c42defc8c4e6dba58af9d932518c090af349c16b6aca13d0a6a1086e: Status 404 returned error can't find the container with id aed46246c42defc8c4e6dba58af9d932518c090af349c16b6aca13d0a6a1086e Dec 11 08:13:51 crc kubenswrapper[4860]: I1211 08:13:51.446997 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 11 08:13:51 crc kubenswrapper[4860]: I1211 08:13:51.674303 4860 patch_prober.go:28] interesting pod/router-default-5444994796-n8wrd container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 11 08:13:51 crc kubenswrapper[4860]: [-]has-synced failed: reason withheld Dec 11 08:13:51 crc kubenswrapper[4860]: [+]process-running ok Dec 11 08:13:51 crc kubenswrapper[4860]: healthz check failed Dec 11 08:13:51 crc kubenswrapper[4860]: I1211 08:13:51.674392 4860 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-n8wrd" podUID="30bec2ef-4635-4766-a34c-423c2e7f471b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 11 08:13:52 crc kubenswrapper[4860]: I1211 08:13:52.022837 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-gjhlj" event={"ID":"5cfd640a-f3d2-4b28-87e1-5f49f6341e21","Type":"ContainerStarted","Data":"aed46246c42defc8c4e6dba58af9d932518c090af349c16b6aca13d0a6a1086e"} Dec 11 08:13:52 crc kubenswrapper[4860]: W1211 08:13:52.066515 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod1d2c560f_1648_48c1_847a_54da5b58a777.slice/crio-e674d842465c9047e38deabda077a2f11ae7e276fa19992e9e7bf42445da356c WatchSource:0}: Error finding container e674d842465c9047e38deabda077a2f11ae7e276fa19992e9e7bf42445da356c: Status 404 returned error can't find the container with id e674d842465c9047e38deabda077a2f11ae7e276fa19992e9e7bf42445da356c Dec 11 08:13:52 crc kubenswrapper[4860]: I1211 08:13:52.133403 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-sbcxd" Dec 11 08:13:52 crc kubenswrapper[4860]: I1211 08:13:52.676305 4860 patch_prober.go:28] interesting pod/router-default-5444994796-n8wrd container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 11 08:13:52 crc kubenswrapper[4860]: [-]has-synced failed: reason withheld Dec 11 08:13:52 crc kubenswrapper[4860]: [+]process-running ok Dec 11 08:13:52 crc kubenswrapper[4860]: healthz check failed Dec 11 08:13:52 crc kubenswrapper[4860]: I1211 08:13:52.676495 4860 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-n8wrd" podUID="30bec2ef-4635-4766-a34c-423c2e7f471b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 11 08:13:53 crc kubenswrapper[4860]: I1211 08:13:53.030448 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-gjhlj" event={"ID":"5cfd640a-f3d2-4b28-87e1-5f49f6341e21","Type":"ContainerStarted","Data":"2ba5ae2099611871c116e4b7eb1338635e09e9abce17a255826ff011d9a78fc6"} Dec 11 08:13:53 crc kubenswrapper[4860]: I1211 08:13:53.035149 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"1d2c560f-1648-48c1-847a-54da5b58a777","Type":"ContainerStarted","Data":"e674d842465c9047e38deabda077a2f11ae7e276fa19992e9e7bf42445da356c"} Dec 11 08:13:53 crc kubenswrapper[4860]: I1211 08:13:53.678177 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-n8wrd" Dec 11 08:13:53 crc kubenswrapper[4860]: I1211 08:13:53.684161 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-n8wrd" Dec 11 08:13:54 crc kubenswrapper[4860]: I1211 08:13:54.147330 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-gjhlj" event={"ID":"5cfd640a-f3d2-4b28-87e1-5f49f6341e21","Type":"ContainerStarted","Data":"10543b8505496956296fe2883b8e7149bda681d432b0abeadaa8e4333d1b5c61"} Dec 11 08:13:54 crc kubenswrapper[4860]: I1211 08:13:54.157346 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"1d2c560f-1648-48c1-847a-54da5b58a777","Type":"ContainerStarted","Data":"5440abf9fbb463cf5eb7392c41c0727da56660669f5e73c3df9318b056f4d5bf"} Dec 11 08:13:55 crc kubenswrapper[4860]: I1211 08:13:55.176729 4860 generic.go:334] "Generic (PLEG): container finished" podID="1d2c560f-1648-48c1-847a-54da5b58a777" containerID="5440abf9fbb463cf5eb7392c41c0727da56660669f5e73c3df9318b056f4d5bf" exitCode=0 Dec 11 08:13:55 crc kubenswrapper[4860]: I1211 08:13:55.177565 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"1d2c560f-1648-48c1-847a-54da5b58a777","Type":"ContainerDied","Data":"5440abf9fbb463cf5eb7392c41c0727da56660669f5e73c3df9318b056f4d5bf"} Dec 11 08:13:55 crc kubenswrapper[4860]: I1211 08:13:55.210563 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-gjhlj" podStartSLOduration=147.210540035 podStartE2EDuration="2m27.210540035s" podCreationTimestamp="2025-12-11 08:11:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:13:55.198907737 +0000 UTC m=+167.927426792" watchObservedRunningTime="2025-12-11 08:13:55.210540035 +0000 UTC m=+167.939059100" Dec 11 08:13:55 crc kubenswrapper[4860]: I1211 08:13:55.819164 4860 patch_prober.go:28] interesting pod/console-f9d7485db-rhl5v container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.11:8443/health\": dial tcp 10.217.0.11:8443: connect: connection refused" start-of-body= Dec 11 08:13:55 crc kubenswrapper[4860]: I1211 08:13:55.819228 4860 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-rhl5v" podUID="d7d1536c-ed04-4199-b839-b94e9482a63c" containerName="console" probeResult="failure" output="Get \"https://10.217.0.11:8443/health\": dial tcp 10.217.0.11:8443: connect: connection refused" Dec 11 08:13:56 crc kubenswrapper[4860]: I1211 08:13:56.628401 4860 patch_prober.go:28] interesting pod/downloads-7954f5f757-mj2tr container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.25:8080/\": dial tcp 10.217.0.25:8080: connect: connection refused" start-of-body= Dec 11 08:13:56 crc kubenswrapper[4860]: I1211 08:13:56.628728 4860 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-mj2tr" podUID="cf1bd6cb-c047-423c-9d36-7246a6108844" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.25:8080/\": dial tcp 10.217.0.25:8080: connect: connection refused" Dec 11 08:13:56 crc kubenswrapper[4860]: I1211 08:13:56.628481 4860 patch_prober.go:28] interesting pod/downloads-7954f5f757-mj2tr container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.25:8080/\": dial tcp 10.217.0.25:8080: connect: connection refused" start-of-body= Dec 11 08:13:56 crc kubenswrapper[4860]: I1211 08:13:56.628840 4860 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-mj2tr" podUID="cf1bd6cb-c047-423c-9d36-7246a6108844" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.25:8080/\": dial tcp 10.217.0.25:8080: connect: connection refused" Dec 11 08:14:03 crc kubenswrapper[4860]: I1211 08:14:03.470866 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-zf572" Dec 11 08:14:05 crc kubenswrapper[4860]: I1211 08:14:05.862496 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-rhl5v" Dec 11 08:14:05 crc kubenswrapper[4860]: I1211 08:14:05.871714 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-rhl5v" Dec 11 08:14:06 crc kubenswrapper[4860]: I1211 08:14:06.633477 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-mj2tr" Dec 11 08:14:08 crc kubenswrapper[4860]: I1211 08:14:08.795400 4860 patch_prober.go:28] interesting pod/machine-config-daemon-99qgp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 08:14:08 crc kubenswrapper[4860]: I1211 08:14:08.795682 4860 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 08:14:13 crc kubenswrapper[4860]: I1211 08:14:13.534128 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 11 08:14:13 crc kubenswrapper[4860]: I1211 08:14:13.586482 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1d2c560f-1648-48c1-847a-54da5b58a777-kubelet-dir\") pod \"1d2c560f-1648-48c1-847a-54da5b58a777\" (UID: \"1d2c560f-1648-48c1-847a-54da5b58a777\") " Dec 11 08:14:13 crc kubenswrapper[4860]: I1211 08:14:13.586567 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1d2c560f-1648-48c1-847a-54da5b58a777-kube-api-access\") pod \"1d2c560f-1648-48c1-847a-54da5b58a777\" (UID: \"1d2c560f-1648-48c1-847a-54da5b58a777\") " Dec 11 08:14:13 crc kubenswrapper[4860]: I1211 08:14:13.588841 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1d2c560f-1648-48c1-847a-54da5b58a777-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "1d2c560f-1648-48c1-847a-54da5b58a777" (UID: "1d2c560f-1648-48c1-847a-54da5b58a777"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 08:14:13 crc kubenswrapper[4860]: I1211 08:14:13.594860 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d2c560f-1648-48c1-847a-54da5b58a777-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1d2c560f-1648-48c1-847a-54da5b58a777" (UID: "1d2c560f-1648-48c1-847a-54da5b58a777"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:14:13 crc kubenswrapper[4860]: I1211 08:14:13.688023 4860 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1d2c560f-1648-48c1-847a-54da5b58a777-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 11 08:14:13 crc kubenswrapper[4860]: I1211 08:14:13.688103 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1d2c560f-1648-48c1-847a-54da5b58a777-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 11 08:14:14 crc kubenswrapper[4860]: I1211 08:14:14.334768 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"1d2c560f-1648-48c1-847a-54da5b58a777","Type":"ContainerDied","Data":"e674d842465c9047e38deabda077a2f11ae7e276fa19992e9e7bf42445da356c"} Dec 11 08:14:14 crc kubenswrapper[4860]: I1211 08:14:14.335111 4860 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e674d842465c9047e38deabda077a2f11ae7e276fa19992e9e7bf42445da356c" Dec 11 08:14:14 crc kubenswrapper[4860]: I1211 08:14:14.334903 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 11 08:14:14 crc kubenswrapper[4860]: I1211 08:14:14.901468 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 11 08:14:16 crc kubenswrapper[4860]: I1211 08:14:16.726456 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5zqnk" Dec 11 08:14:16 crc kubenswrapper[4860]: I1211 08:14:16.913817 4860 patch_prober.go:28] interesting pod/authentication-operator-69f744f599-j5kxr container/authentication-operator namespace/openshift-authentication-operator: Liveness probe status=failure output="Get \"https://10.217.0.13:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 11 08:14:16 crc kubenswrapper[4860]: I1211 08:14:16.913875 4860 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-authentication-operator/authentication-operator-69f744f599-j5kxr" podUID="2c04dbaf-f385-4cee-bccf-8506ecdeb680" containerName="authentication-operator" probeResult="failure" output="Get \"https://10.217.0.13:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 11 08:14:21 crc kubenswrapper[4860]: I1211 08:14:21.964875 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 11 08:14:21 crc kubenswrapper[4860]: E1211 08:14:21.965921 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d2c560f-1648-48c1-847a-54da5b58a777" containerName="pruner" Dec 11 08:14:21 crc kubenswrapper[4860]: I1211 08:14:21.965938 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d2c560f-1648-48c1-847a-54da5b58a777" containerName="pruner" Dec 11 08:14:21 crc kubenswrapper[4860]: I1211 08:14:21.966064 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d2c560f-1648-48c1-847a-54da5b58a777" containerName="pruner" Dec 11 08:14:21 crc kubenswrapper[4860]: I1211 08:14:21.966408 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 11 08:14:21 crc kubenswrapper[4860]: I1211 08:14:21.971816 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 11 08:14:21 crc kubenswrapper[4860]: I1211 08:14:21.971982 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 11 08:14:21 crc kubenswrapper[4860]: I1211 08:14:21.976447 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 11 08:14:22 crc kubenswrapper[4860]: I1211 08:14:22.008939 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/934cd13b-9314-4299-8b5b-5a8e03810781-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"934cd13b-9314-4299-8b5b-5a8e03810781\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 11 08:14:22 crc kubenswrapper[4860]: I1211 08:14:22.009026 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/934cd13b-9314-4299-8b5b-5a8e03810781-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"934cd13b-9314-4299-8b5b-5a8e03810781\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 11 08:14:22 crc kubenswrapper[4860]: I1211 08:14:22.110860 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/934cd13b-9314-4299-8b5b-5a8e03810781-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"934cd13b-9314-4299-8b5b-5a8e03810781\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 11 08:14:22 crc kubenswrapper[4860]: I1211 08:14:22.111298 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/934cd13b-9314-4299-8b5b-5a8e03810781-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"934cd13b-9314-4299-8b5b-5a8e03810781\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 11 08:14:22 crc kubenswrapper[4860]: I1211 08:14:22.111066 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/934cd13b-9314-4299-8b5b-5a8e03810781-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"934cd13b-9314-4299-8b5b-5a8e03810781\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 11 08:14:22 crc kubenswrapper[4860]: I1211 08:14:22.135802 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/934cd13b-9314-4299-8b5b-5a8e03810781-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"934cd13b-9314-4299-8b5b-5a8e03810781\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 11 08:14:22 crc kubenswrapper[4860]: I1211 08:14:22.293467 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 11 08:14:25 crc kubenswrapper[4860]: E1211 08:14:25.165417 4860 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 11 08:14:25 crc kubenswrapper[4860]: E1211 08:14:25.165829 4860 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-f92fq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-kqwgm_openshift-marketplace(093a9903-7041-410b-bfb2-c74ef741e0be): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 11 08:14:25 crc kubenswrapper[4860]: E1211 08:14:25.167050 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-kqwgm" podUID="093a9903-7041-410b-bfb2-c74ef741e0be" Dec 11 08:14:26 crc kubenswrapper[4860]: I1211 08:14:26.162531 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 11 08:14:26 crc kubenswrapper[4860]: I1211 08:14:26.163959 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 11 08:14:26 crc kubenswrapper[4860]: I1211 08:14:26.176416 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 11 08:14:26 crc kubenswrapper[4860]: I1211 08:14:26.285131 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/3bea0969-d146-4f96-8f02-b95bfe17aac3-var-lock\") pod \"installer-9-crc\" (UID: \"3bea0969-d146-4f96-8f02-b95bfe17aac3\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 11 08:14:26 crc kubenswrapper[4860]: I1211 08:14:26.285193 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3bea0969-d146-4f96-8f02-b95bfe17aac3-kubelet-dir\") pod \"installer-9-crc\" (UID: \"3bea0969-d146-4f96-8f02-b95bfe17aac3\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 11 08:14:26 crc kubenswrapper[4860]: I1211 08:14:26.285458 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3bea0969-d146-4f96-8f02-b95bfe17aac3-kube-api-access\") pod \"installer-9-crc\" (UID: \"3bea0969-d146-4f96-8f02-b95bfe17aac3\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 11 08:14:26 crc kubenswrapper[4860]: I1211 08:14:26.386448 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/3bea0969-d146-4f96-8f02-b95bfe17aac3-var-lock\") pod \"installer-9-crc\" (UID: \"3bea0969-d146-4f96-8f02-b95bfe17aac3\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 11 08:14:26 crc kubenswrapper[4860]: I1211 08:14:26.386497 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3bea0969-d146-4f96-8f02-b95bfe17aac3-kubelet-dir\") pod \"installer-9-crc\" (UID: \"3bea0969-d146-4f96-8f02-b95bfe17aac3\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 11 08:14:26 crc kubenswrapper[4860]: I1211 08:14:26.386546 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3bea0969-d146-4f96-8f02-b95bfe17aac3-kube-api-access\") pod \"installer-9-crc\" (UID: \"3bea0969-d146-4f96-8f02-b95bfe17aac3\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 11 08:14:26 crc kubenswrapper[4860]: I1211 08:14:26.386858 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/3bea0969-d146-4f96-8f02-b95bfe17aac3-var-lock\") pod \"installer-9-crc\" (UID: \"3bea0969-d146-4f96-8f02-b95bfe17aac3\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 11 08:14:26 crc kubenswrapper[4860]: I1211 08:14:26.386879 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3bea0969-d146-4f96-8f02-b95bfe17aac3-kubelet-dir\") pod \"installer-9-crc\" (UID: \"3bea0969-d146-4f96-8f02-b95bfe17aac3\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 11 08:14:26 crc kubenswrapper[4860]: I1211 08:14:26.401915 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3bea0969-d146-4f96-8f02-b95bfe17aac3-kube-api-access\") pod \"installer-9-crc\" (UID: \"3bea0969-d146-4f96-8f02-b95bfe17aac3\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 11 08:14:26 crc kubenswrapper[4860]: I1211 08:14:26.490392 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 11 08:14:27 crc kubenswrapper[4860]: E1211 08:14:27.625698 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-kqwgm" podUID="093a9903-7041-410b-bfb2-c74ef741e0be" Dec 11 08:14:29 crc kubenswrapper[4860]: E1211 08:14:29.127613 4860 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 11 08:14:29 crc kubenswrapper[4860]: E1211 08:14:29.128121 4860 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tf7hl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-pqtvk_openshift-marketplace(7ac89442-1754-45d3-b67a-aa6cc31a1235): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 11 08:14:29 crc kubenswrapper[4860]: E1211 08:14:29.129304 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-pqtvk" podUID="7ac89442-1754-45d3-b67a-aa6cc31a1235" Dec 11 08:14:29 crc kubenswrapper[4860]: E1211 08:14:29.143819 4860 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 11 08:14:29 crc kubenswrapper[4860]: E1211 08:14:29.143980 4860 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-whc24,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-p6f75_openshift-marketplace(f9b23a65-98b8-4d83-a71f-2316fc1d46a0): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 11 08:14:29 crc kubenswrapper[4860]: E1211 08:14:29.145182 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-p6f75" podUID="f9b23a65-98b8-4d83-a71f-2316fc1d46a0" Dec 11 08:14:32 crc kubenswrapper[4860]: E1211 08:14:32.138130 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-pqtvk" podUID="7ac89442-1754-45d3-b67a-aa6cc31a1235" Dec 11 08:14:32 crc kubenswrapper[4860]: E1211 08:14:32.138135 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-p6f75" podUID="f9b23a65-98b8-4d83-a71f-2316fc1d46a0" Dec 11 08:14:32 crc kubenswrapper[4860]: E1211 08:14:32.234107 4860 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 11 08:14:32 crc kubenswrapper[4860]: E1211 08:14:32.234281 4860 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4n7mg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-8p4mb_openshift-marketplace(01481957-a690-40f4-8c89-352b7350d327): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 11 08:14:32 crc kubenswrapper[4860]: E1211 08:14:32.236206 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-8p4mb" podUID="01481957-a690-40f4-8c89-352b7350d327" Dec 11 08:14:32 crc kubenswrapper[4860]: E1211 08:14:32.237490 4860 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 11 08:14:32 crc kubenswrapper[4860]: E1211 08:14:32.237752 4860 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-gf8wg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-55jtz_openshift-marketplace(b498b689-633b-4fea-9981-159362f44053): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 11 08:14:32 crc kubenswrapper[4860]: E1211 08:14:32.238948 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-55jtz" podUID="b498b689-633b-4fea-9981-159362f44053" Dec 11 08:14:32 crc kubenswrapper[4860]: E1211 08:14:32.270630 4860 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 11 08:14:32 crc kubenswrapper[4860]: E1211 08:14:32.270801 4860 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-fnfjs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-mgfr4_openshift-marketplace(30107273-766a-4575-8cb1-3de387248437): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 11 08:14:32 crc kubenswrapper[4860]: E1211 08:14:32.271992 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-mgfr4" podUID="30107273-766a-4575-8cb1-3de387248437" Dec 11 08:14:32 crc kubenswrapper[4860]: E1211 08:14:32.436954 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-55jtz" podUID="b498b689-633b-4fea-9981-159362f44053" Dec 11 08:14:32 crc kubenswrapper[4860]: E1211 08:14:32.436988 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-8p4mb" podUID="01481957-a690-40f4-8c89-352b7350d327" Dec 11 08:14:32 crc kubenswrapper[4860]: E1211 08:14:32.437025 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-mgfr4" podUID="30107273-766a-4575-8cb1-3de387248437" Dec 11 08:14:36 crc kubenswrapper[4860]: E1211 08:14:36.532170 4860 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 11 08:14:36 crc kubenswrapper[4860]: E1211 08:14:36.532615 4860 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ftgqm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-dzg4x_openshift-marketplace(294792dd-5104-4308-a10f-f3ab0c4504e9): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 11 08:14:36 crc kubenswrapper[4860]: E1211 08:14:36.534477 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-dzg4x" podUID="294792dd-5104-4308-a10f-f3ab0c4504e9" Dec 11 08:14:36 crc kubenswrapper[4860]: E1211 08:14:36.539937 4860 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 11 08:14:36 crc kubenswrapper[4860]: E1211 08:14:36.540096 4860 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-x55qf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-dcccn_openshift-marketplace(8e052d7c-59d1-43a4-ac7a-e9daf2a88c65): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 11 08:14:36 crc kubenswrapper[4860]: E1211 08:14:36.541290 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-dcccn" podUID="8e052d7c-59d1-43a4-ac7a-e9daf2a88c65" Dec 11 08:14:36 crc kubenswrapper[4860]: I1211 08:14:36.723249 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 11 08:14:36 crc kubenswrapper[4860]: W1211 08:14:36.747252 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod934cd13b_9314_4299_8b5b_5a8e03810781.slice/crio-bb8f795ee28d8b694be1d1c1eb441230498472aefd965db859ac65dd991aa9f1 WatchSource:0}: Error finding container bb8f795ee28d8b694be1d1c1eb441230498472aefd965db859ac65dd991aa9f1: Status 404 returned error can't find the container with id bb8f795ee28d8b694be1d1c1eb441230498472aefd965db859ac65dd991aa9f1 Dec 11 08:14:36 crc kubenswrapper[4860]: I1211 08:14:36.887025 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 11 08:14:36 crc kubenswrapper[4860]: W1211 08:14:36.912032 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod3bea0969_d146_4f96_8f02_b95bfe17aac3.slice/crio-73d4aed6dce133a753fe1f5a48696f5eeefba1694e560471a71c2b08e8f01b48 WatchSource:0}: Error finding container 73d4aed6dce133a753fe1f5a48696f5eeefba1694e560471a71c2b08e8f01b48: Status 404 returned error can't find the container with id 73d4aed6dce133a753fe1f5a48696f5eeefba1694e560471a71c2b08e8f01b48 Dec 11 08:14:37 crc kubenswrapper[4860]: I1211 08:14:37.474164 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"934cd13b-9314-4299-8b5b-5a8e03810781","Type":"ContainerStarted","Data":"a93aaba1419b8869ffe60bda0d3c6c37ad4345af08c8b66c76b301aafa3611e8"} Dec 11 08:14:37 crc kubenswrapper[4860]: I1211 08:14:37.474456 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"934cd13b-9314-4299-8b5b-5a8e03810781","Type":"ContainerStarted","Data":"bb8f795ee28d8b694be1d1c1eb441230498472aefd965db859ac65dd991aa9f1"} Dec 11 08:14:37 crc kubenswrapper[4860]: I1211 08:14:37.477089 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"3bea0969-d146-4f96-8f02-b95bfe17aac3","Type":"ContainerStarted","Data":"7d866182c7851f4fab0b8fe46f5ac4afb5266a3665af4234c0509e1fb87b399f"} Dec 11 08:14:37 crc kubenswrapper[4860]: I1211 08:14:37.478407 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"3bea0969-d146-4f96-8f02-b95bfe17aac3","Type":"ContainerStarted","Data":"73d4aed6dce133a753fe1f5a48696f5eeefba1694e560471a71c2b08e8f01b48"} Dec 11 08:14:37 crc kubenswrapper[4860]: E1211 08:14:37.480300 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-dcccn" podUID="8e052d7c-59d1-43a4-ac7a-e9daf2a88c65" Dec 11 08:14:37 crc kubenswrapper[4860]: E1211 08:14:37.480653 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-dzg4x" podUID="294792dd-5104-4308-a10f-f3ab0c4504e9" Dec 11 08:14:37 crc kubenswrapper[4860]: I1211 08:14:37.490891 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-9-crc" podStartSLOduration=16.490869626 podStartE2EDuration="16.490869626s" podCreationTimestamp="2025-12-11 08:14:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:14:37.48890784 +0000 UTC m=+210.217426915" watchObservedRunningTime="2025-12-11 08:14:37.490869626 +0000 UTC m=+210.219388681" Dec 11 08:14:37 crc kubenswrapper[4860]: I1211 08:14:37.521873 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=11.52184764 podStartE2EDuration="11.52184764s" podCreationTimestamp="2025-12-11 08:14:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:14:37.519110017 +0000 UTC m=+210.247629082" watchObservedRunningTime="2025-12-11 08:14:37.52184764 +0000 UTC m=+210.250366695" Dec 11 08:14:38 crc kubenswrapper[4860]: I1211 08:14:38.484051 4860 generic.go:334] "Generic (PLEG): container finished" podID="934cd13b-9314-4299-8b5b-5a8e03810781" containerID="a93aaba1419b8869ffe60bda0d3c6c37ad4345af08c8b66c76b301aafa3611e8" exitCode=0 Dec 11 08:14:38 crc kubenswrapper[4860]: I1211 08:14:38.484105 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"934cd13b-9314-4299-8b5b-5a8e03810781","Type":"ContainerDied","Data":"a93aaba1419b8869ffe60bda0d3c6c37ad4345af08c8b66c76b301aafa3611e8"} Dec 11 08:14:38 crc kubenswrapper[4860]: I1211 08:14:38.795079 4860 patch_prober.go:28] interesting pod/machine-config-daemon-99qgp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 08:14:38 crc kubenswrapper[4860]: I1211 08:14:38.795148 4860 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 08:14:38 crc kubenswrapper[4860]: I1211 08:14:38.795205 4860 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" Dec 11 08:14:38 crc kubenswrapper[4860]: I1211 08:14:38.795857 4860 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b2d4d5de7106175087d91acb1f0c7552014f0c08bfcd65cbbf184d3f8d7826ff"} pod="openshift-machine-config-operator/machine-config-daemon-99qgp" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 11 08:14:38 crc kubenswrapper[4860]: I1211 08:14:38.795981 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" containerName="machine-config-daemon" containerID="cri-o://b2d4d5de7106175087d91acb1f0c7552014f0c08bfcd65cbbf184d3f8d7826ff" gracePeriod=600 Dec 11 08:14:39 crc kubenswrapper[4860]: I1211 08:14:39.491111 4860 generic.go:334] "Generic (PLEG): container finished" podID="31c30642-6e60-41b4-a477-0d802424e0aa" containerID="b2d4d5de7106175087d91acb1f0c7552014f0c08bfcd65cbbf184d3f8d7826ff" exitCode=0 Dec 11 08:14:39 crc kubenswrapper[4860]: I1211 08:14:39.491196 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" event={"ID":"31c30642-6e60-41b4-a477-0d802424e0aa","Type":"ContainerDied","Data":"b2d4d5de7106175087d91acb1f0c7552014f0c08bfcd65cbbf184d3f8d7826ff"} Dec 11 08:14:39 crc kubenswrapper[4860]: I1211 08:14:39.491622 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" event={"ID":"31c30642-6e60-41b4-a477-0d802424e0aa","Type":"ContainerStarted","Data":"cddf60b7695e04c9618f575369ea700b35bf017d924547b72625c2bacba04e02"} Dec 11 08:14:39 crc kubenswrapper[4860]: I1211 08:14:39.711920 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 11 08:14:39 crc kubenswrapper[4860]: I1211 08:14:39.772398 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/934cd13b-9314-4299-8b5b-5a8e03810781-kube-api-access\") pod \"934cd13b-9314-4299-8b5b-5a8e03810781\" (UID: \"934cd13b-9314-4299-8b5b-5a8e03810781\") " Dec 11 08:14:39 crc kubenswrapper[4860]: I1211 08:14:39.772460 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/934cd13b-9314-4299-8b5b-5a8e03810781-kubelet-dir\") pod \"934cd13b-9314-4299-8b5b-5a8e03810781\" (UID: \"934cd13b-9314-4299-8b5b-5a8e03810781\") " Dec 11 08:14:39 crc kubenswrapper[4860]: I1211 08:14:39.772754 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/934cd13b-9314-4299-8b5b-5a8e03810781-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "934cd13b-9314-4299-8b5b-5a8e03810781" (UID: "934cd13b-9314-4299-8b5b-5a8e03810781"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 08:14:39 crc kubenswrapper[4860]: I1211 08:14:39.780157 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/934cd13b-9314-4299-8b5b-5a8e03810781-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "934cd13b-9314-4299-8b5b-5a8e03810781" (UID: "934cd13b-9314-4299-8b5b-5a8e03810781"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:14:39 crc kubenswrapper[4860]: I1211 08:14:39.873693 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/934cd13b-9314-4299-8b5b-5a8e03810781-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 11 08:14:39 crc kubenswrapper[4860]: I1211 08:14:39.873747 4860 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/934cd13b-9314-4299-8b5b-5a8e03810781-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 11 08:14:40 crc kubenswrapper[4860]: I1211 08:14:40.500387 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"934cd13b-9314-4299-8b5b-5a8e03810781","Type":"ContainerDied","Data":"bb8f795ee28d8b694be1d1c1eb441230498472aefd965db859ac65dd991aa9f1"} Dec 11 08:14:40 crc kubenswrapper[4860]: I1211 08:14:40.500701 4860 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bb8f795ee28d8b694be1d1c1eb441230498472aefd965db859ac65dd991aa9f1" Dec 11 08:14:40 crc kubenswrapper[4860]: I1211 08:14:40.500629 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 11 08:14:42 crc kubenswrapper[4860]: I1211 08:14:42.511882 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kqwgm" event={"ID":"093a9903-7041-410b-bfb2-c74ef741e0be","Type":"ContainerStarted","Data":"733a892d4a4dafcbbc2de14f8a486771dc23a0037161b96eb03eff048301e384"} Dec 11 08:14:43 crc kubenswrapper[4860]: I1211 08:14:43.517734 4860 generic.go:334] "Generic (PLEG): container finished" podID="093a9903-7041-410b-bfb2-c74ef741e0be" containerID="733a892d4a4dafcbbc2de14f8a486771dc23a0037161b96eb03eff048301e384" exitCode=0 Dec 11 08:14:43 crc kubenswrapper[4860]: I1211 08:14:43.517824 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kqwgm" event={"ID":"093a9903-7041-410b-bfb2-c74ef741e0be","Type":"ContainerDied","Data":"733a892d4a4dafcbbc2de14f8a486771dc23a0037161b96eb03eff048301e384"} Dec 11 08:14:43 crc kubenswrapper[4860]: I1211 08:14:43.521673 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p6f75" event={"ID":"f9b23a65-98b8-4d83-a71f-2316fc1d46a0","Type":"ContainerStarted","Data":"d56b94ad7f779326751392fd4ea22b63b32b176c75e8d6387550d7393c40e173"} Dec 11 08:14:44 crc kubenswrapper[4860]: I1211 08:14:44.529659 4860 generic.go:334] "Generic (PLEG): container finished" podID="f9b23a65-98b8-4d83-a71f-2316fc1d46a0" containerID="d56b94ad7f779326751392fd4ea22b63b32b176c75e8d6387550d7393c40e173" exitCode=0 Dec 11 08:14:44 crc kubenswrapper[4860]: I1211 08:14:44.529747 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p6f75" event={"ID":"f9b23a65-98b8-4d83-a71f-2316fc1d46a0","Type":"ContainerDied","Data":"d56b94ad7f779326751392fd4ea22b63b32b176c75e8d6387550d7393c40e173"} Dec 11 08:14:45 crc kubenswrapper[4860]: I1211 08:14:45.544000 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p6f75" event={"ID":"f9b23a65-98b8-4d83-a71f-2316fc1d46a0","Type":"ContainerStarted","Data":"c9c0c514837b2c179afb0c70eb50b7427f0b1cdfa4f763026250e8a797e2e8a0"} Dec 11 08:14:45 crc kubenswrapper[4860]: I1211 08:14:45.548062 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kqwgm" event={"ID":"093a9903-7041-410b-bfb2-c74ef741e0be","Type":"ContainerStarted","Data":"41c0f2c9f5203be9afc00992cd7768f8a3896e585a81b5618856b40455c598eb"} Dec 11 08:14:45 crc kubenswrapper[4860]: I1211 08:14:45.567028 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-p6f75" podStartSLOduration=4.467574116 podStartE2EDuration="1m3.567010953s" podCreationTimestamp="2025-12-11 08:13:42 +0000 UTC" firstStartedPulling="2025-12-11 08:13:45.850632741 +0000 UTC m=+158.579151796" lastFinishedPulling="2025-12-11 08:14:44.950069578 +0000 UTC m=+217.678588633" observedRunningTime="2025-12-11 08:14:45.565016435 +0000 UTC m=+218.293535480" watchObservedRunningTime="2025-12-11 08:14:45.567010953 +0000 UTC m=+218.295530008" Dec 11 08:14:45 crc kubenswrapper[4860]: I1211 08:14:45.587587 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-kqwgm" podStartSLOduration=2.920136879 podStartE2EDuration="1m3.587569812s" podCreationTimestamp="2025-12-11 08:13:42 +0000 UTC" firstStartedPulling="2025-12-11 08:13:43.748742053 +0000 UTC m=+156.477261108" lastFinishedPulling="2025-12-11 08:14:44.416174986 +0000 UTC m=+217.144694041" observedRunningTime="2025-12-11 08:14:45.586207725 +0000 UTC m=+218.314726780" watchObservedRunningTime="2025-12-11 08:14:45.587569812 +0000 UTC m=+218.316088857" Dec 11 08:14:47 crc kubenswrapper[4860]: I1211 08:14:47.561382 4860 generic.go:334] "Generic (PLEG): container finished" podID="01481957-a690-40f4-8c89-352b7350d327" containerID="b4534c18fcfa223bf76f764f0fb5c2a190fe3875b8959c266ed75e731ee682b3" exitCode=0 Dec 11 08:14:47 crc kubenswrapper[4860]: I1211 08:14:47.561463 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8p4mb" event={"ID":"01481957-a690-40f4-8c89-352b7350d327","Type":"ContainerDied","Data":"b4534c18fcfa223bf76f764f0fb5c2a190fe3875b8959c266ed75e731ee682b3"} Dec 11 08:14:47 crc kubenswrapper[4860]: I1211 08:14:47.564468 4860 generic.go:334] "Generic (PLEG): container finished" podID="b498b689-633b-4fea-9981-159362f44053" containerID="c0c0e7688411dc667e3662bab7a995ed924b7b0081d1817049f17f8eddb6facb" exitCode=0 Dec 11 08:14:47 crc kubenswrapper[4860]: I1211 08:14:47.564504 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-55jtz" event={"ID":"b498b689-633b-4fea-9981-159362f44053","Type":"ContainerDied","Data":"c0c0e7688411dc667e3662bab7a995ed924b7b0081d1817049f17f8eddb6facb"} Dec 11 08:14:48 crc kubenswrapper[4860]: I1211 08:14:48.574698 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-55jtz" event={"ID":"b498b689-633b-4fea-9981-159362f44053","Type":"ContainerStarted","Data":"48780a65d92d24e4e66a3d0c38e551061aae518f6d4282361955c53d297ac635"} Dec 11 08:14:48 crc kubenswrapper[4860]: I1211 08:14:48.596480 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-55jtz" podStartSLOduration=3.22509794 podStartE2EDuration="1m6.596461943s" podCreationTimestamp="2025-12-11 08:13:42 +0000 UTC" firstStartedPulling="2025-12-11 08:13:44.771936175 +0000 UTC m=+157.500455230" lastFinishedPulling="2025-12-11 08:14:48.143300138 +0000 UTC m=+220.871819233" observedRunningTime="2025-12-11 08:14:48.591590877 +0000 UTC m=+221.320109932" watchObservedRunningTime="2025-12-11 08:14:48.596461943 +0000 UTC m=+221.324980998" Dec 11 08:14:49 crc kubenswrapper[4860]: I1211 08:14:49.586053 4860 generic.go:334] "Generic (PLEG): container finished" podID="7ac89442-1754-45d3-b67a-aa6cc31a1235" containerID="2eec0fca2445e8ac1ce738b2d8ec1e8ab0d3f134b7d3fc0fa8c11102842021c6" exitCode=0 Dec 11 08:14:49 crc kubenswrapper[4860]: I1211 08:14:49.588635 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pqtvk" event={"ID":"7ac89442-1754-45d3-b67a-aa6cc31a1235","Type":"ContainerDied","Data":"2eec0fca2445e8ac1ce738b2d8ec1e8ab0d3f134b7d3fc0fa8c11102842021c6"} Dec 11 08:14:49 crc kubenswrapper[4860]: I1211 08:14:49.589789 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8p4mb" event={"ID":"01481957-a690-40f4-8c89-352b7350d327","Type":"ContainerStarted","Data":"d9777d9cd9cfe1b753ed0336374b4591e9f9e22227f397713e583f8e99a68df8"} Dec 11 08:14:49 crc kubenswrapper[4860]: I1211 08:14:49.594832 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mgfr4" event={"ID":"30107273-766a-4575-8cb1-3de387248437","Type":"ContainerStarted","Data":"a1981178d5f6012d6286ef994fb1180c51510e9a39d27894f17ec36a76a2cf05"} Dec 11 08:14:49 crc kubenswrapper[4860]: I1211 08:14:49.620493 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-8p4mb" podStartSLOduration=2.992500119 podStartE2EDuration="1m4.620460835s" podCreationTimestamp="2025-12-11 08:13:45 +0000 UTC" firstStartedPulling="2025-12-11 08:13:46.953937491 +0000 UTC m=+159.682456546" lastFinishedPulling="2025-12-11 08:14:48.581898207 +0000 UTC m=+221.310417262" observedRunningTime="2025-12-11 08:14:49.619262664 +0000 UTC m=+222.347781719" watchObservedRunningTime="2025-12-11 08:14:49.620460835 +0000 UTC m=+222.348979900" Dec 11 08:14:50 crc kubenswrapper[4860]: I1211 08:14:50.603952 4860 generic.go:334] "Generic (PLEG): container finished" podID="30107273-766a-4575-8cb1-3de387248437" containerID="a1981178d5f6012d6286ef994fb1180c51510e9a39d27894f17ec36a76a2cf05" exitCode=0 Dec 11 08:14:50 crc kubenswrapper[4860]: I1211 08:14:50.604002 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mgfr4" event={"ID":"30107273-766a-4575-8cb1-3de387248437","Type":"ContainerDied","Data":"a1981178d5f6012d6286ef994fb1180c51510e9a39d27894f17ec36a76a2cf05"} Dec 11 08:14:52 crc kubenswrapper[4860]: I1211 08:14:52.528301 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-kqwgm" Dec 11 08:14:52 crc kubenswrapper[4860]: I1211 08:14:52.528709 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-kqwgm" Dec 11 08:14:52 crc kubenswrapper[4860]: I1211 08:14:52.739997 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-kqwgm" Dec 11 08:14:52 crc kubenswrapper[4860]: I1211 08:14:52.813572 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-kqwgm" Dec 11 08:14:52 crc kubenswrapper[4860]: I1211 08:14:52.960347 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-55jtz" Dec 11 08:14:52 crc kubenswrapper[4860]: I1211 08:14:52.960730 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-55jtz" Dec 11 08:14:53 crc kubenswrapper[4860]: I1211 08:14:53.005151 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-55jtz" Dec 11 08:14:53 crc kubenswrapper[4860]: I1211 08:14:53.178385 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-p6f75" Dec 11 08:14:53 crc kubenswrapper[4860]: I1211 08:14:53.178534 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-p6f75" Dec 11 08:14:53 crc kubenswrapper[4860]: I1211 08:14:53.241373 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-p6f75" Dec 11 08:14:53 crc kubenswrapper[4860]: I1211 08:14:53.670702 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-p6f75" Dec 11 08:14:53 crc kubenswrapper[4860]: I1211 08:14:53.688078 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-55jtz" Dec 11 08:14:54 crc kubenswrapper[4860]: I1211 08:14:54.805610 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-p6f75"] Dec 11 08:14:55 crc kubenswrapper[4860]: I1211 08:14:55.634379 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pqtvk" event={"ID":"7ac89442-1754-45d3-b67a-aa6cc31a1235","Type":"ContainerStarted","Data":"f1a5ae4b9e0d2bfe7ffbc7fd778a4d1ed530e9954a2675c84cc36798d229c6f8"} Dec 11 08:14:55 crc kubenswrapper[4860]: I1211 08:14:55.659456 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-pqtvk" podStartSLOduration=4.092787822 podStartE2EDuration="1m13.659431336s" podCreationTimestamp="2025-12-11 08:13:42 +0000 UTC" firstStartedPulling="2025-12-11 08:13:44.760944588 +0000 UTC m=+157.489463643" lastFinishedPulling="2025-12-11 08:14:54.327588072 +0000 UTC m=+227.056107157" observedRunningTime="2025-12-11 08:14:55.657676916 +0000 UTC m=+228.386195991" watchObservedRunningTime="2025-12-11 08:14:55.659431336 +0000 UTC m=+228.387950391" Dec 11 08:14:55 crc kubenswrapper[4860]: I1211 08:14:55.760588 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-8p4mb" Dec 11 08:14:55 crc kubenswrapper[4860]: I1211 08:14:55.760817 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-8p4mb" Dec 11 08:14:55 crc kubenswrapper[4860]: I1211 08:14:55.809483 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-8p4mb" Dec 11 08:14:56 crc kubenswrapper[4860]: I1211 08:14:56.640221 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-p6f75" podUID="f9b23a65-98b8-4d83-a71f-2316fc1d46a0" containerName="registry-server" containerID="cri-o://c9c0c514837b2c179afb0c70eb50b7427f0b1cdfa4f763026250e8a797e2e8a0" gracePeriod=2 Dec 11 08:14:56 crc kubenswrapper[4860]: I1211 08:14:56.686091 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-8p4mb" Dec 11 08:14:57 crc kubenswrapper[4860]: I1211 08:14:57.007818 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-55jtz"] Dec 11 08:14:57 crc kubenswrapper[4860]: I1211 08:14:57.008129 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-55jtz" podUID="b498b689-633b-4fea-9981-159362f44053" containerName="registry-server" containerID="cri-o://48780a65d92d24e4e66a3d0c38e551061aae518f6d4282361955c53d297ac635" gracePeriod=2 Dec 11 08:14:59 crc kubenswrapper[4860]: I1211 08:14:59.672092 4860 generic.go:334] "Generic (PLEG): container finished" podID="b498b689-633b-4fea-9981-159362f44053" containerID="48780a65d92d24e4e66a3d0c38e551061aae518f6d4282361955c53d297ac635" exitCode=0 Dec 11 08:14:59 crc kubenswrapper[4860]: I1211 08:14:59.672169 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-55jtz" event={"ID":"b498b689-633b-4fea-9981-159362f44053","Type":"ContainerDied","Data":"48780a65d92d24e4e66a3d0c38e551061aae518f6d4282361955c53d297ac635"} Dec 11 08:14:59 crc kubenswrapper[4860]: I1211 08:14:59.675249 4860 generic.go:334] "Generic (PLEG): container finished" podID="f9b23a65-98b8-4d83-a71f-2316fc1d46a0" containerID="c9c0c514837b2c179afb0c70eb50b7427f0b1cdfa4f763026250e8a797e2e8a0" exitCode=0 Dec 11 08:14:59 crc kubenswrapper[4860]: I1211 08:14:59.675289 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p6f75" event={"ID":"f9b23a65-98b8-4d83-a71f-2316fc1d46a0","Type":"ContainerDied","Data":"c9c0c514837b2c179afb0c70eb50b7427f0b1cdfa4f763026250e8a797e2e8a0"} Dec 11 08:15:00 crc kubenswrapper[4860]: I1211 08:15:00.141750 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29424015-j4cd9"] Dec 11 08:15:00 crc kubenswrapper[4860]: E1211 08:15:00.142160 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="934cd13b-9314-4299-8b5b-5a8e03810781" containerName="pruner" Dec 11 08:15:00 crc kubenswrapper[4860]: I1211 08:15:00.142199 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="934cd13b-9314-4299-8b5b-5a8e03810781" containerName="pruner" Dec 11 08:15:00 crc kubenswrapper[4860]: I1211 08:15:00.142474 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="934cd13b-9314-4299-8b5b-5a8e03810781" containerName="pruner" Dec 11 08:15:00 crc kubenswrapper[4860]: I1211 08:15:00.143142 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29424015-j4cd9" Dec 11 08:15:00 crc kubenswrapper[4860]: I1211 08:15:00.146008 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 11 08:15:00 crc kubenswrapper[4860]: I1211 08:15:00.147184 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 11 08:15:00 crc kubenswrapper[4860]: I1211 08:15:00.159592 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29424015-j4cd9"] Dec 11 08:15:00 crc kubenswrapper[4860]: I1211 08:15:00.296566 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dhmlx\" (UniqueName: \"kubernetes.io/projected/a32ff4fb-1fd8-4097-88cf-74c67cf6fa56-kube-api-access-dhmlx\") pod \"collect-profiles-29424015-j4cd9\" (UID: \"a32ff4fb-1fd8-4097-88cf-74c67cf6fa56\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424015-j4cd9" Dec 11 08:15:00 crc kubenswrapper[4860]: I1211 08:15:00.296661 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a32ff4fb-1fd8-4097-88cf-74c67cf6fa56-secret-volume\") pod \"collect-profiles-29424015-j4cd9\" (UID: \"a32ff4fb-1fd8-4097-88cf-74c67cf6fa56\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424015-j4cd9" Dec 11 08:15:00 crc kubenswrapper[4860]: I1211 08:15:00.297008 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a32ff4fb-1fd8-4097-88cf-74c67cf6fa56-config-volume\") pod \"collect-profiles-29424015-j4cd9\" (UID: \"a32ff4fb-1fd8-4097-88cf-74c67cf6fa56\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424015-j4cd9" Dec 11 08:15:00 crc kubenswrapper[4860]: I1211 08:15:00.398390 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a32ff4fb-1fd8-4097-88cf-74c67cf6fa56-secret-volume\") pod \"collect-profiles-29424015-j4cd9\" (UID: \"a32ff4fb-1fd8-4097-88cf-74c67cf6fa56\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424015-j4cd9" Dec 11 08:15:00 crc kubenswrapper[4860]: I1211 08:15:00.398459 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a32ff4fb-1fd8-4097-88cf-74c67cf6fa56-config-volume\") pod \"collect-profiles-29424015-j4cd9\" (UID: \"a32ff4fb-1fd8-4097-88cf-74c67cf6fa56\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424015-j4cd9" Dec 11 08:15:00 crc kubenswrapper[4860]: I1211 08:15:00.398547 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dhmlx\" (UniqueName: \"kubernetes.io/projected/a32ff4fb-1fd8-4097-88cf-74c67cf6fa56-kube-api-access-dhmlx\") pod \"collect-profiles-29424015-j4cd9\" (UID: \"a32ff4fb-1fd8-4097-88cf-74c67cf6fa56\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424015-j4cd9" Dec 11 08:15:00 crc kubenswrapper[4860]: I1211 08:15:00.399940 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a32ff4fb-1fd8-4097-88cf-74c67cf6fa56-config-volume\") pod \"collect-profiles-29424015-j4cd9\" (UID: \"a32ff4fb-1fd8-4097-88cf-74c67cf6fa56\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424015-j4cd9" Dec 11 08:15:00 crc kubenswrapper[4860]: I1211 08:15:00.407674 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a32ff4fb-1fd8-4097-88cf-74c67cf6fa56-secret-volume\") pod \"collect-profiles-29424015-j4cd9\" (UID: \"a32ff4fb-1fd8-4097-88cf-74c67cf6fa56\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424015-j4cd9" Dec 11 08:15:00 crc kubenswrapper[4860]: I1211 08:15:00.420824 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dhmlx\" (UniqueName: \"kubernetes.io/projected/a32ff4fb-1fd8-4097-88cf-74c67cf6fa56-kube-api-access-dhmlx\") pod \"collect-profiles-29424015-j4cd9\" (UID: \"a32ff4fb-1fd8-4097-88cf-74c67cf6fa56\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424015-j4cd9" Dec 11 08:15:00 crc kubenswrapper[4860]: I1211 08:15:00.515069 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29424015-j4cd9" Dec 11 08:15:02 crc kubenswrapper[4860]: I1211 08:15:02.704174 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-pqtvk" Dec 11 08:15:02 crc kubenswrapper[4860]: I1211 08:15:02.704500 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-pqtvk" Dec 11 08:15:02 crc kubenswrapper[4860]: I1211 08:15:02.745452 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-pqtvk" Dec 11 08:15:02 crc kubenswrapper[4860]: E1211 08:15:02.960853 4860 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 48780a65d92d24e4e66a3d0c38e551061aae518f6d4282361955c53d297ac635 is running failed: container process not found" containerID="48780a65d92d24e4e66a3d0c38e551061aae518f6d4282361955c53d297ac635" cmd=["grpc_health_probe","-addr=:50051"] Dec 11 08:15:02 crc kubenswrapper[4860]: E1211 08:15:02.961613 4860 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 48780a65d92d24e4e66a3d0c38e551061aae518f6d4282361955c53d297ac635 is running failed: container process not found" containerID="48780a65d92d24e4e66a3d0c38e551061aae518f6d4282361955c53d297ac635" cmd=["grpc_health_probe","-addr=:50051"] Dec 11 08:15:02 crc kubenswrapper[4860]: E1211 08:15:02.962080 4860 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 48780a65d92d24e4e66a3d0c38e551061aae518f6d4282361955c53d297ac635 is running failed: container process not found" containerID="48780a65d92d24e4e66a3d0c38e551061aae518f6d4282361955c53d297ac635" cmd=["grpc_health_probe","-addr=:50051"] Dec 11 08:15:02 crc kubenswrapper[4860]: E1211 08:15:02.962160 4860 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 48780a65d92d24e4e66a3d0c38e551061aae518f6d4282361955c53d297ac635 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/certified-operators-55jtz" podUID="b498b689-633b-4fea-9981-159362f44053" containerName="registry-server" Dec 11 08:15:03 crc kubenswrapper[4860]: E1211 08:15:03.179277 4860 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of c9c0c514837b2c179afb0c70eb50b7427f0b1cdfa4f763026250e8a797e2e8a0 is running failed: container process not found" containerID="c9c0c514837b2c179afb0c70eb50b7427f0b1cdfa4f763026250e8a797e2e8a0" cmd=["grpc_health_probe","-addr=:50051"] Dec 11 08:15:03 crc kubenswrapper[4860]: E1211 08:15:03.180439 4860 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of c9c0c514837b2c179afb0c70eb50b7427f0b1cdfa4f763026250e8a797e2e8a0 is running failed: container process not found" containerID="c9c0c514837b2c179afb0c70eb50b7427f0b1cdfa4f763026250e8a797e2e8a0" cmd=["grpc_health_probe","-addr=:50051"] Dec 11 08:15:03 crc kubenswrapper[4860]: E1211 08:15:03.180954 4860 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of c9c0c514837b2c179afb0c70eb50b7427f0b1cdfa4f763026250e8a797e2e8a0 is running failed: container process not found" containerID="c9c0c514837b2c179afb0c70eb50b7427f0b1cdfa4f763026250e8a797e2e8a0" cmd=["grpc_health_probe","-addr=:50051"] Dec 11 08:15:03 crc kubenswrapper[4860]: E1211 08:15:03.181001 4860 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of c9c0c514837b2c179afb0c70eb50b7427f0b1cdfa4f763026250e8a797e2e8a0 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/community-operators-p6f75" podUID="f9b23a65-98b8-4d83-a71f-2316fc1d46a0" containerName="registry-server" Dec 11 08:15:03 crc kubenswrapper[4860]: I1211 08:15:03.739963 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-pqtvk" Dec 11 08:15:05 crc kubenswrapper[4860]: I1211 08:15:05.233217 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-55jtz" Dec 11 08:15:05 crc kubenswrapper[4860]: I1211 08:15:05.271338 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p6f75" Dec 11 08:15:05 crc kubenswrapper[4860]: I1211 08:15:05.367795 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf8wg\" (UniqueName: \"kubernetes.io/projected/b498b689-633b-4fea-9981-159362f44053-kube-api-access-gf8wg\") pod \"b498b689-633b-4fea-9981-159362f44053\" (UID: \"b498b689-633b-4fea-9981-159362f44053\") " Dec 11 08:15:05 crc kubenswrapper[4860]: I1211 08:15:05.368252 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b498b689-633b-4fea-9981-159362f44053-catalog-content\") pod \"b498b689-633b-4fea-9981-159362f44053\" (UID: \"b498b689-633b-4fea-9981-159362f44053\") " Dec 11 08:15:05 crc kubenswrapper[4860]: I1211 08:15:05.368454 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b498b689-633b-4fea-9981-159362f44053-utilities\") pod \"b498b689-633b-4fea-9981-159362f44053\" (UID: \"b498b689-633b-4fea-9981-159362f44053\") " Dec 11 08:15:05 crc kubenswrapper[4860]: I1211 08:15:05.371213 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b498b689-633b-4fea-9981-159362f44053-utilities" (OuterVolumeSpecName: "utilities") pod "b498b689-633b-4fea-9981-159362f44053" (UID: "b498b689-633b-4fea-9981-159362f44053"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:15:05 crc kubenswrapper[4860]: I1211 08:15:05.376919 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b498b689-633b-4fea-9981-159362f44053-kube-api-access-gf8wg" (OuterVolumeSpecName: "kube-api-access-gf8wg") pod "b498b689-633b-4fea-9981-159362f44053" (UID: "b498b689-633b-4fea-9981-159362f44053"). InnerVolumeSpecName "kube-api-access-gf8wg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:15:05 crc kubenswrapper[4860]: I1211 08:15:05.428732 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b498b689-633b-4fea-9981-159362f44053-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b498b689-633b-4fea-9981-159362f44053" (UID: "b498b689-633b-4fea-9981-159362f44053"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:15:05 crc kubenswrapper[4860]: I1211 08:15:05.437630 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29424015-j4cd9"] Dec 11 08:15:05 crc kubenswrapper[4860]: I1211 08:15:05.470999 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-whc24\" (UniqueName: \"kubernetes.io/projected/f9b23a65-98b8-4d83-a71f-2316fc1d46a0-kube-api-access-whc24\") pod \"f9b23a65-98b8-4d83-a71f-2316fc1d46a0\" (UID: \"f9b23a65-98b8-4d83-a71f-2316fc1d46a0\") " Dec 11 08:15:05 crc kubenswrapper[4860]: I1211 08:15:05.471064 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f9b23a65-98b8-4d83-a71f-2316fc1d46a0-catalog-content\") pod \"f9b23a65-98b8-4d83-a71f-2316fc1d46a0\" (UID: \"f9b23a65-98b8-4d83-a71f-2316fc1d46a0\") " Dec 11 08:15:05 crc kubenswrapper[4860]: I1211 08:15:05.471153 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f9b23a65-98b8-4d83-a71f-2316fc1d46a0-utilities\") pod \"f9b23a65-98b8-4d83-a71f-2316fc1d46a0\" (UID: \"f9b23a65-98b8-4d83-a71f-2316fc1d46a0\") " Dec 11 08:15:05 crc kubenswrapper[4860]: I1211 08:15:05.472159 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f9b23a65-98b8-4d83-a71f-2316fc1d46a0-utilities" (OuterVolumeSpecName: "utilities") pod "f9b23a65-98b8-4d83-a71f-2316fc1d46a0" (UID: "f9b23a65-98b8-4d83-a71f-2316fc1d46a0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:15:05 crc kubenswrapper[4860]: I1211 08:15:05.472603 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf8wg\" (UniqueName: \"kubernetes.io/projected/b498b689-633b-4fea-9981-159362f44053-kube-api-access-gf8wg\") on node \"crc\" DevicePath \"\"" Dec 11 08:15:05 crc kubenswrapper[4860]: I1211 08:15:05.473264 4860 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f9b23a65-98b8-4d83-a71f-2316fc1d46a0-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 08:15:05 crc kubenswrapper[4860]: I1211 08:15:05.473291 4860 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b498b689-633b-4fea-9981-159362f44053-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 08:15:05 crc kubenswrapper[4860]: I1211 08:15:05.473302 4860 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b498b689-633b-4fea-9981-159362f44053-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 08:15:05 crc kubenswrapper[4860]: I1211 08:15:05.477248 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f9b23a65-98b8-4d83-a71f-2316fc1d46a0-kube-api-access-whc24" (OuterVolumeSpecName: "kube-api-access-whc24") pod "f9b23a65-98b8-4d83-a71f-2316fc1d46a0" (UID: "f9b23a65-98b8-4d83-a71f-2316fc1d46a0"). InnerVolumeSpecName "kube-api-access-whc24". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:15:05 crc kubenswrapper[4860]: I1211 08:15:05.537476 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f9b23a65-98b8-4d83-a71f-2316fc1d46a0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f9b23a65-98b8-4d83-a71f-2316fc1d46a0" (UID: "f9b23a65-98b8-4d83-a71f-2316fc1d46a0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:15:05 crc kubenswrapper[4860]: I1211 08:15:05.574730 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-whc24\" (UniqueName: \"kubernetes.io/projected/f9b23a65-98b8-4d83-a71f-2316fc1d46a0-kube-api-access-whc24\") on node \"crc\" DevicePath \"\"" Dec 11 08:15:05 crc kubenswrapper[4860]: I1211 08:15:05.574757 4860 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f9b23a65-98b8-4d83-a71f-2316fc1d46a0-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 08:15:05 crc kubenswrapper[4860]: I1211 08:15:05.706856 4860 generic.go:334] "Generic (PLEG): container finished" podID="8e052d7c-59d1-43a4-ac7a-e9daf2a88c65" containerID="2383f4d15b00b86de2542756c810c6715d4b50e92963ff933972eef2e3906e13" exitCode=0 Dec 11 08:15:05 crc kubenswrapper[4860]: I1211 08:15:05.706922 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dcccn" event={"ID":"8e052d7c-59d1-43a4-ac7a-e9daf2a88c65","Type":"ContainerDied","Data":"2383f4d15b00b86de2542756c810c6715d4b50e92963ff933972eef2e3906e13"} Dec 11 08:15:05 crc kubenswrapper[4860]: I1211 08:15:05.709728 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-55jtz" event={"ID":"b498b689-633b-4fea-9981-159362f44053","Type":"ContainerDied","Data":"c2f4f4cd1623e2a47b32453685b7ff3d774fe09f1d80bb0181e5ee581550c48d"} Dec 11 08:15:05 crc kubenswrapper[4860]: I1211 08:15:05.709764 4860 scope.go:117] "RemoveContainer" containerID="48780a65d92d24e4e66a3d0c38e551061aae518f6d4282361955c53d297ac635" Dec 11 08:15:05 crc kubenswrapper[4860]: I1211 08:15:05.709860 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-55jtz" Dec 11 08:15:05 crc kubenswrapper[4860]: I1211 08:15:05.717062 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mgfr4" event={"ID":"30107273-766a-4575-8cb1-3de387248437","Type":"ContainerStarted","Data":"14e1cbe7def3e65e27f664f13218ca12aabceb8a2a528dacbf116993fd9b73b9"} Dec 11 08:15:05 crc kubenswrapper[4860]: I1211 08:15:05.718946 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p6f75" event={"ID":"f9b23a65-98b8-4d83-a71f-2316fc1d46a0","Type":"ContainerDied","Data":"a8c240a8856f6d5e175157582fa03875d3ba8ea02a4fa6a5f8fdeefc96eea465"} Dec 11 08:15:05 crc kubenswrapper[4860]: I1211 08:15:05.718992 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p6f75" Dec 11 08:15:05 crc kubenswrapper[4860]: I1211 08:15:05.720023 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29424015-j4cd9" event={"ID":"a32ff4fb-1fd8-4097-88cf-74c67cf6fa56","Type":"ContainerStarted","Data":"d0289548ad4f995ea757eb89d9adc13ae624e1e3565fb874661e50fd36d76e30"} Dec 11 08:15:05 crc kubenswrapper[4860]: I1211 08:15:05.720131 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29424015-j4cd9" event={"ID":"a32ff4fb-1fd8-4097-88cf-74c67cf6fa56","Type":"ContainerStarted","Data":"c9ca262e760f24601d048784dc10ec9786221e4c38332ae1ab667e5ff3e2adcf"} Dec 11 08:15:05 crc kubenswrapper[4860]: I1211 08:15:05.726834 4860 generic.go:334] "Generic (PLEG): container finished" podID="294792dd-5104-4308-a10f-f3ab0c4504e9" containerID="fdf701a14de86401e1b762180ef4ada3e5a21e1a0c9001e9c146d6971a79594f" exitCode=0 Dec 11 08:15:05 crc kubenswrapper[4860]: I1211 08:15:05.726877 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dzg4x" event={"ID":"294792dd-5104-4308-a10f-f3ab0c4504e9","Type":"ContainerDied","Data":"fdf701a14de86401e1b762180ef4ada3e5a21e1a0c9001e9c146d6971a79594f"} Dec 11 08:15:05 crc kubenswrapper[4860]: I1211 08:15:05.732800 4860 scope.go:117] "RemoveContainer" containerID="c0c0e7688411dc667e3662bab7a995ed924b7b0081d1817049f17f8eddb6facb" Dec 11 08:15:05 crc kubenswrapper[4860]: I1211 08:15:05.770413 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-mgfr4" podStartSLOduration=3.730820095 podStartE2EDuration="1m20.77039006s" podCreationTimestamp="2025-12-11 08:13:45 +0000 UTC" firstStartedPulling="2025-12-11 08:13:47.924189947 +0000 UTC m=+160.652708992" lastFinishedPulling="2025-12-11 08:15:04.963759902 +0000 UTC m=+237.692278957" observedRunningTime="2025-12-11 08:15:05.768940801 +0000 UTC m=+238.497459856" watchObservedRunningTime="2025-12-11 08:15:05.77039006 +0000 UTC m=+238.498909125" Dec 11 08:15:05 crc kubenswrapper[4860]: I1211 08:15:05.788863 4860 scope.go:117] "RemoveContainer" containerID="0d6492f503343d818f1c697b3a863bfc2e6cf553475503ec2395cae61ed2e3ea" Dec 11 08:15:05 crc kubenswrapper[4860]: I1211 08:15:05.798379 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29424015-j4cd9" podStartSLOduration=5.798361622 podStartE2EDuration="5.798361622s" podCreationTimestamp="2025-12-11 08:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:15:05.789076866 +0000 UTC m=+238.517595921" watchObservedRunningTime="2025-12-11 08:15:05.798361622 +0000 UTC m=+238.526880667" Dec 11 08:15:05 crc kubenswrapper[4860]: I1211 08:15:05.800306 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-55jtz"] Dec 11 08:15:05 crc kubenswrapper[4860]: I1211 08:15:05.804492 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-55jtz"] Dec 11 08:15:05 crc kubenswrapper[4860]: I1211 08:15:05.816628 4860 scope.go:117] "RemoveContainer" containerID="c9c0c514837b2c179afb0c70eb50b7427f0b1cdfa4f763026250e8a797e2e8a0" Dec 11 08:15:05 crc kubenswrapper[4860]: I1211 08:15:05.816818 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-p6f75"] Dec 11 08:15:05 crc kubenswrapper[4860]: I1211 08:15:05.823386 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-p6f75"] Dec 11 08:15:05 crc kubenswrapper[4860]: I1211 08:15:05.843969 4860 scope.go:117] "RemoveContainer" containerID="d56b94ad7f779326751392fd4ea22b63b32b176c75e8d6387550d7393c40e173" Dec 11 08:15:05 crc kubenswrapper[4860]: I1211 08:15:05.858678 4860 scope.go:117] "RemoveContainer" containerID="e7891ca8b7cb19f5e3732a1f072123261d58a224b8793c6a7c0923fe1959b620" Dec 11 08:15:06 crc kubenswrapper[4860]: I1211 08:15:06.154832 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-mgfr4" Dec 11 08:15:06 crc kubenswrapper[4860]: I1211 08:15:06.154975 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-mgfr4" Dec 11 08:15:06 crc kubenswrapper[4860]: I1211 08:15:06.732609 4860 generic.go:334] "Generic (PLEG): container finished" podID="a32ff4fb-1fd8-4097-88cf-74c67cf6fa56" containerID="d0289548ad4f995ea757eb89d9adc13ae624e1e3565fb874661e50fd36d76e30" exitCode=0 Dec 11 08:15:06 crc kubenswrapper[4860]: I1211 08:15:06.732667 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29424015-j4cd9" event={"ID":"a32ff4fb-1fd8-4097-88cf-74c67cf6fa56","Type":"ContainerDied","Data":"d0289548ad4f995ea757eb89d9adc13ae624e1e3565fb874661e50fd36d76e30"} Dec 11 08:15:06 crc kubenswrapper[4860]: I1211 08:15:06.736091 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dzg4x" event={"ID":"294792dd-5104-4308-a10f-f3ab0c4504e9","Type":"ContainerStarted","Data":"8b5bd9e3bd9159c617215631e1bd2e8031d35214329adfb7419a18627358c68d"} Dec 11 08:15:06 crc kubenswrapper[4860]: I1211 08:15:06.737938 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dcccn" event={"ID":"8e052d7c-59d1-43a4-ac7a-e9daf2a88c65","Type":"ContainerStarted","Data":"f96f8c23423f6d30e831662d312b1f6086c55082925cb0383a915df10a158b79"} Dec 11 08:15:06 crc kubenswrapper[4860]: I1211 08:15:06.776607 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-dzg4x" podStartSLOduration=2.112340443 podStartE2EDuration="1m22.776589617s" podCreationTimestamp="2025-12-11 08:13:44 +0000 UTC" firstStartedPulling="2025-12-11 08:13:45.809831542 +0000 UTC m=+158.538350597" lastFinishedPulling="2025-12-11 08:15:06.474080716 +0000 UTC m=+239.202599771" observedRunningTime="2025-12-11 08:15:06.773186522 +0000 UTC m=+239.501705577" watchObservedRunningTime="2025-12-11 08:15:06.776589617 +0000 UTC m=+239.505108672" Dec 11 08:15:06 crc kubenswrapper[4860]: I1211 08:15:06.789603 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-dcccn" podStartSLOduration=2.50054013 podStartE2EDuration="1m22.789584479s" podCreationTimestamp="2025-12-11 08:13:44 +0000 UTC" firstStartedPulling="2025-12-11 08:13:45.823474959 +0000 UTC m=+158.551994014" lastFinishedPulling="2025-12-11 08:15:06.112519318 +0000 UTC m=+238.841038363" observedRunningTime="2025-12-11 08:15:06.788417469 +0000 UTC m=+239.516936534" watchObservedRunningTime="2025-12-11 08:15:06.789584479 +0000 UTC m=+239.518103534" Dec 11 08:15:07 crc kubenswrapper[4860]: I1211 08:15:07.196932 4860 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-mgfr4" podUID="30107273-766a-4575-8cb1-3de387248437" containerName="registry-server" probeResult="failure" output=< Dec 11 08:15:07 crc kubenswrapper[4860]: timeout: failed to connect service ":50051" within 1s Dec 11 08:15:07 crc kubenswrapper[4860]: > Dec 11 08:15:07 crc kubenswrapper[4860]: I1211 08:15:07.593415 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b498b689-633b-4fea-9981-159362f44053" path="/var/lib/kubelet/pods/b498b689-633b-4fea-9981-159362f44053/volumes" Dec 11 08:15:07 crc kubenswrapper[4860]: I1211 08:15:07.594362 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f9b23a65-98b8-4d83-a71f-2316fc1d46a0" path="/var/lib/kubelet/pods/f9b23a65-98b8-4d83-a71f-2316fc1d46a0/volumes" Dec 11 08:15:08 crc kubenswrapper[4860]: I1211 08:15:08.030423 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29424015-j4cd9" Dec 11 08:15:08 crc kubenswrapper[4860]: I1211 08:15:08.212552 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a32ff4fb-1fd8-4097-88cf-74c67cf6fa56-secret-volume\") pod \"a32ff4fb-1fd8-4097-88cf-74c67cf6fa56\" (UID: \"a32ff4fb-1fd8-4097-88cf-74c67cf6fa56\") " Dec 11 08:15:08 crc kubenswrapper[4860]: I1211 08:15:08.212628 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a32ff4fb-1fd8-4097-88cf-74c67cf6fa56-config-volume\") pod \"a32ff4fb-1fd8-4097-88cf-74c67cf6fa56\" (UID: \"a32ff4fb-1fd8-4097-88cf-74c67cf6fa56\") " Dec 11 08:15:08 crc kubenswrapper[4860]: I1211 08:15:08.212737 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dhmlx\" (UniqueName: \"kubernetes.io/projected/a32ff4fb-1fd8-4097-88cf-74c67cf6fa56-kube-api-access-dhmlx\") pod \"a32ff4fb-1fd8-4097-88cf-74c67cf6fa56\" (UID: \"a32ff4fb-1fd8-4097-88cf-74c67cf6fa56\") " Dec 11 08:15:08 crc kubenswrapper[4860]: I1211 08:15:08.213664 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a32ff4fb-1fd8-4097-88cf-74c67cf6fa56-config-volume" (OuterVolumeSpecName: "config-volume") pod "a32ff4fb-1fd8-4097-88cf-74c67cf6fa56" (UID: "a32ff4fb-1fd8-4097-88cf-74c67cf6fa56"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:15:08 crc kubenswrapper[4860]: I1211 08:15:08.217961 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a32ff4fb-1fd8-4097-88cf-74c67cf6fa56-kube-api-access-dhmlx" (OuterVolumeSpecName: "kube-api-access-dhmlx") pod "a32ff4fb-1fd8-4097-88cf-74c67cf6fa56" (UID: "a32ff4fb-1fd8-4097-88cf-74c67cf6fa56"). InnerVolumeSpecName "kube-api-access-dhmlx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:15:08 crc kubenswrapper[4860]: I1211 08:15:08.221427 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a32ff4fb-1fd8-4097-88cf-74c67cf6fa56-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "a32ff4fb-1fd8-4097-88cf-74c67cf6fa56" (UID: "a32ff4fb-1fd8-4097-88cf-74c67cf6fa56"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:15:08 crc kubenswrapper[4860]: I1211 08:15:08.314520 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dhmlx\" (UniqueName: \"kubernetes.io/projected/a32ff4fb-1fd8-4097-88cf-74c67cf6fa56-kube-api-access-dhmlx\") on node \"crc\" DevicePath \"\"" Dec 11 08:15:08 crc kubenswrapper[4860]: I1211 08:15:08.314562 4860 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a32ff4fb-1fd8-4097-88cf-74c67cf6fa56-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 11 08:15:08 crc kubenswrapper[4860]: I1211 08:15:08.314573 4860 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a32ff4fb-1fd8-4097-88cf-74c67cf6fa56-config-volume\") on node \"crc\" DevicePath \"\"" Dec 11 08:15:08 crc kubenswrapper[4860]: I1211 08:15:08.752360 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29424015-j4cd9" event={"ID":"a32ff4fb-1fd8-4097-88cf-74c67cf6fa56","Type":"ContainerDied","Data":"c9ca262e760f24601d048784dc10ec9786221e4c38332ae1ab667e5ff3e2adcf"} Dec 11 08:15:08 crc kubenswrapper[4860]: I1211 08:15:08.752589 4860 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c9ca262e760f24601d048784dc10ec9786221e4c38332ae1ab667e5ff3e2adcf" Dec 11 08:15:08 crc kubenswrapper[4860]: I1211 08:15:08.752399 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29424015-j4cd9" Dec 11 08:15:14 crc kubenswrapper[4860]: I1211 08:15:14.550491 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-dzg4x" Dec 11 08:15:14 crc kubenswrapper[4860]: I1211 08:15:14.550859 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-dzg4x" Dec 11 08:15:14 crc kubenswrapper[4860]: I1211 08:15:14.605210 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-dzg4x" Dec 11 08:15:14 crc kubenswrapper[4860]: I1211 08:15:14.822299 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-dzg4x" Dec 11 08:15:14 crc kubenswrapper[4860]: I1211 08:15:14.939148 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-dcccn" Dec 11 08:15:14 crc kubenswrapper[4860]: I1211 08:15:14.939584 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-dcccn" Dec 11 08:15:14 crc kubenswrapper[4860]: I1211 08:15:14.980123 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-dcccn" Dec 11 08:15:15 crc kubenswrapper[4860]: I1211 08:15:15.011112 4860 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 11 08:15:15 crc kubenswrapper[4860]: E1211 08:15:15.011403 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9b23a65-98b8-4d83-a71f-2316fc1d46a0" containerName="registry-server" Dec 11 08:15:15 crc kubenswrapper[4860]: I1211 08:15:15.011419 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9b23a65-98b8-4d83-a71f-2316fc1d46a0" containerName="registry-server" Dec 11 08:15:15 crc kubenswrapper[4860]: E1211 08:15:15.011433 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a32ff4fb-1fd8-4097-88cf-74c67cf6fa56" containerName="collect-profiles" Dec 11 08:15:15 crc kubenswrapper[4860]: I1211 08:15:15.011442 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="a32ff4fb-1fd8-4097-88cf-74c67cf6fa56" containerName="collect-profiles" Dec 11 08:15:15 crc kubenswrapper[4860]: E1211 08:15:15.011459 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9b23a65-98b8-4d83-a71f-2316fc1d46a0" containerName="extract-utilities" Dec 11 08:15:15 crc kubenswrapper[4860]: I1211 08:15:15.011468 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9b23a65-98b8-4d83-a71f-2316fc1d46a0" containerName="extract-utilities" Dec 11 08:15:15 crc kubenswrapper[4860]: E1211 08:15:15.011482 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9b23a65-98b8-4d83-a71f-2316fc1d46a0" containerName="extract-content" Dec 11 08:15:15 crc kubenswrapper[4860]: I1211 08:15:15.011490 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9b23a65-98b8-4d83-a71f-2316fc1d46a0" containerName="extract-content" Dec 11 08:15:15 crc kubenswrapper[4860]: E1211 08:15:15.011500 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b498b689-633b-4fea-9981-159362f44053" containerName="registry-server" Dec 11 08:15:15 crc kubenswrapper[4860]: I1211 08:15:15.011509 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="b498b689-633b-4fea-9981-159362f44053" containerName="registry-server" Dec 11 08:15:15 crc kubenswrapper[4860]: E1211 08:15:15.011521 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b498b689-633b-4fea-9981-159362f44053" containerName="extract-utilities" Dec 11 08:15:15 crc kubenswrapper[4860]: I1211 08:15:15.011529 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="b498b689-633b-4fea-9981-159362f44053" containerName="extract-utilities" Dec 11 08:15:15 crc kubenswrapper[4860]: E1211 08:15:15.011543 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b498b689-633b-4fea-9981-159362f44053" containerName="extract-content" Dec 11 08:15:15 crc kubenswrapper[4860]: I1211 08:15:15.011551 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="b498b689-633b-4fea-9981-159362f44053" containerName="extract-content" Dec 11 08:15:15 crc kubenswrapper[4860]: I1211 08:15:15.011712 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="a32ff4fb-1fd8-4097-88cf-74c67cf6fa56" containerName="collect-profiles" Dec 11 08:15:15 crc kubenswrapper[4860]: I1211 08:15:15.011735 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="b498b689-633b-4fea-9981-159362f44053" containerName="registry-server" Dec 11 08:15:15 crc kubenswrapper[4860]: I1211 08:15:15.011746 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="f9b23a65-98b8-4d83-a71f-2316fc1d46a0" containerName="registry-server" Dec 11 08:15:15 crc kubenswrapper[4860]: I1211 08:15:15.012143 4860 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 11 08:15:15 crc kubenswrapper[4860]: I1211 08:15:15.012320 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 11 08:15:15 crc kubenswrapper[4860]: I1211 08:15:15.012438 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://e831db1a15976d3967d56479689572c26f2599a766baa0402897a3008428c56f" gracePeriod=15 Dec 11 08:15:15 crc kubenswrapper[4860]: I1211 08:15:15.012461 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://34e91f615b76445ff3241a304bb2964fde59de7e8a8dfff6841b2e628a31e86d" gracePeriod=15 Dec 11 08:15:15 crc kubenswrapper[4860]: I1211 08:15:15.012505 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://f8caebd2f4dfcbb89c597029a07c335171889761d2d200206987537fb0a228db" gracePeriod=15 Dec 11 08:15:15 crc kubenswrapper[4860]: I1211 08:15:15.012495 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://5a498aa90f914e1f0370df01dc5ef174df113a2a16d86b2e3f24b19a1b99b4c0" gracePeriod=15 Dec 11 08:15:15 crc kubenswrapper[4860]: I1211 08:15:15.012536 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://f2274fb298ca7d2b37b060b4dbf9b166d39ba77a108b08edeeb4d67196e72cb3" gracePeriod=15 Dec 11 08:15:15 crc kubenswrapper[4860]: I1211 08:15:15.012892 4860 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 11 08:15:15 crc kubenswrapper[4860]: E1211 08:15:15.013056 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 11 08:15:15 crc kubenswrapper[4860]: I1211 08:15:15.013066 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 11 08:15:15 crc kubenswrapper[4860]: E1211 08:15:15.013077 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 11 08:15:15 crc kubenswrapper[4860]: I1211 08:15:15.013082 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 11 08:15:15 crc kubenswrapper[4860]: E1211 08:15:15.013090 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 11 08:15:15 crc kubenswrapper[4860]: I1211 08:15:15.013095 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 11 08:15:15 crc kubenswrapper[4860]: E1211 08:15:15.013109 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 11 08:15:15 crc kubenswrapper[4860]: I1211 08:15:15.013114 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 11 08:15:15 crc kubenswrapper[4860]: E1211 08:15:15.013126 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 11 08:15:15 crc kubenswrapper[4860]: I1211 08:15:15.013131 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 11 08:15:15 crc kubenswrapper[4860]: E1211 08:15:15.013137 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 11 08:15:15 crc kubenswrapper[4860]: I1211 08:15:15.013145 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 11 08:15:15 crc kubenswrapper[4860]: E1211 08:15:15.013155 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 11 08:15:15 crc kubenswrapper[4860]: I1211 08:15:15.013161 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 11 08:15:15 crc kubenswrapper[4860]: I1211 08:15:15.013261 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 11 08:15:15 crc kubenswrapper[4860]: I1211 08:15:15.013273 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 11 08:15:15 crc kubenswrapper[4860]: I1211 08:15:15.013284 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 11 08:15:15 crc kubenswrapper[4860]: I1211 08:15:15.013293 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 11 08:15:15 crc kubenswrapper[4860]: I1211 08:15:15.013306 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 11 08:15:15 crc kubenswrapper[4860]: I1211 08:15:15.013726 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 11 08:15:15 crc kubenswrapper[4860]: I1211 08:15:15.108442 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 08:15:15 crc kubenswrapper[4860]: I1211 08:15:15.108918 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 11 08:15:15 crc kubenswrapper[4860]: I1211 08:15:15.109019 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 11 08:15:15 crc kubenswrapper[4860]: I1211 08:15:15.109041 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 08:15:15 crc kubenswrapper[4860]: I1211 08:15:15.109062 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 11 08:15:15 crc kubenswrapper[4860]: I1211 08:15:15.109089 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 08:15:15 crc kubenswrapper[4860]: I1211 08:15:15.109112 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 11 08:15:15 crc kubenswrapper[4860]: I1211 08:15:15.109143 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 11 08:15:15 crc kubenswrapper[4860]: I1211 08:15:15.210377 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 11 08:15:15 crc kubenswrapper[4860]: I1211 08:15:15.210442 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 08:15:15 crc kubenswrapper[4860]: I1211 08:15:15.210467 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 11 08:15:15 crc kubenswrapper[4860]: I1211 08:15:15.210494 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 08:15:15 crc kubenswrapper[4860]: I1211 08:15:15.210518 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 11 08:15:15 crc kubenswrapper[4860]: I1211 08:15:15.210539 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 08:15:15 crc kubenswrapper[4860]: I1211 08:15:15.210583 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 11 08:15:15 crc kubenswrapper[4860]: I1211 08:15:15.210606 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 11 08:15:15 crc kubenswrapper[4860]: I1211 08:15:15.210616 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 11 08:15:15 crc kubenswrapper[4860]: I1211 08:15:15.210540 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 11 08:15:15 crc kubenswrapper[4860]: I1211 08:15:15.210552 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 11 08:15:15 crc kubenswrapper[4860]: I1211 08:15:15.210742 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 08:15:15 crc kubenswrapper[4860]: I1211 08:15:15.210783 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 08:15:15 crc kubenswrapper[4860]: I1211 08:15:15.210879 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 11 08:15:15 crc kubenswrapper[4860]: I1211 08:15:15.211010 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 11 08:15:15 crc kubenswrapper[4860]: I1211 08:15:15.211039 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 08:15:15 crc kubenswrapper[4860]: E1211 08:15:15.646412 4860 desired_state_of_world_populator.go:312] "Error processing volume" err="error processing PVC openshift-image-registry/crc-image-registry-storage: failed to fetch PVC from API server: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-image-registry/persistentvolumeclaims/crc-image-registry-storage\": dial tcp 38.102.83.169:6443: connect: connection refused" pod="openshift-image-registry/image-registry-697d97f7c8-zf572" volumeName="registry-storage" Dec 11 08:15:15 crc kubenswrapper[4860]: I1211 08:15:15.793400 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 11 08:15:15 crc kubenswrapper[4860]: I1211 08:15:15.795821 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 11 08:15:15 crc kubenswrapper[4860]: I1211 08:15:15.796726 4860 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="34e91f615b76445ff3241a304bb2964fde59de7e8a8dfff6841b2e628a31e86d" exitCode=0 Dec 11 08:15:15 crc kubenswrapper[4860]: I1211 08:15:15.796763 4860 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="f2274fb298ca7d2b37b060b4dbf9b166d39ba77a108b08edeeb4d67196e72cb3" exitCode=0 Dec 11 08:15:15 crc kubenswrapper[4860]: I1211 08:15:15.796773 4860 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="5a498aa90f914e1f0370df01dc5ef174df113a2a16d86b2e3f24b19a1b99b4c0" exitCode=0 Dec 11 08:15:15 crc kubenswrapper[4860]: I1211 08:15:15.796784 4860 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="f8caebd2f4dfcbb89c597029a07c335171889761d2d200206987537fb0a228db" exitCode=2 Dec 11 08:15:15 crc kubenswrapper[4860]: I1211 08:15:15.796837 4860 scope.go:117] "RemoveContainer" containerID="2090cab02ad74ddfd6bb57ff50eba35d18f56b0cde1c230fe011c01994b693bc" Dec 11 08:15:15 crc kubenswrapper[4860]: I1211 08:15:15.798625 4860 generic.go:334] "Generic (PLEG): container finished" podID="3bea0969-d146-4f96-8f02-b95bfe17aac3" containerID="7d866182c7851f4fab0b8fe46f5ac4afb5266a3665af4234c0509e1fb87b399f" exitCode=0 Dec 11 08:15:15 crc kubenswrapper[4860]: I1211 08:15:15.798672 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"3bea0969-d146-4f96-8f02-b95bfe17aac3","Type":"ContainerDied","Data":"7d866182c7851f4fab0b8fe46f5ac4afb5266a3665af4234c0509e1fb87b399f"} Dec 11 08:15:15 crc kubenswrapper[4860]: I1211 08:15:15.799467 4860 status_manager.go:851] "Failed to get status for pod" podUID="3bea0969-d146-4f96-8f02-b95bfe17aac3" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.169:6443: connect: connection refused" Dec 11 08:15:15 crc kubenswrapper[4860]: I1211 08:15:15.846343 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-dcccn" Dec 11 08:15:15 crc kubenswrapper[4860]: I1211 08:15:15.846976 4860 status_manager.go:851] "Failed to get status for pod" podUID="3bea0969-d146-4f96-8f02-b95bfe17aac3" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.169:6443: connect: connection refused" Dec 11 08:15:15 crc kubenswrapper[4860]: I1211 08:15:15.847334 4860 status_manager.go:851] "Failed to get status for pod" podUID="8e052d7c-59d1-43a4-ac7a-e9daf2a88c65" pod="openshift-marketplace/redhat-marketplace-dcccn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dcccn\": dial tcp 38.102.83.169:6443: connect: connection refused" Dec 11 08:15:15 crc kubenswrapper[4860]: E1211 08:15:15.888831 4860 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:15:15Z\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:15:15Z\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:15:15Z\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:15:15Z\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:34f522750c260aee8d7d3d8c16bba58727f5dfb964b4aecc8b09e3e6f7056f12\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:9acec1ab208005d77c0ac2722e15bf8620aff3b5c4ab7910d45b05a66d2bb912\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1628955991},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:178c13b6a1b34d5a4da4710d46305ff33fc30a390d065c0e2ba191c863238f9e\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:fcefccd5340edefa42f3ec04805e7514cbd84b40e2ad4f0542e25acb4897c5a4\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1232534877},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:b0004ff683181b2b57df13c0ffc42453e10a5dcb1789d938a3f18527b08412d6\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d68762606abc1a4575916f8aec19a1d1c4e07b5c88745bc46602ddbd3b20496c\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1202271579},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:be25e28aabd5a6e06b4df55e58fa4be426c96c57e3387969e0070e6058149d04\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:e6f1bca5d60a93ec9f9bd8ae305cd4ded3f62b2a51bbfdf59e056ea57c0c5b9f\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1154573130},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792}]}}\" for node \"crc\": Patch \"https://api-int.crc.testing:6443/api/v1/nodes/crc/status?timeout=10s\": dial tcp 38.102.83.169:6443: connect: connection refused" Dec 11 08:15:15 crc kubenswrapper[4860]: E1211 08:15:15.889487 4860 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.169:6443: connect: connection refused" Dec 11 08:15:15 crc kubenswrapper[4860]: E1211 08:15:15.889931 4860 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.169:6443: connect: connection refused" Dec 11 08:15:15 crc kubenswrapper[4860]: E1211 08:15:15.890180 4860 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.169:6443: connect: connection refused" Dec 11 08:15:15 crc kubenswrapper[4860]: E1211 08:15:15.890462 4860 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.169:6443: connect: connection refused" Dec 11 08:15:15 crc kubenswrapper[4860]: E1211 08:15:15.890481 4860 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 11 08:15:16 crc kubenswrapper[4860]: I1211 08:15:16.197053 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-mgfr4" Dec 11 08:15:16 crc kubenswrapper[4860]: I1211 08:15:16.197820 4860 status_manager.go:851] "Failed to get status for pod" podUID="3bea0969-d146-4f96-8f02-b95bfe17aac3" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.169:6443: connect: connection refused" Dec 11 08:15:16 crc kubenswrapper[4860]: I1211 08:15:16.198054 4860 status_manager.go:851] "Failed to get status for pod" podUID="8e052d7c-59d1-43a4-ac7a-e9daf2a88c65" pod="openshift-marketplace/redhat-marketplace-dcccn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dcccn\": dial tcp 38.102.83.169:6443: connect: connection refused" Dec 11 08:15:16 crc kubenswrapper[4860]: I1211 08:15:16.198375 4860 status_manager.go:851] "Failed to get status for pod" podUID="30107273-766a-4575-8cb1-3de387248437" pod="openshift-marketplace/redhat-operators-mgfr4" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-mgfr4\": dial tcp 38.102.83.169:6443: connect: connection refused" Dec 11 08:15:16 crc kubenswrapper[4860]: I1211 08:15:16.243151 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-mgfr4" Dec 11 08:15:16 crc kubenswrapper[4860]: I1211 08:15:16.243609 4860 status_manager.go:851] "Failed to get status for pod" podUID="3bea0969-d146-4f96-8f02-b95bfe17aac3" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.169:6443: connect: connection refused" Dec 11 08:15:16 crc kubenswrapper[4860]: I1211 08:15:16.243805 4860 status_manager.go:851] "Failed to get status for pod" podUID="8e052d7c-59d1-43a4-ac7a-e9daf2a88c65" pod="openshift-marketplace/redhat-marketplace-dcccn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dcccn\": dial tcp 38.102.83.169:6443: connect: connection refused" Dec 11 08:15:16 crc kubenswrapper[4860]: I1211 08:15:16.244070 4860 status_manager.go:851] "Failed to get status for pod" podUID="30107273-766a-4575-8cb1-3de387248437" pod="openshift-marketplace/redhat-operators-mgfr4" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-mgfr4\": dial tcp 38.102.83.169:6443: connect: connection refused" Dec 11 08:15:16 crc kubenswrapper[4860]: I1211 08:15:16.806071 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 11 08:15:17 crc kubenswrapper[4860]: I1211 08:15:17.046738 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 11 08:15:17 crc kubenswrapper[4860]: I1211 08:15:17.047264 4860 status_manager.go:851] "Failed to get status for pod" podUID="30107273-766a-4575-8cb1-3de387248437" pod="openshift-marketplace/redhat-operators-mgfr4" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-mgfr4\": dial tcp 38.102.83.169:6443: connect: connection refused" Dec 11 08:15:17 crc kubenswrapper[4860]: I1211 08:15:17.047552 4860 status_manager.go:851] "Failed to get status for pod" podUID="3bea0969-d146-4f96-8f02-b95bfe17aac3" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.169:6443: connect: connection refused" Dec 11 08:15:17 crc kubenswrapper[4860]: I1211 08:15:17.048000 4860 status_manager.go:851] "Failed to get status for pod" podUID="8e052d7c-59d1-43a4-ac7a-e9daf2a88c65" pod="openshift-marketplace/redhat-marketplace-dcccn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dcccn\": dial tcp 38.102.83.169:6443: connect: connection refused" Dec 11 08:15:17 crc kubenswrapper[4860]: I1211 08:15:17.136214 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3bea0969-d146-4f96-8f02-b95bfe17aac3-kubelet-dir\") pod \"3bea0969-d146-4f96-8f02-b95bfe17aac3\" (UID: \"3bea0969-d146-4f96-8f02-b95bfe17aac3\") " Dec 11 08:15:17 crc kubenswrapper[4860]: I1211 08:15:17.136585 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3bea0969-d146-4f96-8f02-b95bfe17aac3-kube-api-access\") pod \"3bea0969-d146-4f96-8f02-b95bfe17aac3\" (UID: \"3bea0969-d146-4f96-8f02-b95bfe17aac3\") " Dec 11 08:15:17 crc kubenswrapper[4860]: I1211 08:15:17.136708 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/3bea0969-d146-4f96-8f02-b95bfe17aac3-var-lock\") pod \"3bea0969-d146-4f96-8f02-b95bfe17aac3\" (UID: \"3bea0969-d146-4f96-8f02-b95bfe17aac3\") " Dec 11 08:15:17 crc kubenswrapper[4860]: I1211 08:15:17.137140 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3bea0969-d146-4f96-8f02-b95bfe17aac3-var-lock" (OuterVolumeSpecName: "var-lock") pod "3bea0969-d146-4f96-8f02-b95bfe17aac3" (UID: "3bea0969-d146-4f96-8f02-b95bfe17aac3"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 08:15:17 crc kubenswrapper[4860]: I1211 08:15:17.137187 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3bea0969-d146-4f96-8f02-b95bfe17aac3-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "3bea0969-d146-4f96-8f02-b95bfe17aac3" (UID: "3bea0969-d146-4f96-8f02-b95bfe17aac3"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 08:15:17 crc kubenswrapper[4860]: I1211 08:15:17.150799 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3bea0969-d146-4f96-8f02-b95bfe17aac3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "3bea0969-d146-4f96-8f02-b95bfe17aac3" (UID: "3bea0969-d146-4f96-8f02-b95bfe17aac3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:15:17 crc kubenswrapper[4860]: I1211 08:15:17.238131 4860 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/3bea0969-d146-4f96-8f02-b95bfe17aac3-var-lock\") on node \"crc\" DevicePath \"\"" Dec 11 08:15:17 crc kubenswrapper[4860]: I1211 08:15:17.238174 4860 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3bea0969-d146-4f96-8f02-b95bfe17aac3-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 11 08:15:17 crc kubenswrapper[4860]: I1211 08:15:17.238187 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3bea0969-d146-4f96-8f02-b95bfe17aac3-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 11 08:15:17 crc kubenswrapper[4860]: I1211 08:15:17.582275 4860 status_manager.go:851] "Failed to get status for pod" podUID="3bea0969-d146-4f96-8f02-b95bfe17aac3" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.169:6443: connect: connection refused" Dec 11 08:15:17 crc kubenswrapper[4860]: I1211 08:15:17.582720 4860 status_manager.go:851] "Failed to get status for pod" podUID="8e052d7c-59d1-43a4-ac7a-e9daf2a88c65" pod="openshift-marketplace/redhat-marketplace-dcccn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dcccn\": dial tcp 38.102.83.169:6443: connect: connection refused" Dec 11 08:15:17 crc kubenswrapper[4860]: I1211 08:15:17.582991 4860 status_manager.go:851] "Failed to get status for pod" podUID="30107273-766a-4575-8cb1-3de387248437" pod="openshift-marketplace/redhat-operators-mgfr4" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-mgfr4\": dial tcp 38.102.83.169:6443: connect: connection refused" Dec 11 08:15:17 crc kubenswrapper[4860]: I1211 08:15:17.818236 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 11 08:15:17 crc kubenswrapper[4860]: I1211 08:15:17.819060 4860 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="e831db1a15976d3967d56479689572c26f2599a766baa0402897a3008428c56f" exitCode=0 Dec 11 08:15:17 crc kubenswrapper[4860]: I1211 08:15:17.822629 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"3bea0969-d146-4f96-8f02-b95bfe17aac3","Type":"ContainerDied","Data":"73d4aed6dce133a753fe1f5a48696f5eeefba1694e560471a71c2b08e8f01b48"} Dec 11 08:15:17 crc kubenswrapper[4860]: I1211 08:15:17.822705 4860 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="73d4aed6dce133a753fe1f5a48696f5eeefba1694e560471a71c2b08e8f01b48" Dec 11 08:15:17 crc kubenswrapper[4860]: I1211 08:15:17.822704 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 11 08:15:17 crc kubenswrapper[4860]: I1211 08:15:17.827389 4860 status_manager.go:851] "Failed to get status for pod" podUID="30107273-766a-4575-8cb1-3de387248437" pod="openshift-marketplace/redhat-operators-mgfr4" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-mgfr4\": dial tcp 38.102.83.169:6443: connect: connection refused" Dec 11 08:15:17 crc kubenswrapper[4860]: I1211 08:15:17.828113 4860 status_manager.go:851] "Failed to get status for pod" podUID="3bea0969-d146-4f96-8f02-b95bfe17aac3" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.169:6443: connect: connection refused" Dec 11 08:15:17 crc kubenswrapper[4860]: I1211 08:15:17.829601 4860 status_manager.go:851] "Failed to get status for pod" podUID="8e052d7c-59d1-43a4-ac7a-e9daf2a88c65" pod="openshift-marketplace/redhat-marketplace-dcccn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dcccn\": dial tcp 38.102.83.169:6443: connect: connection refused" Dec 11 08:15:17 crc kubenswrapper[4860]: I1211 08:15:17.883797 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 11 08:15:17 crc kubenswrapper[4860]: I1211 08:15:17.884993 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 08:15:17 crc kubenswrapper[4860]: I1211 08:15:17.885563 4860 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.169:6443: connect: connection refused" Dec 11 08:15:17 crc kubenswrapper[4860]: I1211 08:15:17.885963 4860 status_manager.go:851] "Failed to get status for pod" podUID="3bea0969-d146-4f96-8f02-b95bfe17aac3" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.169:6443: connect: connection refused" Dec 11 08:15:17 crc kubenswrapper[4860]: I1211 08:15:17.886244 4860 status_manager.go:851] "Failed to get status for pod" podUID="8e052d7c-59d1-43a4-ac7a-e9daf2a88c65" pod="openshift-marketplace/redhat-marketplace-dcccn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dcccn\": dial tcp 38.102.83.169:6443: connect: connection refused" Dec 11 08:15:17 crc kubenswrapper[4860]: I1211 08:15:17.886530 4860 status_manager.go:851] "Failed to get status for pod" podUID="30107273-766a-4575-8cb1-3de387248437" pod="openshift-marketplace/redhat-operators-mgfr4" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-mgfr4\": dial tcp 38.102.83.169:6443: connect: connection refused" Dec 11 08:15:18 crc kubenswrapper[4860]: I1211 08:15:18.046999 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 11 08:15:18 crc kubenswrapper[4860]: I1211 08:15:18.047131 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 11 08:15:18 crc kubenswrapper[4860]: I1211 08:15:18.047171 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 11 08:15:18 crc kubenswrapper[4860]: I1211 08:15:18.047167 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 08:15:18 crc kubenswrapper[4860]: I1211 08:15:18.047189 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 08:15:18 crc kubenswrapper[4860]: I1211 08:15:18.047202 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 08:15:18 crc kubenswrapper[4860]: I1211 08:15:18.047469 4860 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 11 08:15:18 crc kubenswrapper[4860]: I1211 08:15:18.047491 4860 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Dec 11 08:15:18 crc kubenswrapper[4860]: I1211 08:15:18.047504 4860 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 11 08:15:18 crc kubenswrapper[4860]: I1211 08:15:18.832006 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 11 08:15:18 crc kubenswrapper[4860]: I1211 08:15:18.832914 4860 scope.go:117] "RemoveContainer" containerID="34e91f615b76445ff3241a304bb2964fde59de7e8a8dfff6841b2e628a31e86d" Dec 11 08:15:18 crc kubenswrapper[4860]: I1211 08:15:18.833081 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 08:15:18 crc kubenswrapper[4860]: I1211 08:15:18.850497 4860 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.169:6443: connect: connection refused" Dec 11 08:15:18 crc kubenswrapper[4860]: I1211 08:15:18.850711 4860 status_manager.go:851] "Failed to get status for pod" podUID="3bea0969-d146-4f96-8f02-b95bfe17aac3" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.169:6443: connect: connection refused" Dec 11 08:15:18 crc kubenswrapper[4860]: I1211 08:15:18.850858 4860 status_manager.go:851] "Failed to get status for pod" podUID="8e052d7c-59d1-43a4-ac7a-e9daf2a88c65" pod="openshift-marketplace/redhat-marketplace-dcccn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dcccn\": dial tcp 38.102.83.169:6443: connect: connection refused" Dec 11 08:15:18 crc kubenswrapper[4860]: I1211 08:15:18.850991 4860 status_manager.go:851] "Failed to get status for pod" podUID="30107273-766a-4575-8cb1-3de387248437" pod="openshift-marketplace/redhat-operators-mgfr4" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-mgfr4\": dial tcp 38.102.83.169:6443: connect: connection refused" Dec 11 08:15:18 crc kubenswrapper[4860]: I1211 08:15:18.870274 4860 scope.go:117] "RemoveContainer" containerID="f2274fb298ca7d2b37b060b4dbf9b166d39ba77a108b08edeeb4d67196e72cb3" Dec 11 08:15:18 crc kubenswrapper[4860]: I1211 08:15:18.889518 4860 scope.go:117] "RemoveContainer" containerID="5a498aa90f914e1f0370df01dc5ef174df113a2a16d86b2e3f24b19a1b99b4c0" Dec 11 08:15:18 crc kubenswrapper[4860]: I1211 08:15:18.903878 4860 scope.go:117] "RemoveContainer" containerID="f8caebd2f4dfcbb89c597029a07c335171889761d2d200206987537fb0a228db" Dec 11 08:15:18 crc kubenswrapper[4860]: I1211 08:15:18.921391 4860 scope.go:117] "RemoveContainer" containerID="e831db1a15976d3967d56479689572c26f2599a766baa0402897a3008428c56f" Dec 11 08:15:18 crc kubenswrapper[4860]: I1211 08:15:18.936322 4860 scope.go:117] "RemoveContainer" containerID="cf0aff8a6ab816b6617b1feeca9146d41bb5e8808bf1db96f8d493500c135d85" Dec 11 08:15:19 crc kubenswrapper[4860]: E1211 08:15:19.101633 4860 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.169:6443: connect: connection refused" Dec 11 08:15:19 crc kubenswrapper[4860]: E1211 08:15:19.101893 4860 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.169:6443: connect: connection refused" Dec 11 08:15:19 crc kubenswrapper[4860]: E1211 08:15:19.102229 4860 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.169:6443: connect: connection refused" Dec 11 08:15:19 crc kubenswrapper[4860]: E1211 08:15:19.102808 4860 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.169:6443: connect: connection refused" Dec 11 08:15:19 crc kubenswrapper[4860]: E1211 08:15:19.103312 4860 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.169:6443: connect: connection refused" Dec 11 08:15:19 crc kubenswrapper[4860]: I1211 08:15:19.103344 4860 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Dec 11 08:15:19 crc kubenswrapper[4860]: E1211 08:15:19.103582 4860 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.169:6443: connect: connection refused" interval="200ms" Dec 11 08:15:19 crc kubenswrapper[4860]: E1211 08:15:19.304386 4860 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.169:6443: connect: connection refused" interval="400ms" Dec 11 08:15:19 crc kubenswrapper[4860]: I1211 08:15:19.591706 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Dec 11 08:15:19 crc kubenswrapper[4860]: E1211 08:15:19.705179 4860 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.169:6443: connect: connection refused" interval="800ms" Dec 11 08:15:20 crc kubenswrapper[4860]: E1211 08:15:20.050997 4860 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.169:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 11 08:15:20 crc kubenswrapper[4860]: I1211 08:15:20.051894 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 11 08:15:20 crc kubenswrapper[4860]: W1211 08:15:20.076426 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf85e55b1a89d02b0cb034b1ea31ed45a.slice/crio-5820dbbb32034a7ea1923819e6166dfe8c167ca0ee792f6cd781096c5aec56f9 WatchSource:0}: Error finding container 5820dbbb32034a7ea1923819e6166dfe8c167ca0ee792f6cd781096c5aec56f9: Status 404 returned error can't find the container with id 5820dbbb32034a7ea1923819e6166dfe8c167ca0ee792f6cd781096c5aec56f9 Dec 11 08:15:20 crc kubenswrapper[4860]: E1211 08:15:20.082220 4860 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.169:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.18801b29c162feb5 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 08:15:20.081333941 +0000 UTC m=+252.809852996,LastTimestamp:2025-12-11 08:15:20.081333941 +0000 UTC m=+252.809852996,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 08:15:20 crc kubenswrapper[4860]: E1211 08:15:20.506326 4860 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.169:6443: connect: connection refused" interval="1.6s" Dec 11 08:15:20 crc kubenswrapper[4860]: I1211 08:15:20.848213 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"7c8d32000cd3750bd5ea7b6fd6bc17c8db195e315089cd5a1ed0023342f049aa"} Dec 11 08:15:20 crc kubenswrapper[4860]: I1211 08:15:20.848515 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"5820dbbb32034a7ea1923819e6166dfe8c167ca0ee792f6cd781096c5aec56f9"} Dec 11 08:15:20 crc kubenswrapper[4860]: E1211 08:15:20.849044 4860 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.169:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 11 08:15:20 crc kubenswrapper[4860]: I1211 08:15:20.849048 4860 status_manager.go:851] "Failed to get status for pod" podUID="3bea0969-d146-4f96-8f02-b95bfe17aac3" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.169:6443: connect: connection refused" Dec 11 08:15:20 crc kubenswrapper[4860]: I1211 08:15:20.849494 4860 status_manager.go:851] "Failed to get status for pod" podUID="8e052d7c-59d1-43a4-ac7a-e9daf2a88c65" pod="openshift-marketplace/redhat-marketplace-dcccn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dcccn\": dial tcp 38.102.83.169:6443: connect: connection refused" Dec 11 08:15:20 crc kubenswrapper[4860]: I1211 08:15:20.849693 4860 status_manager.go:851] "Failed to get status for pod" podUID="30107273-766a-4575-8cb1-3de387248437" pod="openshift-marketplace/redhat-operators-mgfr4" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-mgfr4\": dial tcp 38.102.83.169:6443: connect: connection refused" Dec 11 08:15:22 crc kubenswrapper[4860]: E1211 08:15:22.108290 4860 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.169:6443: connect: connection refused" interval="3.2s" Dec 11 08:15:22 crc kubenswrapper[4860]: E1211 08:15:22.757209 4860 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.169:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.18801b29c162feb5 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 08:15:20.081333941 +0000 UTC m=+252.809852996,LastTimestamp:2025-12-11 08:15:20.081333941 +0000 UTC m=+252.809852996,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 08:15:25 crc kubenswrapper[4860]: E1211 08:15:25.309223 4860 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.169:6443: connect: connection refused" interval="6.4s" Dec 11 08:15:26 crc kubenswrapper[4860]: E1211 08:15:26.027703 4860 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:15:26Z\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:15:26Z\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:15:26Z\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T08:15:26Z\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:34f522750c260aee8d7d3d8c16bba58727f5dfb964b4aecc8b09e3e6f7056f12\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:9acec1ab208005d77c0ac2722e15bf8620aff3b5c4ab7910d45b05a66d2bb912\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1628955991},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:178c13b6a1b34d5a4da4710d46305ff33fc30a390d065c0e2ba191c863238f9e\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:fcefccd5340edefa42f3ec04805e7514cbd84b40e2ad4f0542e25acb4897c5a4\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1232534877},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:b0004ff683181b2b57df13c0ffc42453e10a5dcb1789d938a3f18527b08412d6\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d68762606abc1a4575916f8aec19a1d1c4e07b5c88745bc46602ddbd3b20496c\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1202271579},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:be25e28aabd5a6e06b4df55e58fa4be426c96c57e3387969e0070e6058149d04\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:e6f1bca5d60a93ec9f9bd8ae305cd4ded3f62b2a51bbfdf59e056ea57c0c5b9f\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1154573130},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792}]}}\" for node \"crc\": Patch \"https://api-int.crc.testing:6443/api/v1/nodes/crc/status?timeout=10s\": dial tcp 38.102.83.169:6443: connect: connection refused" Dec 11 08:15:26 crc kubenswrapper[4860]: E1211 08:15:26.028317 4860 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.169:6443: connect: connection refused" Dec 11 08:15:26 crc kubenswrapper[4860]: E1211 08:15:26.028705 4860 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.169:6443: connect: connection refused" Dec 11 08:15:26 crc kubenswrapper[4860]: E1211 08:15:26.029220 4860 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.169:6443: connect: connection refused" Dec 11 08:15:26 crc kubenswrapper[4860]: E1211 08:15:26.029921 4860 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.169:6443: connect: connection refused" Dec 11 08:15:26 crc kubenswrapper[4860]: E1211 08:15:26.029960 4860 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 11 08:15:27 crc kubenswrapper[4860]: I1211 08:15:27.579106 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 08:15:27 crc kubenswrapper[4860]: I1211 08:15:27.583455 4860 status_manager.go:851] "Failed to get status for pod" podUID="3bea0969-d146-4f96-8f02-b95bfe17aac3" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.169:6443: connect: connection refused" Dec 11 08:15:27 crc kubenswrapper[4860]: I1211 08:15:27.583959 4860 status_manager.go:851] "Failed to get status for pod" podUID="8e052d7c-59d1-43a4-ac7a-e9daf2a88c65" pod="openshift-marketplace/redhat-marketplace-dcccn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dcccn\": dial tcp 38.102.83.169:6443: connect: connection refused" Dec 11 08:15:27 crc kubenswrapper[4860]: I1211 08:15:27.584635 4860 status_manager.go:851] "Failed to get status for pod" podUID="30107273-766a-4575-8cb1-3de387248437" pod="openshift-marketplace/redhat-operators-mgfr4" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-mgfr4\": dial tcp 38.102.83.169:6443: connect: connection refused" Dec 11 08:15:27 crc kubenswrapper[4860]: I1211 08:15:27.585768 4860 status_manager.go:851] "Failed to get status for pod" podUID="3bea0969-d146-4f96-8f02-b95bfe17aac3" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.169:6443: connect: connection refused" Dec 11 08:15:27 crc kubenswrapper[4860]: I1211 08:15:27.586168 4860 status_manager.go:851] "Failed to get status for pod" podUID="8e052d7c-59d1-43a4-ac7a-e9daf2a88c65" pod="openshift-marketplace/redhat-marketplace-dcccn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dcccn\": dial tcp 38.102.83.169:6443: connect: connection refused" Dec 11 08:15:27 crc kubenswrapper[4860]: I1211 08:15:27.586728 4860 status_manager.go:851] "Failed to get status for pod" podUID="30107273-766a-4575-8cb1-3de387248437" pod="openshift-marketplace/redhat-operators-mgfr4" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-mgfr4\": dial tcp 38.102.83.169:6443: connect: connection refused" Dec 11 08:15:27 crc kubenswrapper[4860]: I1211 08:15:27.605903 4860 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e5ec3758-85f1-4a22-a4ec-c805472607a1" Dec 11 08:15:27 crc kubenswrapper[4860]: I1211 08:15:27.605946 4860 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e5ec3758-85f1-4a22-a4ec-c805472607a1" Dec 11 08:15:27 crc kubenswrapper[4860]: E1211 08:15:27.606459 4860 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.169:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 08:15:27 crc kubenswrapper[4860]: I1211 08:15:27.607155 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 08:15:27 crc kubenswrapper[4860]: W1211 08:15:27.636530 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71bb4a3aecc4ba5b26c4b7318770ce13.slice/crio-0424133265b8285e3bc77e9e3cd685ccb681f9954084b4a0e854b27103f42947 WatchSource:0}: Error finding container 0424133265b8285e3bc77e9e3cd685ccb681f9954084b4a0e854b27103f42947: Status 404 returned error can't find the container with id 0424133265b8285e3bc77e9e3cd685ccb681f9954084b4a0e854b27103f42947 Dec 11 08:15:27 crc kubenswrapper[4860]: I1211 08:15:27.898430 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"0424133265b8285e3bc77e9e3cd685ccb681f9954084b4a0e854b27103f42947"} Dec 11 08:15:28 crc kubenswrapper[4860]: I1211 08:15:28.905548 4860 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="7846cd90acaf6486110d73955b947387e65880b7c9b4021d281a5946542bba62" exitCode=0 Dec 11 08:15:28 crc kubenswrapper[4860]: I1211 08:15:28.905593 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"7846cd90acaf6486110d73955b947387e65880b7c9b4021d281a5946542bba62"} Dec 11 08:15:28 crc kubenswrapper[4860]: I1211 08:15:28.905866 4860 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e5ec3758-85f1-4a22-a4ec-c805472607a1" Dec 11 08:15:28 crc kubenswrapper[4860]: I1211 08:15:28.905882 4860 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e5ec3758-85f1-4a22-a4ec-c805472607a1" Dec 11 08:15:28 crc kubenswrapper[4860]: I1211 08:15:28.906256 4860 status_manager.go:851] "Failed to get status for pod" podUID="3bea0969-d146-4f96-8f02-b95bfe17aac3" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.169:6443: connect: connection refused" Dec 11 08:15:28 crc kubenswrapper[4860]: E1211 08:15:28.906417 4860 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.169:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 08:15:28 crc kubenswrapper[4860]: I1211 08:15:28.906633 4860 status_manager.go:851] "Failed to get status for pod" podUID="8e052d7c-59d1-43a4-ac7a-e9daf2a88c65" pod="openshift-marketplace/redhat-marketplace-dcccn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dcccn\": dial tcp 38.102.83.169:6443: connect: connection refused" Dec 11 08:15:28 crc kubenswrapper[4860]: I1211 08:15:28.907174 4860 status_manager.go:851] "Failed to get status for pod" podUID="30107273-766a-4575-8cb1-3de387248437" pod="openshift-marketplace/redhat-operators-mgfr4" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-mgfr4\": dial tcp 38.102.83.169:6443: connect: connection refused" Dec 11 08:15:29 crc kubenswrapper[4860]: I1211 08:15:29.590852 4860 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Readiness probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Dec 11 08:15:29 crc kubenswrapper[4860]: I1211 08:15:29.591116 4860 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Dec 11 08:15:29 crc kubenswrapper[4860]: I1211 08:15:29.916042 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 11 08:15:29 crc kubenswrapper[4860]: I1211 08:15:29.916087 4860 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="077aa1dd2fe5189ff60a0f2365e47a6a545cca0de7aa6379370908db4ea5838a" exitCode=1 Dec 11 08:15:29 crc kubenswrapper[4860]: I1211 08:15:29.916133 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"077aa1dd2fe5189ff60a0f2365e47a6a545cca0de7aa6379370908db4ea5838a"} Dec 11 08:15:29 crc kubenswrapper[4860]: I1211 08:15:29.916466 4860 scope.go:117] "RemoveContainer" containerID="077aa1dd2fe5189ff60a0f2365e47a6a545cca0de7aa6379370908db4ea5838a" Dec 11 08:15:29 crc kubenswrapper[4860]: I1211 08:15:29.923384 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"b232528a4fcbf09eca7ea6514467e22c7db4ac8cf49d2b517569e8af48e96af3"} Dec 11 08:15:29 crc kubenswrapper[4860]: I1211 08:15:29.923424 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"82c246e576a631895c9682f450f00418151ea90e52ccd676b913862b39c08f89"} Dec 11 08:15:29 crc kubenswrapper[4860]: I1211 08:15:29.923434 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"89b989624f1088fbbd5a28c878a724ff7d5bebfc04131b0582d28774151079bf"} Dec 11 08:15:29 crc kubenswrapper[4860]: I1211 08:15:29.923443 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"5d5655c39d33d544732103e242f78ff2095e84868a06641b3892e7859a38ee73"} Dec 11 08:15:30 crc kubenswrapper[4860]: I1211 08:15:30.932368 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 11 08:15:30 crc kubenswrapper[4860]: I1211 08:15:30.932705 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"85c2e4d7f46755df050f0f2f1218016d25e54a5241e76ddd840b25ace404577b"} Dec 11 08:15:30 crc kubenswrapper[4860]: I1211 08:15:30.936973 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"209b4b7d9d6115754b2c446783fadd2957314f442203ea4f3ca2b7cb984734f3"} Dec 11 08:15:30 crc kubenswrapper[4860]: I1211 08:15:30.937242 4860 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e5ec3758-85f1-4a22-a4ec-c805472607a1" Dec 11 08:15:30 crc kubenswrapper[4860]: I1211 08:15:30.937264 4860 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e5ec3758-85f1-4a22-a4ec-c805472607a1" Dec 11 08:15:30 crc kubenswrapper[4860]: I1211 08:15:30.937510 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 08:15:30 crc kubenswrapper[4860]: I1211 08:15:30.991378 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 11 08:15:31 crc kubenswrapper[4860]: I1211 08:15:31.371297 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 11 08:15:31 crc kubenswrapper[4860]: I1211 08:15:31.942008 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 11 08:15:32 crc kubenswrapper[4860]: I1211 08:15:32.607730 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 08:15:32 crc kubenswrapper[4860]: I1211 08:15:32.608064 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 08:15:32 crc kubenswrapper[4860]: I1211 08:15:32.612540 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 08:15:35 crc kubenswrapper[4860]: I1211 08:15:35.946898 4860 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 08:15:36 crc kubenswrapper[4860]: I1211 08:15:36.973607 4860 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e5ec3758-85f1-4a22-a4ec-c805472607a1" Dec 11 08:15:36 crc kubenswrapper[4860]: I1211 08:15:36.973652 4860 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e5ec3758-85f1-4a22-a4ec-c805472607a1" Dec 11 08:15:36 crc kubenswrapper[4860]: I1211 08:15:36.978180 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 08:15:37 crc kubenswrapper[4860]: I1211 08:15:37.611425 4860 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="ac59cdaa-9c25-434a-a1a7-2a76e950a8e7" Dec 11 08:15:37 crc kubenswrapper[4860]: I1211 08:15:37.985812 4860 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e5ec3758-85f1-4a22-a4ec-c805472607a1" Dec 11 08:15:37 crc kubenswrapper[4860]: I1211 08:15:37.985864 4860 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e5ec3758-85f1-4a22-a4ec-c805472607a1" Dec 11 08:15:37 crc kubenswrapper[4860]: I1211 08:15:37.990970 4860 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="ac59cdaa-9c25-434a-a1a7-2a76e950a8e7" Dec 11 08:15:45 crc kubenswrapper[4860]: I1211 08:15:45.700069 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 11 08:15:46 crc kubenswrapper[4860]: I1211 08:15:46.563972 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 11 08:15:46 crc kubenswrapper[4860]: I1211 08:15:46.768785 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 11 08:15:47 crc kubenswrapper[4860]: I1211 08:15:47.145408 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 11 08:15:47 crc kubenswrapper[4860]: I1211 08:15:47.489572 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 11 08:15:47 crc kubenswrapper[4860]: I1211 08:15:47.664674 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 11 08:15:47 crc kubenswrapper[4860]: I1211 08:15:47.769179 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 11 08:15:47 crc kubenswrapper[4860]: I1211 08:15:47.842113 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 11 08:15:48 crc kubenswrapper[4860]: I1211 08:15:48.094229 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 11 08:15:48 crc kubenswrapper[4860]: I1211 08:15:48.563587 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 11 08:15:48 crc kubenswrapper[4860]: I1211 08:15:48.751622 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 11 08:15:48 crc kubenswrapper[4860]: I1211 08:15:48.795954 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 11 08:15:48 crc kubenswrapper[4860]: I1211 08:15:48.808724 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 11 08:15:48 crc kubenswrapper[4860]: I1211 08:15:48.923789 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 11 08:15:48 crc kubenswrapper[4860]: I1211 08:15:48.926176 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 11 08:15:49 crc kubenswrapper[4860]: I1211 08:15:49.148450 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 11 08:15:49 crc kubenswrapper[4860]: I1211 08:15:49.150107 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 11 08:15:49 crc kubenswrapper[4860]: I1211 08:15:49.171697 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 11 08:15:49 crc kubenswrapper[4860]: I1211 08:15:49.258101 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 11 08:15:49 crc kubenswrapper[4860]: I1211 08:15:49.297675 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 11 08:15:49 crc kubenswrapper[4860]: I1211 08:15:49.346178 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 11 08:15:49 crc kubenswrapper[4860]: I1211 08:15:49.407942 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 11 08:15:49 crc kubenswrapper[4860]: I1211 08:15:49.416610 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 11 08:15:49 crc kubenswrapper[4860]: I1211 08:15:49.449115 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 11 08:15:49 crc kubenswrapper[4860]: I1211 08:15:49.455708 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 11 08:15:49 crc kubenswrapper[4860]: I1211 08:15:49.566196 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 11 08:15:49 crc kubenswrapper[4860]: I1211 08:15:49.577337 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 11 08:15:49 crc kubenswrapper[4860]: I1211 08:15:49.593564 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 11 08:15:49 crc kubenswrapper[4860]: I1211 08:15:49.594339 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 11 08:15:49 crc kubenswrapper[4860]: I1211 08:15:49.655098 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 11 08:15:49 crc kubenswrapper[4860]: I1211 08:15:49.668493 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 11 08:15:49 crc kubenswrapper[4860]: I1211 08:15:49.714759 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 11 08:15:49 crc kubenswrapper[4860]: I1211 08:15:49.801780 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 11 08:15:49 crc kubenswrapper[4860]: I1211 08:15:49.861720 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 11 08:15:49 crc kubenswrapper[4860]: I1211 08:15:49.862959 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 11 08:15:49 crc kubenswrapper[4860]: I1211 08:15:49.887725 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 11 08:15:49 crc kubenswrapper[4860]: I1211 08:15:49.938827 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 11 08:15:50 crc kubenswrapper[4860]: I1211 08:15:50.179079 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 11 08:15:50 crc kubenswrapper[4860]: I1211 08:15:50.359244 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 11 08:15:50 crc kubenswrapper[4860]: I1211 08:15:50.502938 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 11 08:15:50 crc kubenswrapper[4860]: I1211 08:15:50.648127 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 11 08:15:50 crc kubenswrapper[4860]: I1211 08:15:50.740456 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 11 08:15:50 crc kubenswrapper[4860]: I1211 08:15:50.741777 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 11 08:15:50 crc kubenswrapper[4860]: I1211 08:15:50.787153 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 11 08:15:50 crc kubenswrapper[4860]: I1211 08:15:50.814479 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 11 08:15:51 crc kubenswrapper[4860]: I1211 08:15:51.034553 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 11 08:15:51 crc kubenswrapper[4860]: I1211 08:15:51.042326 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 11 08:15:51 crc kubenswrapper[4860]: I1211 08:15:51.080266 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 11 08:15:51 crc kubenswrapper[4860]: I1211 08:15:51.090026 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 11 08:15:51 crc kubenswrapper[4860]: I1211 08:15:51.130767 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 11 08:15:51 crc kubenswrapper[4860]: I1211 08:15:51.133900 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 11 08:15:51 crc kubenswrapper[4860]: I1211 08:15:51.258260 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 11 08:15:51 crc kubenswrapper[4860]: I1211 08:15:51.263691 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 11 08:15:51 crc kubenswrapper[4860]: I1211 08:15:51.388151 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 11 08:15:51 crc kubenswrapper[4860]: I1211 08:15:51.444913 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 11 08:15:51 crc kubenswrapper[4860]: I1211 08:15:51.473608 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 11 08:15:51 crc kubenswrapper[4860]: I1211 08:15:51.531998 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 11 08:15:51 crc kubenswrapper[4860]: I1211 08:15:51.545781 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 11 08:15:51 crc kubenswrapper[4860]: I1211 08:15:51.598637 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 11 08:15:51 crc kubenswrapper[4860]: I1211 08:15:51.630315 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 11 08:15:51 crc kubenswrapper[4860]: I1211 08:15:51.658542 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 11 08:15:51 crc kubenswrapper[4860]: I1211 08:15:51.709355 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 11 08:15:51 crc kubenswrapper[4860]: I1211 08:15:51.752904 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 11 08:15:51 crc kubenswrapper[4860]: I1211 08:15:51.765589 4860 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 11 08:15:51 crc kubenswrapper[4860]: I1211 08:15:51.766737 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 11 08:15:51 crc kubenswrapper[4860]: I1211 08:15:51.903914 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 11 08:15:51 crc kubenswrapper[4860]: I1211 08:15:51.910555 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 11 08:15:51 crc kubenswrapper[4860]: I1211 08:15:51.974735 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 11 08:15:52 crc kubenswrapper[4860]: I1211 08:15:52.039906 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 11 08:15:52 crc kubenswrapper[4860]: I1211 08:15:52.100891 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 11 08:15:52 crc kubenswrapper[4860]: I1211 08:15:52.216601 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 11 08:15:52 crc kubenswrapper[4860]: I1211 08:15:52.235277 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 11 08:15:52 crc kubenswrapper[4860]: I1211 08:15:52.246894 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 11 08:15:52 crc kubenswrapper[4860]: I1211 08:15:52.250041 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 11 08:15:52 crc kubenswrapper[4860]: I1211 08:15:52.255032 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 11 08:15:52 crc kubenswrapper[4860]: I1211 08:15:52.282728 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 11 08:15:52 crc kubenswrapper[4860]: I1211 08:15:52.310204 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 11 08:15:52 crc kubenswrapper[4860]: I1211 08:15:52.336150 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 11 08:15:52 crc kubenswrapper[4860]: I1211 08:15:52.344311 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 11 08:15:52 crc kubenswrapper[4860]: I1211 08:15:52.348430 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 11 08:15:52 crc kubenswrapper[4860]: I1211 08:15:52.400395 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 11 08:15:52 crc kubenswrapper[4860]: I1211 08:15:52.503416 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 11 08:15:52 crc kubenswrapper[4860]: I1211 08:15:52.630131 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 11 08:15:52 crc kubenswrapper[4860]: I1211 08:15:52.694496 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 11 08:15:52 crc kubenswrapper[4860]: I1211 08:15:52.724432 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 11 08:15:52 crc kubenswrapper[4860]: I1211 08:15:52.768069 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 11 08:15:52 crc kubenswrapper[4860]: I1211 08:15:52.792865 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 11 08:15:52 crc kubenswrapper[4860]: I1211 08:15:52.820640 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 11 08:15:52 crc kubenswrapper[4860]: I1211 08:15:52.829508 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 11 08:15:52 crc kubenswrapper[4860]: I1211 08:15:52.872051 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 11 08:15:53 crc kubenswrapper[4860]: I1211 08:15:53.038821 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 11 08:15:53 crc kubenswrapper[4860]: I1211 08:15:53.045877 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 11 08:15:53 crc kubenswrapper[4860]: I1211 08:15:53.048978 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 11 08:15:53 crc kubenswrapper[4860]: I1211 08:15:53.089970 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 11 08:15:53 crc kubenswrapper[4860]: I1211 08:15:53.170899 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 11 08:15:53 crc kubenswrapper[4860]: I1211 08:15:53.185750 4860 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 11 08:15:53 crc kubenswrapper[4860]: I1211 08:15:53.191004 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 11 08:15:53 crc kubenswrapper[4860]: I1211 08:15:53.191066 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc","openshift-marketplace/marketplace-operator-79b997595-trp75"] Dec 11 08:15:53 crc kubenswrapper[4860]: E1211 08:15:53.191320 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3bea0969-d146-4f96-8f02-b95bfe17aac3" containerName="installer" Dec 11 08:15:53 crc kubenswrapper[4860]: I1211 08:15:53.191343 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="3bea0969-d146-4f96-8f02-b95bfe17aac3" containerName="installer" Dec 11 08:15:53 crc kubenswrapper[4860]: I1211 08:15:53.191552 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="3bea0969-d146-4f96-8f02-b95bfe17aac3" containerName="installer" Dec 11 08:15:53 crc kubenswrapper[4860]: I1211 08:15:53.191985 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dzg4x","openshift-marketplace/community-operators-pqtvk","openshift-marketplace/certified-operators-kqwgm","openshift-marketplace/redhat-operators-mgfr4","openshift-marketplace/redhat-marketplace-dcccn","openshift-marketplace/marketplace-operator-79b997595-ttrm2","openshift-marketplace/redhat-operators-8p4mb"] Dec 11 08:15:53 crc kubenswrapper[4860]: I1211 08:15:53.192148 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-trp75" Dec 11 08:15:53 crc kubenswrapper[4860]: I1211 08:15:53.192239 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-dzg4x" podUID="294792dd-5104-4308-a10f-f3ab0c4504e9" containerName="registry-server" containerID="cri-o://8b5bd9e3bd9159c617215631e1bd2e8031d35214329adfb7419a18627358c68d" gracePeriod=30 Dec 11 08:15:53 crc kubenswrapper[4860]: I1211 08:15:53.192459 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-ttrm2" podUID="ad724960-18d1-4e83-944c-4cdeb033a436" containerName="marketplace-operator" containerID="cri-o://cafda173280b05a3de720d8e7f8927fc27b4c53b7500ba9deb3878e2112d3d8c" gracePeriod=30 Dec 11 08:15:53 crc kubenswrapper[4860]: I1211 08:15:53.192689 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-8p4mb" podUID="01481957-a690-40f4-8c89-352b7350d327" containerName="registry-server" containerID="cri-o://d9777d9cd9cfe1b753ed0336374b4591e9f9e22227f397713e583f8e99a68df8" gracePeriod=30 Dec 11 08:15:53 crc kubenswrapper[4860]: I1211 08:15:53.192627 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-dcccn" podUID="8e052d7c-59d1-43a4-ac7a-e9daf2a88c65" containerName="registry-server" containerID="cri-o://f96f8c23423f6d30e831662d312b1f6086c55082925cb0383a915df10a158b79" gracePeriod=30 Dec 11 08:15:53 crc kubenswrapper[4860]: I1211 08:15:53.192809 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-kqwgm" podUID="093a9903-7041-410b-bfb2-c74ef741e0be" containerName="registry-server" containerID="cri-o://41c0f2c9f5203be9afc00992cd7768f8a3896e585a81b5618856b40455c598eb" gracePeriod=30 Dec 11 08:15:53 crc kubenswrapper[4860]: I1211 08:15:53.192902 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-mgfr4" podUID="30107273-766a-4575-8cb1-3de387248437" containerName="registry-server" containerID="cri-o://14e1cbe7def3e65e27f664f13218ca12aabceb8a2a528dacbf116993fd9b73b9" gracePeriod=30 Dec 11 08:15:53 crc kubenswrapper[4860]: I1211 08:15:53.192994 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-pqtvk" podUID="7ac89442-1754-45d3-b67a-aa6cc31a1235" containerName="registry-server" containerID="cri-o://f1a5ae4b9e0d2bfe7ffbc7fd778a4d1ed530e9954a2675c84cc36798d229c6f8" gracePeriod=30 Dec 11 08:15:53 crc kubenswrapper[4860]: I1211 08:15:53.200026 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 08:15:53 crc kubenswrapper[4860]: I1211 08:15:53.212793 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 11 08:15:53 crc kubenswrapper[4860]: I1211 08:15:53.218045 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 11 08:15:53 crc kubenswrapper[4860]: I1211 08:15:53.221508 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=18.221487286 podStartE2EDuration="18.221487286s" podCreationTimestamp="2025-12-11 08:15:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:15:53.21873426 +0000 UTC m=+285.947253335" watchObservedRunningTime="2025-12-11 08:15:53.221487286 +0000 UTC m=+285.950006361" Dec 11 08:15:53 crc kubenswrapper[4860]: I1211 08:15:53.270626 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 11 08:15:53 crc kubenswrapper[4860]: I1211 08:15:53.328299 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/dd2052c5-96c6-4226-858e-61e3ddbfeef9-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-trp75\" (UID: \"dd2052c5-96c6-4226-858e-61e3ddbfeef9\") " pod="openshift-marketplace/marketplace-operator-79b997595-trp75" Dec 11 08:15:53 crc kubenswrapper[4860]: I1211 08:15:53.328411 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vhk55\" (UniqueName: \"kubernetes.io/projected/dd2052c5-96c6-4226-858e-61e3ddbfeef9-kube-api-access-vhk55\") pod \"marketplace-operator-79b997595-trp75\" (UID: \"dd2052c5-96c6-4226-858e-61e3ddbfeef9\") " pod="openshift-marketplace/marketplace-operator-79b997595-trp75" Dec 11 08:15:53 crc kubenswrapper[4860]: I1211 08:15:53.328452 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/dd2052c5-96c6-4226-858e-61e3ddbfeef9-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-trp75\" (UID: \"dd2052c5-96c6-4226-858e-61e3ddbfeef9\") " pod="openshift-marketplace/marketplace-operator-79b997595-trp75" Dec 11 08:15:53 crc kubenswrapper[4860]: I1211 08:15:53.334022 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 11 08:15:53 crc kubenswrapper[4860]: I1211 08:15:53.403228 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 11 08:15:53 crc kubenswrapper[4860]: I1211 08:15:53.430358 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vhk55\" (UniqueName: \"kubernetes.io/projected/dd2052c5-96c6-4226-858e-61e3ddbfeef9-kube-api-access-vhk55\") pod \"marketplace-operator-79b997595-trp75\" (UID: \"dd2052c5-96c6-4226-858e-61e3ddbfeef9\") " pod="openshift-marketplace/marketplace-operator-79b997595-trp75" Dec 11 08:15:53 crc kubenswrapper[4860]: I1211 08:15:53.430427 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/dd2052c5-96c6-4226-858e-61e3ddbfeef9-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-trp75\" (UID: \"dd2052c5-96c6-4226-858e-61e3ddbfeef9\") " pod="openshift-marketplace/marketplace-operator-79b997595-trp75" Dec 11 08:15:53 crc kubenswrapper[4860]: I1211 08:15:53.430499 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/dd2052c5-96c6-4226-858e-61e3ddbfeef9-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-trp75\" (UID: \"dd2052c5-96c6-4226-858e-61e3ddbfeef9\") " pod="openshift-marketplace/marketplace-operator-79b997595-trp75" Dec 11 08:15:53 crc kubenswrapper[4860]: I1211 08:15:53.432404 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/dd2052c5-96c6-4226-858e-61e3ddbfeef9-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-trp75\" (UID: \"dd2052c5-96c6-4226-858e-61e3ddbfeef9\") " pod="openshift-marketplace/marketplace-operator-79b997595-trp75" Dec 11 08:15:53 crc kubenswrapper[4860]: I1211 08:15:53.454714 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/dd2052c5-96c6-4226-858e-61e3ddbfeef9-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-trp75\" (UID: \"dd2052c5-96c6-4226-858e-61e3ddbfeef9\") " pod="openshift-marketplace/marketplace-operator-79b997595-trp75" Dec 11 08:15:53 crc kubenswrapper[4860]: I1211 08:15:53.466182 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vhk55\" (UniqueName: \"kubernetes.io/projected/dd2052c5-96c6-4226-858e-61e3ddbfeef9-kube-api-access-vhk55\") pod \"marketplace-operator-79b997595-trp75\" (UID: \"dd2052c5-96c6-4226-858e-61e3ddbfeef9\") " pod="openshift-marketplace/marketplace-operator-79b997595-trp75" Dec 11 08:15:53 crc kubenswrapper[4860]: I1211 08:15:53.513109 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-trp75" Dec 11 08:15:53 crc kubenswrapper[4860]: I1211 08:15:53.578594 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 11 08:15:53 crc kubenswrapper[4860]: I1211 08:15:53.620843 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 11 08:15:53 crc kubenswrapper[4860]: I1211 08:15:53.651845 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 11 08:15:53 crc kubenswrapper[4860]: I1211 08:15:53.671493 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 11 08:15:53 crc kubenswrapper[4860]: I1211 08:15:53.702375 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-ttrm2" Dec 11 08:15:53 crc kubenswrapper[4860]: I1211 08:15:53.772954 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mgfr4" Dec 11 08:15:53 crc kubenswrapper[4860]: I1211 08:15:53.783587 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dcccn" Dec 11 08:15:53 crc kubenswrapper[4860]: I1211 08:15:53.787735 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dzg4x" Dec 11 08:15:53 crc kubenswrapper[4860]: I1211 08:15:53.812175 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pqtvk" Dec 11 08:15:53 crc kubenswrapper[4860]: I1211 08:15:53.813953 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8p4mb" Dec 11 08:15:53 crc kubenswrapper[4860]: I1211 08:15:53.819489 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kqwgm" Dec 11 08:15:53 crc kubenswrapper[4860]: I1211 08:15:53.837505 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ad724960-18d1-4e83-944c-4cdeb033a436-marketplace-trusted-ca\") pod \"ad724960-18d1-4e83-944c-4cdeb033a436\" (UID: \"ad724960-18d1-4e83-944c-4cdeb033a436\") " Dec 11 08:15:53 crc kubenswrapper[4860]: I1211 08:15:53.837773 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/ad724960-18d1-4e83-944c-4cdeb033a436-marketplace-operator-metrics\") pod \"ad724960-18d1-4e83-944c-4cdeb033a436\" (UID: \"ad724960-18d1-4e83-944c-4cdeb033a436\") " Dec 11 08:15:53 crc kubenswrapper[4860]: I1211 08:15:53.838010 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-krqtd\" (UniqueName: \"kubernetes.io/projected/ad724960-18d1-4e83-944c-4cdeb033a436-kube-api-access-krqtd\") pod \"ad724960-18d1-4e83-944c-4cdeb033a436\" (UID: \"ad724960-18d1-4e83-944c-4cdeb033a436\") " Dec 11 08:15:53 crc kubenswrapper[4860]: I1211 08:15:53.839397 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ad724960-18d1-4e83-944c-4cdeb033a436-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "ad724960-18d1-4e83-944c-4cdeb033a436" (UID: "ad724960-18d1-4e83-944c-4cdeb033a436"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:15:53 crc kubenswrapper[4860]: I1211 08:15:53.845496 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad724960-18d1-4e83-944c-4cdeb033a436-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "ad724960-18d1-4e83-944c-4cdeb033a436" (UID: "ad724960-18d1-4e83-944c-4cdeb033a436"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:15:53 crc kubenswrapper[4860]: I1211 08:15:53.846016 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad724960-18d1-4e83-944c-4cdeb033a436-kube-api-access-krqtd" (OuterVolumeSpecName: "kube-api-access-krqtd") pod "ad724960-18d1-4e83-944c-4cdeb033a436" (UID: "ad724960-18d1-4e83-944c-4cdeb033a436"). InnerVolumeSpecName "kube-api-access-krqtd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:15:53 crc kubenswrapper[4860]: I1211 08:15:53.933662 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 11 08:15:53 crc kubenswrapper[4860]: I1211 08:15:53.939782 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/093a9903-7041-410b-bfb2-c74ef741e0be-utilities\") pod \"093a9903-7041-410b-bfb2-c74ef741e0be\" (UID: \"093a9903-7041-410b-bfb2-c74ef741e0be\") " Dec 11 08:15:53 crc kubenswrapper[4860]: I1211 08:15:53.940013 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/294792dd-5104-4308-a10f-f3ab0c4504e9-catalog-content\") pod \"294792dd-5104-4308-a10f-f3ab0c4504e9\" (UID: \"294792dd-5104-4308-a10f-f3ab0c4504e9\") " Dec 11 08:15:53 crc kubenswrapper[4860]: I1211 08:15:53.940169 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ac89442-1754-45d3-b67a-aa6cc31a1235-catalog-content\") pod \"7ac89442-1754-45d3-b67a-aa6cc31a1235\" (UID: \"7ac89442-1754-45d3-b67a-aa6cc31a1235\") " Dec 11 08:15:53 crc kubenswrapper[4860]: I1211 08:15:53.940297 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e052d7c-59d1-43a4-ac7a-e9daf2a88c65-catalog-content\") pod \"8e052d7c-59d1-43a4-ac7a-e9daf2a88c65\" (UID: \"8e052d7c-59d1-43a4-ac7a-e9daf2a88c65\") " Dec 11 08:15:53 crc kubenswrapper[4860]: I1211 08:15:53.940429 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30107273-766a-4575-8cb1-3de387248437-catalog-content\") pod \"30107273-766a-4575-8cb1-3de387248437\" (UID: \"30107273-766a-4575-8cb1-3de387248437\") " Dec 11 08:15:53 crc kubenswrapper[4860]: I1211 08:15:53.941950 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/294792dd-5104-4308-a10f-f3ab0c4504e9-utilities\") pod \"294792dd-5104-4308-a10f-f3ab0c4504e9\" (UID: \"294792dd-5104-4308-a10f-f3ab0c4504e9\") " Dec 11 08:15:53 crc kubenswrapper[4860]: I1211 08:15:53.942015 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x55qf\" (UniqueName: \"kubernetes.io/projected/8e052d7c-59d1-43a4-ac7a-e9daf2a88c65-kube-api-access-x55qf\") pod \"8e052d7c-59d1-43a4-ac7a-e9daf2a88c65\" (UID: \"8e052d7c-59d1-43a4-ac7a-e9daf2a88c65\") " Dec 11 08:15:53 crc kubenswrapper[4860]: I1211 08:15:53.942052 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ac89442-1754-45d3-b67a-aa6cc31a1235-utilities\") pod \"7ac89442-1754-45d3-b67a-aa6cc31a1235\" (UID: \"7ac89442-1754-45d3-b67a-aa6cc31a1235\") " Dec 11 08:15:53 crc kubenswrapper[4860]: I1211 08:15:53.942088 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/093a9903-7041-410b-bfb2-c74ef741e0be-catalog-content\") pod \"093a9903-7041-410b-bfb2-c74ef741e0be\" (UID: \"093a9903-7041-410b-bfb2-c74ef741e0be\") " Dec 11 08:15:53 crc kubenswrapper[4860]: I1211 08:15:53.942164 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e052d7c-59d1-43a4-ac7a-e9daf2a88c65-utilities\") pod \"8e052d7c-59d1-43a4-ac7a-e9daf2a88c65\" (UID: \"8e052d7c-59d1-43a4-ac7a-e9daf2a88c65\") " Dec 11 08:15:53 crc kubenswrapper[4860]: I1211 08:15:53.942182 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30107273-766a-4575-8cb1-3de387248437-utilities\") pod \"30107273-766a-4575-8cb1-3de387248437\" (UID: \"30107273-766a-4575-8cb1-3de387248437\") " Dec 11 08:15:53 crc kubenswrapper[4860]: I1211 08:15:53.942205 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tf7hl\" (UniqueName: \"kubernetes.io/projected/7ac89442-1754-45d3-b67a-aa6cc31a1235-kube-api-access-tf7hl\") pod \"7ac89442-1754-45d3-b67a-aa6cc31a1235\" (UID: \"7ac89442-1754-45d3-b67a-aa6cc31a1235\") " Dec 11 08:15:53 crc kubenswrapper[4860]: I1211 08:15:53.942229 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fnfjs\" (UniqueName: \"kubernetes.io/projected/30107273-766a-4575-8cb1-3de387248437-kube-api-access-fnfjs\") pod \"30107273-766a-4575-8cb1-3de387248437\" (UID: \"30107273-766a-4575-8cb1-3de387248437\") " Dec 11 08:15:53 crc kubenswrapper[4860]: I1211 08:15:53.942258 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f92fq\" (UniqueName: \"kubernetes.io/projected/093a9903-7041-410b-bfb2-c74ef741e0be-kube-api-access-f92fq\") pod \"093a9903-7041-410b-bfb2-c74ef741e0be\" (UID: \"093a9903-7041-410b-bfb2-c74ef741e0be\") " Dec 11 08:15:53 crc kubenswrapper[4860]: I1211 08:15:53.942284 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ftgqm\" (UniqueName: \"kubernetes.io/projected/294792dd-5104-4308-a10f-f3ab0c4504e9-kube-api-access-ftgqm\") pod \"294792dd-5104-4308-a10f-f3ab0c4504e9\" (UID: \"294792dd-5104-4308-a10f-f3ab0c4504e9\") " Dec 11 08:15:53 crc kubenswrapper[4860]: I1211 08:15:53.942322 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/01481957-a690-40f4-8c89-352b7350d327-catalog-content\") pod \"01481957-a690-40f4-8c89-352b7350d327\" (UID: \"01481957-a690-40f4-8c89-352b7350d327\") " Dec 11 08:15:53 crc kubenswrapper[4860]: I1211 08:15:53.942351 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/01481957-a690-40f4-8c89-352b7350d327-utilities\") pod \"01481957-a690-40f4-8c89-352b7350d327\" (UID: \"01481957-a690-40f4-8c89-352b7350d327\") " Dec 11 08:15:53 crc kubenswrapper[4860]: I1211 08:15:53.942370 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4n7mg\" (UniqueName: \"kubernetes.io/projected/01481957-a690-40f4-8c89-352b7350d327-kube-api-access-4n7mg\") pod \"01481957-a690-40f4-8c89-352b7350d327\" (UID: \"01481957-a690-40f4-8c89-352b7350d327\") " Dec 11 08:15:53 crc kubenswrapper[4860]: I1211 08:15:53.941326 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/093a9903-7041-410b-bfb2-c74ef741e0be-utilities" (OuterVolumeSpecName: "utilities") pod "093a9903-7041-410b-bfb2-c74ef741e0be" (UID: "093a9903-7041-410b-bfb2-c74ef741e0be"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:15:53 crc kubenswrapper[4860]: I1211 08:15:53.942907 4860 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/093a9903-7041-410b-bfb2-c74ef741e0be-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 08:15:53 crc kubenswrapper[4860]: I1211 08:15:53.942931 4860 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ad724960-18d1-4e83-944c-4cdeb033a436-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 11 08:15:53 crc kubenswrapper[4860]: I1211 08:15:53.942943 4860 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/ad724960-18d1-4e83-944c-4cdeb033a436-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 11 08:15:53 crc kubenswrapper[4860]: I1211 08:15:53.942956 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-krqtd\" (UniqueName: \"kubernetes.io/projected/ad724960-18d1-4e83-944c-4cdeb033a436-kube-api-access-krqtd\") on node \"crc\" DevicePath \"\"" Dec 11 08:15:53 crc kubenswrapper[4860]: I1211 08:15:53.945211 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01481957-a690-40f4-8c89-352b7350d327-kube-api-access-4n7mg" (OuterVolumeSpecName: "kube-api-access-4n7mg") pod "01481957-a690-40f4-8c89-352b7350d327" (UID: "01481957-a690-40f4-8c89-352b7350d327"). InnerVolumeSpecName "kube-api-access-4n7mg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:15:53 crc kubenswrapper[4860]: I1211 08:15:53.945326 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e052d7c-59d1-43a4-ac7a-e9daf2a88c65-kube-api-access-x55qf" (OuterVolumeSpecName: "kube-api-access-x55qf") pod "8e052d7c-59d1-43a4-ac7a-e9daf2a88c65" (UID: "8e052d7c-59d1-43a4-ac7a-e9daf2a88c65"). InnerVolumeSpecName "kube-api-access-x55qf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:15:53 crc kubenswrapper[4860]: I1211 08:15:53.945914 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/294792dd-5104-4308-a10f-f3ab0c4504e9-utilities" (OuterVolumeSpecName: "utilities") pod "294792dd-5104-4308-a10f-f3ab0c4504e9" (UID: "294792dd-5104-4308-a10f-f3ab0c4504e9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:15:53 crc kubenswrapper[4860]: I1211 08:15:53.946191 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8e052d7c-59d1-43a4-ac7a-e9daf2a88c65-utilities" (OuterVolumeSpecName: "utilities") pod "8e052d7c-59d1-43a4-ac7a-e9daf2a88c65" (UID: "8e052d7c-59d1-43a4-ac7a-e9daf2a88c65"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:15:53 crc kubenswrapper[4860]: I1211 08:15:53.946611 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7ac89442-1754-45d3-b67a-aa6cc31a1235-utilities" (OuterVolumeSpecName: "utilities") pod "7ac89442-1754-45d3-b67a-aa6cc31a1235" (UID: "7ac89442-1754-45d3-b67a-aa6cc31a1235"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:15:53 crc kubenswrapper[4860]: I1211 08:15:53.947454 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/01481957-a690-40f4-8c89-352b7350d327-utilities" (OuterVolumeSpecName: "utilities") pod "01481957-a690-40f4-8c89-352b7350d327" (UID: "01481957-a690-40f4-8c89-352b7350d327"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:15:53 crc kubenswrapper[4860]: I1211 08:15:53.948091 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/093a9903-7041-410b-bfb2-c74ef741e0be-kube-api-access-f92fq" (OuterVolumeSpecName: "kube-api-access-f92fq") pod "093a9903-7041-410b-bfb2-c74ef741e0be" (UID: "093a9903-7041-410b-bfb2-c74ef741e0be"). InnerVolumeSpecName "kube-api-access-f92fq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:15:53 crc kubenswrapper[4860]: I1211 08:15:53.948471 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/30107273-766a-4575-8cb1-3de387248437-kube-api-access-fnfjs" (OuterVolumeSpecName: "kube-api-access-fnfjs") pod "30107273-766a-4575-8cb1-3de387248437" (UID: "30107273-766a-4575-8cb1-3de387248437"). InnerVolumeSpecName "kube-api-access-fnfjs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:15:53 crc kubenswrapper[4860]: I1211 08:15:53.949433 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/30107273-766a-4575-8cb1-3de387248437-utilities" (OuterVolumeSpecName: "utilities") pod "30107273-766a-4575-8cb1-3de387248437" (UID: "30107273-766a-4575-8cb1-3de387248437"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:15:53 crc kubenswrapper[4860]: I1211 08:15:53.950032 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/294792dd-5104-4308-a10f-f3ab0c4504e9-kube-api-access-ftgqm" (OuterVolumeSpecName: "kube-api-access-ftgqm") pod "294792dd-5104-4308-a10f-f3ab0c4504e9" (UID: "294792dd-5104-4308-a10f-f3ab0c4504e9"). InnerVolumeSpecName "kube-api-access-ftgqm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:15:53 crc kubenswrapper[4860]: I1211 08:15:53.958378 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ac89442-1754-45d3-b67a-aa6cc31a1235-kube-api-access-tf7hl" (OuterVolumeSpecName: "kube-api-access-tf7hl") pod "7ac89442-1754-45d3-b67a-aa6cc31a1235" (UID: "7ac89442-1754-45d3-b67a-aa6cc31a1235"). InnerVolumeSpecName "kube-api-access-tf7hl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:15:53 crc kubenswrapper[4860]: I1211 08:15:53.966222 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/294792dd-5104-4308-a10f-f3ab0c4504e9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "294792dd-5104-4308-a10f-f3ab0c4504e9" (UID: "294792dd-5104-4308-a10f-f3ab0c4504e9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:15:53 crc kubenswrapper[4860]: I1211 08:15:53.977174 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8e052d7c-59d1-43a4-ac7a-e9daf2a88c65-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8e052d7c-59d1-43a4-ac7a-e9daf2a88c65" (UID: "8e052d7c-59d1-43a4-ac7a-e9daf2a88c65"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.007145 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.007288 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.018800 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7ac89442-1754-45d3-b67a-aa6cc31a1235-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7ac89442-1754-45d3-b67a-aa6cc31a1235" (UID: "7ac89442-1754-45d3-b67a-aa6cc31a1235"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.023338 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.041528 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/093a9903-7041-410b-bfb2-c74ef741e0be-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "093a9903-7041-410b-bfb2-c74ef741e0be" (UID: "093a9903-7041-410b-bfb2-c74ef741e0be"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.044240 4860 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/01481957-a690-40f4-8c89-352b7350d327-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.044301 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4n7mg\" (UniqueName: \"kubernetes.io/projected/01481957-a690-40f4-8c89-352b7350d327-kube-api-access-4n7mg\") on node \"crc\" DevicePath \"\"" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.044315 4860 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/294792dd-5104-4308-a10f-f3ab0c4504e9-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.044323 4860 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ac89442-1754-45d3-b67a-aa6cc31a1235-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.044332 4860 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e052d7c-59d1-43a4-ac7a-e9daf2a88c65-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.044340 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x55qf\" (UniqueName: \"kubernetes.io/projected/8e052d7c-59d1-43a4-ac7a-e9daf2a88c65-kube-api-access-x55qf\") on node \"crc\" DevicePath \"\"" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.044349 4860 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/294792dd-5104-4308-a10f-f3ab0c4504e9-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.044357 4860 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ac89442-1754-45d3-b67a-aa6cc31a1235-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.044365 4860 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/093a9903-7041-410b-bfb2-c74ef741e0be-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.044373 4860 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e052d7c-59d1-43a4-ac7a-e9daf2a88c65-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.044381 4860 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30107273-766a-4575-8cb1-3de387248437-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.044389 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tf7hl\" (UniqueName: \"kubernetes.io/projected/7ac89442-1754-45d3-b67a-aa6cc31a1235-kube-api-access-tf7hl\") on node \"crc\" DevicePath \"\"" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.044398 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fnfjs\" (UniqueName: \"kubernetes.io/projected/30107273-766a-4575-8cb1-3de387248437-kube-api-access-fnfjs\") on node \"crc\" DevicePath \"\"" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.044407 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f92fq\" (UniqueName: \"kubernetes.io/projected/093a9903-7041-410b-bfb2-c74ef741e0be-kube-api-access-f92fq\") on node \"crc\" DevicePath \"\"" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.044416 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ftgqm\" (UniqueName: \"kubernetes.io/projected/294792dd-5104-4308-a10f-f3ab0c4504e9-kube-api-access-ftgqm\") on node \"crc\" DevicePath \"\"" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.073230 4860 generic.go:334] "Generic (PLEG): container finished" podID="7ac89442-1754-45d3-b67a-aa6cc31a1235" containerID="f1a5ae4b9e0d2bfe7ffbc7fd778a4d1ed530e9954a2675c84cc36798d229c6f8" exitCode=0 Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.073318 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pqtvk" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.073635 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pqtvk" event={"ID":"7ac89442-1754-45d3-b67a-aa6cc31a1235","Type":"ContainerDied","Data":"f1a5ae4b9e0d2bfe7ffbc7fd778a4d1ed530e9954a2675c84cc36798d229c6f8"} Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.073756 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pqtvk" event={"ID":"7ac89442-1754-45d3-b67a-aa6cc31a1235","Type":"ContainerDied","Data":"bf503ca2ebe64d9958c793d4c050d3487e5fb82e8e8c12d45bb1796ee873902f"} Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.073855 4860 scope.go:117] "RemoveContainer" containerID="f1a5ae4b9e0d2bfe7ffbc7fd778a4d1ed530e9954a2675c84cc36798d229c6f8" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.074887 4860 generic.go:334] "Generic (PLEG): container finished" podID="ad724960-18d1-4e83-944c-4cdeb033a436" containerID="cafda173280b05a3de720d8e7f8927fc27b4c53b7500ba9deb3878e2112d3d8c" exitCode=0 Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.074997 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-ttrm2" event={"ID":"ad724960-18d1-4e83-944c-4cdeb033a436","Type":"ContainerDied","Data":"cafda173280b05a3de720d8e7f8927fc27b4c53b7500ba9deb3878e2112d3d8c"} Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.075080 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-ttrm2" event={"ID":"ad724960-18d1-4e83-944c-4cdeb033a436","Type":"ContainerDied","Data":"0e4f02a9720f30652b572acc71b8074196ee665d1a1bc67339fc4accc42ddee2"} Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.075202 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-ttrm2" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.079190 4860 generic.go:334] "Generic (PLEG): container finished" podID="01481957-a690-40f4-8c89-352b7350d327" containerID="d9777d9cd9cfe1b753ed0336374b4591e9f9e22227f397713e583f8e99a68df8" exitCode=0 Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.079382 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8p4mb" event={"ID":"01481957-a690-40f4-8c89-352b7350d327","Type":"ContainerDied","Data":"d9777d9cd9cfe1b753ed0336374b4591e9f9e22227f397713e583f8e99a68df8"} Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.079467 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8p4mb" event={"ID":"01481957-a690-40f4-8c89-352b7350d327","Type":"ContainerDied","Data":"09109e706dbeaf7ff53ad518750d66808c1f71b08f2a704acbd51462026a8a49"} Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.079593 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8p4mb" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.088077 4860 generic.go:334] "Generic (PLEG): container finished" podID="294792dd-5104-4308-a10f-f3ab0c4504e9" containerID="8b5bd9e3bd9159c617215631e1bd2e8031d35214329adfb7419a18627358c68d" exitCode=0 Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.088169 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dzg4x" event={"ID":"294792dd-5104-4308-a10f-f3ab0c4504e9","Type":"ContainerDied","Data":"8b5bd9e3bd9159c617215631e1bd2e8031d35214329adfb7419a18627358c68d"} Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.088193 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dzg4x" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.088325 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dzg4x" event={"ID":"294792dd-5104-4308-a10f-f3ab0c4504e9","Type":"ContainerDied","Data":"c989cbc39fe1f5dd125136d91c8ece373ef71dc48fb73cff33382e4d60655abc"} Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.090823 4860 generic.go:334] "Generic (PLEG): container finished" podID="8e052d7c-59d1-43a4-ac7a-e9daf2a88c65" containerID="f96f8c23423f6d30e831662d312b1f6086c55082925cb0383a915df10a158b79" exitCode=0 Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.090878 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dcccn" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.090915 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dcccn" event={"ID":"8e052d7c-59d1-43a4-ac7a-e9daf2a88c65","Type":"ContainerDied","Data":"f96f8c23423f6d30e831662d312b1f6086c55082925cb0383a915df10a158b79"} Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.092587 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dcccn" event={"ID":"8e052d7c-59d1-43a4-ac7a-e9daf2a88c65","Type":"ContainerDied","Data":"e978765edc94f035346011ecb76709f38a004e2a46dfcfa9c3b24c676dbdcbad"} Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.091341 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/30107273-766a-4575-8cb1-3de387248437-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "30107273-766a-4575-8cb1-3de387248437" (UID: "30107273-766a-4575-8cb1-3de387248437"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.101104 4860 scope.go:117] "RemoveContainer" containerID="2eec0fca2445e8ac1ce738b2d8ec1e8ab0d3f134b7d3fc0fa8c11102842021c6" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.102149 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/01481957-a690-40f4-8c89-352b7350d327-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "01481957-a690-40f4-8c89-352b7350d327" (UID: "01481957-a690-40f4-8c89-352b7350d327"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.102362 4860 generic.go:334] "Generic (PLEG): container finished" podID="30107273-766a-4575-8cb1-3de387248437" containerID="14e1cbe7def3e65e27f664f13218ca12aabceb8a2a528dacbf116993fd9b73b9" exitCode=0 Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.102441 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mgfr4" event={"ID":"30107273-766a-4575-8cb1-3de387248437","Type":"ContainerDied","Data":"14e1cbe7def3e65e27f664f13218ca12aabceb8a2a528dacbf116993fd9b73b9"} Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.102476 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mgfr4" event={"ID":"30107273-766a-4575-8cb1-3de387248437","Type":"ContainerDied","Data":"df3f1a03bb7d67e0dd6af440ad08068302ae497e58fefd9296b73bfb75c129cc"} Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.102562 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mgfr4" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.108954 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pqtvk"] Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.109473 4860 generic.go:334] "Generic (PLEG): container finished" podID="093a9903-7041-410b-bfb2-c74ef741e0be" containerID="41c0f2c9f5203be9afc00992cd7768f8a3896e585a81b5618856b40455c598eb" exitCode=0 Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.110815 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kqwgm" event={"ID":"093a9903-7041-410b-bfb2-c74ef741e0be","Type":"ContainerDied","Data":"41c0f2c9f5203be9afc00992cd7768f8a3896e585a81b5618856b40455c598eb"} Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.111339 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kqwgm" event={"ID":"093a9903-7041-410b-bfb2-c74ef741e0be","Type":"ContainerDied","Data":"687246172b9ae710d66e39478fe0475ab20c8db92fe23f13ad0072894b373ba6"} Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.110948 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kqwgm" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.115335 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-pqtvk"] Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.132551 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-ttrm2"] Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.137449 4860 scope.go:117] "RemoveContainer" containerID="684bf006e94de3d89bc7b8ef53686ca15a7aa10ba602e2881e0cee03ab001acb" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.137767 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-ttrm2"] Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.141632 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dcccn"] Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.142826 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.145942 4860 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/01481957-a690-40f4-8c89-352b7350d327-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.145987 4860 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30107273-766a-4575-8cb1-3de387248437-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.146098 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-dcccn"] Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.153906 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-mgfr4"] Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.159568 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-mgfr4"] Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.164378 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dzg4x"] Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.168740 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-dzg4x"] Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.171123 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kqwgm"] Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.171727 4860 scope.go:117] "RemoveContainer" containerID="f1a5ae4b9e0d2bfe7ffbc7fd778a4d1ed530e9954a2675c84cc36798d229c6f8" Dec 11 08:15:54 crc kubenswrapper[4860]: E1211 08:15:54.172535 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f1a5ae4b9e0d2bfe7ffbc7fd778a4d1ed530e9954a2675c84cc36798d229c6f8\": container with ID starting with f1a5ae4b9e0d2bfe7ffbc7fd778a4d1ed530e9954a2675c84cc36798d229c6f8 not found: ID does not exist" containerID="f1a5ae4b9e0d2bfe7ffbc7fd778a4d1ed530e9954a2675c84cc36798d229c6f8" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.172575 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f1a5ae4b9e0d2bfe7ffbc7fd778a4d1ed530e9954a2675c84cc36798d229c6f8"} err="failed to get container status \"f1a5ae4b9e0d2bfe7ffbc7fd778a4d1ed530e9954a2675c84cc36798d229c6f8\": rpc error: code = NotFound desc = could not find container \"f1a5ae4b9e0d2bfe7ffbc7fd778a4d1ed530e9954a2675c84cc36798d229c6f8\": container with ID starting with f1a5ae4b9e0d2bfe7ffbc7fd778a4d1ed530e9954a2675c84cc36798d229c6f8 not found: ID does not exist" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.172607 4860 scope.go:117] "RemoveContainer" containerID="2eec0fca2445e8ac1ce738b2d8ec1e8ab0d3f134b7d3fc0fa8c11102842021c6" Dec 11 08:15:54 crc kubenswrapper[4860]: E1211 08:15:54.172992 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2eec0fca2445e8ac1ce738b2d8ec1e8ab0d3f134b7d3fc0fa8c11102842021c6\": container with ID starting with 2eec0fca2445e8ac1ce738b2d8ec1e8ab0d3f134b7d3fc0fa8c11102842021c6 not found: ID does not exist" containerID="2eec0fca2445e8ac1ce738b2d8ec1e8ab0d3f134b7d3fc0fa8c11102842021c6" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.173018 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2eec0fca2445e8ac1ce738b2d8ec1e8ab0d3f134b7d3fc0fa8c11102842021c6"} err="failed to get container status \"2eec0fca2445e8ac1ce738b2d8ec1e8ab0d3f134b7d3fc0fa8c11102842021c6\": rpc error: code = NotFound desc = could not find container \"2eec0fca2445e8ac1ce738b2d8ec1e8ab0d3f134b7d3fc0fa8c11102842021c6\": container with ID starting with 2eec0fca2445e8ac1ce738b2d8ec1e8ab0d3f134b7d3fc0fa8c11102842021c6 not found: ID does not exist" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.173062 4860 scope.go:117] "RemoveContainer" containerID="684bf006e94de3d89bc7b8ef53686ca15a7aa10ba602e2881e0cee03ab001acb" Dec 11 08:15:54 crc kubenswrapper[4860]: E1211 08:15:54.173321 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"684bf006e94de3d89bc7b8ef53686ca15a7aa10ba602e2881e0cee03ab001acb\": container with ID starting with 684bf006e94de3d89bc7b8ef53686ca15a7aa10ba602e2881e0cee03ab001acb not found: ID does not exist" containerID="684bf006e94de3d89bc7b8ef53686ca15a7aa10ba602e2881e0cee03ab001acb" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.173351 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"684bf006e94de3d89bc7b8ef53686ca15a7aa10ba602e2881e0cee03ab001acb"} err="failed to get container status \"684bf006e94de3d89bc7b8ef53686ca15a7aa10ba602e2881e0cee03ab001acb\": rpc error: code = NotFound desc = could not find container \"684bf006e94de3d89bc7b8ef53686ca15a7aa10ba602e2881e0cee03ab001acb\": container with ID starting with 684bf006e94de3d89bc7b8ef53686ca15a7aa10ba602e2881e0cee03ab001acb not found: ID does not exist" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.173369 4860 scope.go:117] "RemoveContainer" containerID="cafda173280b05a3de720d8e7f8927fc27b4c53b7500ba9deb3878e2112d3d8c" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.174739 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-kqwgm"] Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.177984 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.190825 4860 scope.go:117] "RemoveContainer" containerID="cafda173280b05a3de720d8e7f8927fc27b4c53b7500ba9deb3878e2112d3d8c" Dec 11 08:15:54 crc kubenswrapper[4860]: E1211 08:15:54.191437 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cafda173280b05a3de720d8e7f8927fc27b4c53b7500ba9deb3878e2112d3d8c\": container with ID starting with cafda173280b05a3de720d8e7f8927fc27b4c53b7500ba9deb3878e2112d3d8c not found: ID does not exist" containerID="cafda173280b05a3de720d8e7f8927fc27b4c53b7500ba9deb3878e2112d3d8c" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.191491 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cafda173280b05a3de720d8e7f8927fc27b4c53b7500ba9deb3878e2112d3d8c"} err="failed to get container status \"cafda173280b05a3de720d8e7f8927fc27b4c53b7500ba9deb3878e2112d3d8c\": rpc error: code = NotFound desc = could not find container \"cafda173280b05a3de720d8e7f8927fc27b4c53b7500ba9deb3878e2112d3d8c\": container with ID starting with cafda173280b05a3de720d8e7f8927fc27b4c53b7500ba9deb3878e2112d3d8c not found: ID does not exist" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.191556 4860 scope.go:117] "RemoveContainer" containerID="d9777d9cd9cfe1b753ed0336374b4591e9f9e22227f397713e583f8e99a68df8" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.206412 4860 scope.go:117] "RemoveContainer" containerID="b4534c18fcfa223bf76f764f0fb5c2a190fe3875b8959c266ed75e731ee682b3" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.220467 4860 scope.go:117] "RemoveContainer" containerID="a9587c21a079451102e0829c093f0967af046ffa260e000b16b6bcfc2d8a3dbe" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.236365 4860 scope.go:117] "RemoveContainer" containerID="d9777d9cd9cfe1b753ed0336374b4591e9f9e22227f397713e583f8e99a68df8" Dec 11 08:15:54 crc kubenswrapper[4860]: E1211 08:15:54.236931 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d9777d9cd9cfe1b753ed0336374b4591e9f9e22227f397713e583f8e99a68df8\": container with ID starting with d9777d9cd9cfe1b753ed0336374b4591e9f9e22227f397713e583f8e99a68df8 not found: ID does not exist" containerID="d9777d9cd9cfe1b753ed0336374b4591e9f9e22227f397713e583f8e99a68df8" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.236976 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d9777d9cd9cfe1b753ed0336374b4591e9f9e22227f397713e583f8e99a68df8"} err="failed to get container status \"d9777d9cd9cfe1b753ed0336374b4591e9f9e22227f397713e583f8e99a68df8\": rpc error: code = NotFound desc = could not find container \"d9777d9cd9cfe1b753ed0336374b4591e9f9e22227f397713e583f8e99a68df8\": container with ID starting with d9777d9cd9cfe1b753ed0336374b4591e9f9e22227f397713e583f8e99a68df8 not found: ID does not exist" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.237005 4860 scope.go:117] "RemoveContainer" containerID="b4534c18fcfa223bf76f764f0fb5c2a190fe3875b8959c266ed75e731ee682b3" Dec 11 08:15:54 crc kubenswrapper[4860]: E1211 08:15:54.237436 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b4534c18fcfa223bf76f764f0fb5c2a190fe3875b8959c266ed75e731ee682b3\": container with ID starting with b4534c18fcfa223bf76f764f0fb5c2a190fe3875b8959c266ed75e731ee682b3 not found: ID does not exist" containerID="b4534c18fcfa223bf76f764f0fb5c2a190fe3875b8959c266ed75e731ee682b3" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.237475 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b4534c18fcfa223bf76f764f0fb5c2a190fe3875b8959c266ed75e731ee682b3"} err="failed to get container status \"b4534c18fcfa223bf76f764f0fb5c2a190fe3875b8959c266ed75e731ee682b3\": rpc error: code = NotFound desc = could not find container \"b4534c18fcfa223bf76f764f0fb5c2a190fe3875b8959c266ed75e731ee682b3\": container with ID starting with b4534c18fcfa223bf76f764f0fb5c2a190fe3875b8959c266ed75e731ee682b3 not found: ID does not exist" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.237504 4860 scope.go:117] "RemoveContainer" containerID="a9587c21a079451102e0829c093f0967af046ffa260e000b16b6bcfc2d8a3dbe" Dec 11 08:15:54 crc kubenswrapper[4860]: E1211 08:15:54.237863 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a9587c21a079451102e0829c093f0967af046ffa260e000b16b6bcfc2d8a3dbe\": container with ID starting with a9587c21a079451102e0829c093f0967af046ffa260e000b16b6bcfc2d8a3dbe not found: ID does not exist" containerID="a9587c21a079451102e0829c093f0967af046ffa260e000b16b6bcfc2d8a3dbe" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.237962 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a9587c21a079451102e0829c093f0967af046ffa260e000b16b6bcfc2d8a3dbe"} err="failed to get container status \"a9587c21a079451102e0829c093f0967af046ffa260e000b16b6bcfc2d8a3dbe\": rpc error: code = NotFound desc = could not find container \"a9587c21a079451102e0829c093f0967af046ffa260e000b16b6bcfc2d8a3dbe\": container with ID starting with a9587c21a079451102e0829c093f0967af046ffa260e000b16b6bcfc2d8a3dbe not found: ID does not exist" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.238043 4860 scope.go:117] "RemoveContainer" containerID="8b5bd9e3bd9159c617215631e1bd2e8031d35214329adfb7419a18627358c68d" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.251477 4860 scope.go:117] "RemoveContainer" containerID="fdf701a14de86401e1b762180ef4ada3e5a21e1a0c9001e9c146d6971a79594f" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.252407 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.265762 4860 scope.go:117] "RemoveContainer" containerID="1f93e20350117e6aec17bfb51b14de14dc965a6d4a235499361eed82a2fe74d7" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.279674 4860 scope.go:117] "RemoveContainer" containerID="8b5bd9e3bd9159c617215631e1bd2e8031d35214329adfb7419a18627358c68d" Dec 11 08:15:54 crc kubenswrapper[4860]: E1211 08:15:54.280209 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8b5bd9e3bd9159c617215631e1bd2e8031d35214329adfb7419a18627358c68d\": container with ID starting with 8b5bd9e3bd9159c617215631e1bd2e8031d35214329adfb7419a18627358c68d not found: ID does not exist" containerID="8b5bd9e3bd9159c617215631e1bd2e8031d35214329adfb7419a18627358c68d" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.280257 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b5bd9e3bd9159c617215631e1bd2e8031d35214329adfb7419a18627358c68d"} err="failed to get container status \"8b5bd9e3bd9159c617215631e1bd2e8031d35214329adfb7419a18627358c68d\": rpc error: code = NotFound desc = could not find container \"8b5bd9e3bd9159c617215631e1bd2e8031d35214329adfb7419a18627358c68d\": container with ID starting with 8b5bd9e3bd9159c617215631e1bd2e8031d35214329adfb7419a18627358c68d not found: ID does not exist" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.280286 4860 scope.go:117] "RemoveContainer" containerID="fdf701a14de86401e1b762180ef4ada3e5a21e1a0c9001e9c146d6971a79594f" Dec 11 08:15:54 crc kubenswrapper[4860]: E1211 08:15:54.280993 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fdf701a14de86401e1b762180ef4ada3e5a21e1a0c9001e9c146d6971a79594f\": container with ID starting with fdf701a14de86401e1b762180ef4ada3e5a21e1a0c9001e9c146d6971a79594f not found: ID does not exist" containerID="fdf701a14de86401e1b762180ef4ada3e5a21e1a0c9001e9c146d6971a79594f" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.281071 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fdf701a14de86401e1b762180ef4ada3e5a21e1a0c9001e9c146d6971a79594f"} err="failed to get container status \"fdf701a14de86401e1b762180ef4ada3e5a21e1a0c9001e9c146d6971a79594f\": rpc error: code = NotFound desc = could not find container \"fdf701a14de86401e1b762180ef4ada3e5a21e1a0c9001e9c146d6971a79594f\": container with ID starting with fdf701a14de86401e1b762180ef4ada3e5a21e1a0c9001e9c146d6971a79594f not found: ID does not exist" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.281129 4860 scope.go:117] "RemoveContainer" containerID="1f93e20350117e6aec17bfb51b14de14dc965a6d4a235499361eed82a2fe74d7" Dec 11 08:15:54 crc kubenswrapper[4860]: E1211 08:15:54.281507 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1f93e20350117e6aec17bfb51b14de14dc965a6d4a235499361eed82a2fe74d7\": container with ID starting with 1f93e20350117e6aec17bfb51b14de14dc965a6d4a235499361eed82a2fe74d7 not found: ID does not exist" containerID="1f93e20350117e6aec17bfb51b14de14dc965a6d4a235499361eed82a2fe74d7" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.281629 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1f93e20350117e6aec17bfb51b14de14dc965a6d4a235499361eed82a2fe74d7"} err="failed to get container status \"1f93e20350117e6aec17bfb51b14de14dc965a6d4a235499361eed82a2fe74d7\": rpc error: code = NotFound desc = could not find container \"1f93e20350117e6aec17bfb51b14de14dc965a6d4a235499361eed82a2fe74d7\": container with ID starting with 1f93e20350117e6aec17bfb51b14de14dc965a6d4a235499361eed82a2fe74d7 not found: ID does not exist" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.281738 4860 scope.go:117] "RemoveContainer" containerID="f96f8c23423f6d30e831662d312b1f6086c55082925cb0383a915df10a158b79" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.308323 4860 scope.go:117] "RemoveContainer" containerID="2383f4d15b00b86de2542756c810c6715d4b50e92963ff933972eef2e3906e13" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.309624 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.327743 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.338450 4860 scope.go:117] "RemoveContainer" containerID="3d49cb2b0cb5b10ea530e20f72a5bce70bd0b2bf06405ab7b75dc3260495cba3" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.348099 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.355677 4860 scope.go:117] "RemoveContainer" containerID="f96f8c23423f6d30e831662d312b1f6086c55082925cb0383a915df10a158b79" Dec 11 08:15:54 crc kubenswrapper[4860]: E1211 08:15:54.362084 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f96f8c23423f6d30e831662d312b1f6086c55082925cb0383a915df10a158b79\": container with ID starting with f96f8c23423f6d30e831662d312b1f6086c55082925cb0383a915df10a158b79 not found: ID does not exist" containerID="f96f8c23423f6d30e831662d312b1f6086c55082925cb0383a915df10a158b79" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.362129 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f96f8c23423f6d30e831662d312b1f6086c55082925cb0383a915df10a158b79"} err="failed to get container status \"f96f8c23423f6d30e831662d312b1f6086c55082925cb0383a915df10a158b79\": rpc error: code = NotFound desc = could not find container \"f96f8c23423f6d30e831662d312b1f6086c55082925cb0383a915df10a158b79\": container with ID starting with f96f8c23423f6d30e831662d312b1f6086c55082925cb0383a915df10a158b79 not found: ID does not exist" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.362184 4860 scope.go:117] "RemoveContainer" containerID="2383f4d15b00b86de2542756c810c6715d4b50e92963ff933972eef2e3906e13" Dec 11 08:15:54 crc kubenswrapper[4860]: E1211 08:15:54.362596 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2383f4d15b00b86de2542756c810c6715d4b50e92963ff933972eef2e3906e13\": container with ID starting with 2383f4d15b00b86de2542756c810c6715d4b50e92963ff933972eef2e3906e13 not found: ID does not exist" containerID="2383f4d15b00b86de2542756c810c6715d4b50e92963ff933972eef2e3906e13" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.362655 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2383f4d15b00b86de2542756c810c6715d4b50e92963ff933972eef2e3906e13"} err="failed to get container status \"2383f4d15b00b86de2542756c810c6715d4b50e92963ff933972eef2e3906e13\": rpc error: code = NotFound desc = could not find container \"2383f4d15b00b86de2542756c810c6715d4b50e92963ff933972eef2e3906e13\": container with ID starting with 2383f4d15b00b86de2542756c810c6715d4b50e92963ff933972eef2e3906e13 not found: ID does not exist" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.362685 4860 scope.go:117] "RemoveContainer" containerID="3d49cb2b0cb5b10ea530e20f72a5bce70bd0b2bf06405ab7b75dc3260495cba3" Dec 11 08:15:54 crc kubenswrapper[4860]: E1211 08:15:54.363047 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3d49cb2b0cb5b10ea530e20f72a5bce70bd0b2bf06405ab7b75dc3260495cba3\": container with ID starting with 3d49cb2b0cb5b10ea530e20f72a5bce70bd0b2bf06405ab7b75dc3260495cba3 not found: ID does not exist" containerID="3d49cb2b0cb5b10ea530e20f72a5bce70bd0b2bf06405ab7b75dc3260495cba3" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.363102 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d49cb2b0cb5b10ea530e20f72a5bce70bd0b2bf06405ab7b75dc3260495cba3"} err="failed to get container status \"3d49cb2b0cb5b10ea530e20f72a5bce70bd0b2bf06405ab7b75dc3260495cba3\": rpc error: code = NotFound desc = could not find container \"3d49cb2b0cb5b10ea530e20f72a5bce70bd0b2bf06405ab7b75dc3260495cba3\": container with ID starting with 3d49cb2b0cb5b10ea530e20f72a5bce70bd0b2bf06405ab7b75dc3260495cba3 not found: ID does not exist" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.363137 4860 scope.go:117] "RemoveContainer" containerID="14e1cbe7def3e65e27f664f13218ca12aabceb8a2a528dacbf116993fd9b73b9" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.367065 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.388743 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.410284 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.419771 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8p4mb"] Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.419777 4860 scope.go:117] "RemoveContainer" containerID="a1981178d5f6012d6286ef994fb1180c51510e9a39d27894f17ec36a76a2cf05" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.423899 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-8p4mb"] Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.436882 4860 scope.go:117] "RemoveContainer" containerID="5fd620a265f48ef3dbf7fbe3d0076723092d08f33ac19e073d2931edfd8efaa9" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.460898 4860 scope.go:117] "RemoveContainer" containerID="14e1cbe7def3e65e27f664f13218ca12aabceb8a2a528dacbf116993fd9b73b9" Dec 11 08:15:54 crc kubenswrapper[4860]: E1211 08:15:54.461393 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"14e1cbe7def3e65e27f664f13218ca12aabceb8a2a528dacbf116993fd9b73b9\": container with ID starting with 14e1cbe7def3e65e27f664f13218ca12aabceb8a2a528dacbf116993fd9b73b9 not found: ID does not exist" containerID="14e1cbe7def3e65e27f664f13218ca12aabceb8a2a528dacbf116993fd9b73b9" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.461431 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"14e1cbe7def3e65e27f664f13218ca12aabceb8a2a528dacbf116993fd9b73b9"} err="failed to get container status \"14e1cbe7def3e65e27f664f13218ca12aabceb8a2a528dacbf116993fd9b73b9\": rpc error: code = NotFound desc = could not find container \"14e1cbe7def3e65e27f664f13218ca12aabceb8a2a528dacbf116993fd9b73b9\": container with ID starting with 14e1cbe7def3e65e27f664f13218ca12aabceb8a2a528dacbf116993fd9b73b9 not found: ID does not exist" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.461458 4860 scope.go:117] "RemoveContainer" containerID="a1981178d5f6012d6286ef994fb1180c51510e9a39d27894f17ec36a76a2cf05" Dec 11 08:15:54 crc kubenswrapper[4860]: E1211 08:15:54.461820 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a1981178d5f6012d6286ef994fb1180c51510e9a39d27894f17ec36a76a2cf05\": container with ID starting with a1981178d5f6012d6286ef994fb1180c51510e9a39d27894f17ec36a76a2cf05 not found: ID does not exist" containerID="a1981178d5f6012d6286ef994fb1180c51510e9a39d27894f17ec36a76a2cf05" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.461839 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a1981178d5f6012d6286ef994fb1180c51510e9a39d27894f17ec36a76a2cf05"} err="failed to get container status \"a1981178d5f6012d6286ef994fb1180c51510e9a39d27894f17ec36a76a2cf05\": rpc error: code = NotFound desc = could not find container \"a1981178d5f6012d6286ef994fb1180c51510e9a39d27894f17ec36a76a2cf05\": container with ID starting with a1981178d5f6012d6286ef994fb1180c51510e9a39d27894f17ec36a76a2cf05 not found: ID does not exist" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.461852 4860 scope.go:117] "RemoveContainer" containerID="5fd620a265f48ef3dbf7fbe3d0076723092d08f33ac19e073d2931edfd8efaa9" Dec 11 08:15:54 crc kubenswrapper[4860]: E1211 08:15:54.462153 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5fd620a265f48ef3dbf7fbe3d0076723092d08f33ac19e073d2931edfd8efaa9\": container with ID starting with 5fd620a265f48ef3dbf7fbe3d0076723092d08f33ac19e073d2931edfd8efaa9 not found: ID does not exist" containerID="5fd620a265f48ef3dbf7fbe3d0076723092d08f33ac19e073d2931edfd8efaa9" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.462171 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5fd620a265f48ef3dbf7fbe3d0076723092d08f33ac19e073d2931edfd8efaa9"} err="failed to get container status \"5fd620a265f48ef3dbf7fbe3d0076723092d08f33ac19e073d2931edfd8efaa9\": rpc error: code = NotFound desc = could not find container \"5fd620a265f48ef3dbf7fbe3d0076723092d08f33ac19e073d2931edfd8efaa9\": container with ID starting with 5fd620a265f48ef3dbf7fbe3d0076723092d08f33ac19e073d2931edfd8efaa9 not found: ID does not exist" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.462183 4860 scope.go:117] "RemoveContainer" containerID="41c0f2c9f5203be9afc00992cd7768f8a3896e585a81b5618856b40455c598eb" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.477999 4860 scope.go:117] "RemoveContainer" containerID="733a892d4a4dafcbbc2de14f8a486771dc23a0037161b96eb03eff048301e384" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.491238 4860 scope.go:117] "RemoveContainer" containerID="1c218b10241590e53b4f67b40eb1342404ed52f9336ace9cde6ab0c86e632357" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.491705 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.499483 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.512372 4860 scope.go:117] "RemoveContainer" containerID="41c0f2c9f5203be9afc00992cd7768f8a3896e585a81b5618856b40455c598eb" Dec 11 08:15:54 crc kubenswrapper[4860]: E1211 08:15:54.512788 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"41c0f2c9f5203be9afc00992cd7768f8a3896e585a81b5618856b40455c598eb\": container with ID starting with 41c0f2c9f5203be9afc00992cd7768f8a3896e585a81b5618856b40455c598eb not found: ID does not exist" containerID="41c0f2c9f5203be9afc00992cd7768f8a3896e585a81b5618856b40455c598eb" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.512818 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"41c0f2c9f5203be9afc00992cd7768f8a3896e585a81b5618856b40455c598eb"} err="failed to get container status \"41c0f2c9f5203be9afc00992cd7768f8a3896e585a81b5618856b40455c598eb\": rpc error: code = NotFound desc = could not find container \"41c0f2c9f5203be9afc00992cd7768f8a3896e585a81b5618856b40455c598eb\": container with ID starting with 41c0f2c9f5203be9afc00992cd7768f8a3896e585a81b5618856b40455c598eb not found: ID does not exist" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.512842 4860 scope.go:117] "RemoveContainer" containerID="733a892d4a4dafcbbc2de14f8a486771dc23a0037161b96eb03eff048301e384" Dec 11 08:15:54 crc kubenswrapper[4860]: E1211 08:15:54.513241 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"733a892d4a4dafcbbc2de14f8a486771dc23a0037161b96eb03eff048301e384\": container with ID starting with 733a892d4a4dafcbbc2de14f8a486771dc23a0037161b96eb03eff048301e384 not found: ID does not exist" containerID="733a892d4a4dafcbbc2de14f8a486771dc23a0037161b96eb03eff048301e384" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.513265 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"733a892d4a4dafcbbc2de14f8a486771dc23a0037161b96eb03eff048301e384"} err="failed to get container status \"733a892d4a4dafcbbc2de14f8a486771dc23a0037161b96eb03eff048301e384\": rpc error: code = NotFound desc = could not find container \"733a892d4a4dafcbbc2de14f8a486771dc23a0037161b96eb03eff048301e384\": container with ID starting with 733a892d4a4dafcbbc2de14f8a486771dc23a0037161b96eb03eff048301e384 not found: ID does not exist" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.513285 4860 scope.go:117] "RemoveContainer" containerID="1c218b10241590e53b4f67b40eb1342404ed52f9336ace9cde6ab0c86e632357" Dec 11 08:15:54 crc kubenswrapper[4860]: E1211 08:15:54.513611 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1c218b10241590e53b4f67b40eb1342404ed52f9336ace9cde6ab0c86e632357\": container with ID starting with 1c218b10241590e53b4f67b40eb1342404ed52f9336ace9cde6ab0c86e632357 not found: ID does not exist" containerID="1c218b10241590e53b4f67b40eb1342404ed52f9336ace9cde6ab0c86e632357" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.513682 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c218b10241590e53b4f67b40eb1342404ed52f9336ace9cde6ab0c86e632357"} err="failed to get container status \"1c218b10241590e53b4f67b40eb1342404ed52f9336ace9cde6ab0c86e632357\": rpc error: code = NotFound desc = could not find container \"1c218b10241590e53b4f67b40eb1342404ed52f9336ace9cde6ab0c86e632357\": container with ID starting with 1c218b10241590e53b4f67b40eb1342404ed52f9336ace9cde6ab0c86e632357 not found: ID does not exist" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.589034 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.618058 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.649391 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.671531 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.709920 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.712735 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.773237 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.842636 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.872707 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.884240 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.942662 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 11 08:15:54 crc kubenswrapper[4860]: I1211 08:15:54.988462 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 11 08:15:55 crc kubenswrapper[4860]: I1211 08:15:55.010588 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 11 08:15:55 crc kubenswrapper[4860]: I1211 08:15:55.027844 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 11 08:15:55 crc kubenswrapper[4860]: I1211 08:15:55.111542 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 11 08:15:55 crc kubenswrapper[4860]: I1211 08:15:55.125049 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 11 08:15:55 crc kubenswrapper[4860]: I1211 08:15:55.144609 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 11 08:15:55 crc kubenswrapper[4860]: I1211 08:15:55.152911 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 11 08:15:55 crc kubenswrapper[4860]: I1211 08:15:55.226092 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 11 08:15:55 crc kubenswrapper[4860]: I1211 08:15:55.288905 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 11 08:15:55 crc kubenswrapper[4860]: I1211 08:15:55.352573 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-trp75"] Dec 11 08:15:55 crc kubenswrapper[4860]: I1211 08:15:55.410142 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 11 08:15:55 crc kubenswrapper[4860]: I1211 08:15:55.419194 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 11 08:15:55 crc kubenswrapper[4860]: I1211 08:15:55.447850 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 11 08:15:55 crc kubenswrapper[4860]: I1211 08:15:55.471466 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 11 08:15:55 crc kubenswrapper[4860]: I1211 08:15:55.510430 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 11 08:15:55 crc kubenswrapper[4860]: I1211 08:15:55.585164 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01481957-a690-40f4-8c89-352b7350d327" path="/var/lib/kubelet/pods/01481957-a690-40f4-8c89-352b7350d327/volumes" Dec 11 08:15:55 crc kubenswrapper[4860]: I1211 08:15:55.585919 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="093a9903-7041-410b-bfb2-c74ef741e0be" path="/var/lib/kubelet/pods/093a9903-7041-410b-bfb2-c74ef741e0be/volumes" Dec 11 08:15:55 crc kubenswrapper[4860]: I1211 08:15:55.586395 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 11 08:15:55 crc kubenswrapper[4860]: I1211 08:15:55.586764 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="294792dd-5104-4308-a10f-f3ab0c4504e9" path="/var/lib/kubelet/pods/294792dd-5104-4308-a10f-f3ab0c4504e9/volumes" Dec 11 08:15:55 crc kubenswrapper[4860]: I1211 08:15:55.588209 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="30107273-766a-4575-8cb1-3de387248437" path="/var/lib/kubelet/pods/30107273-766a-4575-8cb1-3de387248437/volumes" Dec 11 08:15:55 crc kubenswrapper[4860]: I1211 08:15:55.588892 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ac89442-1754-45d3-b67a-aa6cc31a1235" path="/var/lib/kubelet/pods/7ac89442-1754-45d3-b67a-aa6cc31a1235/volumes" Dec 11 08:15:55 crc kubenswrapper[4860]: I1211 08:15:55.590038 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8e052d7c-59d1-43a4-ac7a-e9daf2a88c65" path="/var/lib/kubelet/pods/8e052d7c-59d1-43a4-ac7a-e9daf2a88c65/volumes" Dec 11 08:15:55 crc kubenswrapper[4860]: I1211 08:15:55.590806 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ad724960-18d1-4e83-944c-4cdeb033a436" path="/var/lib/kubelet/pods/ad724960-18d1-4e83-944c-4cdeb033a436/volumes" Dec 11 08:15:55 crc kubenswrapper[4860]: I1211 08:15:55.598018 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 11 08:15:55 crc kubenswrapper[4860]: I1211 08:15:55.659966 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 11 08:15:55 crc kubenswrapper[4860]: I1211 08:15:55.688049 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 11 08:15:55 crc kubenswrapper[4860]: I1211 08:15:55.718594 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 11 08:15:55 crc kubenswrapper[4860]: I1211 08:15:55.722574 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 11 08:15:55 crc kubenswrapper[4860]: I1211 08:15:55.877423 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 11 08:15:55 crc kubenswrapper[4860]: I1211 08:15:55.891410 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 11 08:15:55 crc kubenswrapper[4860]: I1211 08:15:55.915621 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 11 08:15:55 crc kubenswrapper[4860]: I1211 08:15:55.932504 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 11 08:15:55 crc kubenswrapper[4860]: I1211 08:15:55.987625 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 11 08:15:56 crc kubenswrapper[4860]: I1211 08:15:56.011108 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 11 08:15:56 crc kubenswrapper[4860]: I1211 08:15:56.127039 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-trp75" event={"ID":"dd2052c5-96c6-4226-858e-61e3ddbfeef9","Type":"ContainerStarted","Data":"be1960fe953541d56ca59316d13a13345e0c2e5e48dc04934b9969e12e67a023"} Dec 11 08:15:56 crc kubenswrapper[4860]: I1211 08:15:56.127086 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-trp75" event={"ID":"dd2052c5-96c6-4226-858e-61e3ddbfeef9","Type":"ContainerStarted","Data":"480503244b1841fd743d669fb9854d231707c265a5b9d245cec375181ff8abfc"} Dec 11 08:15:56 crc kubenswrapper[4860]: I1211 08:15:56.127409 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-trp75" Dec 11 08:15:56 crc kubenswrapper[4860]: I1211 08:15:56.130896 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-trp75" Dec 11 08:15:56 crc kubenswrapper[4860]: I1211 08:15:56.148004 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-trp75" podStartSLOduration=4.14797937 podStartE2EDuration="4.14797937s" podCreationTimestamp="2025-12-11 08:15:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:15:56.141714332 +0000 UTC m=+288.870233397" watchObservedRunningTime="2025-12-11 08:15:56.14797937 +0000 UTC m=+288.876498465" Dec 11 08:15:56 crc kubenswrapper[4860]: I1211 08:15:56.203843 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 11 08:15:56 crc kubenswrapper[4860]: I1211 08:15:56.210743 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 11 08:15:56 crc kubenswrapper[4860]: I1211 08:15:56.215560 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 11 08:15:56 crc kubenswrapper[4860]: I1211 08:15:56.229066 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 11 08:15:56 crc kubenswrapper[4860]: I1211 08:15:56.251776 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 11 08:15:56 crc kubenswrapper[4860]: I1211 08:15:56.256010 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 11 08:15:56 crc kubenswrapper[4860]: I1211 08:15:56.264472 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 11 08:15:56 crc kubenswrapper[4860]: I1211 08:15:56.265286 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 11 08:15:56 crc kubenswrapper[4860]: I1211 08:15:56.284819 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 11 08:15:56 crc kubenswrapper[4860]: I1211 08:15:56.419976 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 11 08:15:56 crc kubenswrapper[4860]: I1211 08:15:56.452849 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 11 08:15:56 crc kubenswrapper[4860]: I1211 08:15:56.502053 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 11 08:15:56 crc kubenswrapper[4860]: I1211 08:15:56.545382 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 11 08:15:56 crc kubenswrapper[4860]: I1211 08:15:56.661416 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 11 08:15:56 crc kubenswrapper[4860]: I1211 08:15:56.772475 4860 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 11 08:15:56 crc kubenswrapper[4860]: I1211 08:15:56.802957 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 11 08:15:56 crc kubenswrapper[4860]: I1211 08:15:56.836958 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 11 08:15:56 crc kubenswrapper[4860]: I1211 08:15:56.845561 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 11 08:15:56 crc kubenswrapper[4860]: I1211 08:15:56.855081 4860 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 11 08:15:56 crc kubenswrapper[4860]: I1211 08:15:56.986206 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 11 08:15:57 crc kubenswrapper[4860]: I1211 08:15:57.007629 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 11 08:15:57 crc kubenswrapper[4860]: I1211 08:15:57.016375 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 11 08:15:57 crc kubenswrapper[4860]: I1211 08:15:57.034520 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 11 08:15:57 crc kubenswrapper[4860]: I1211 08:15:57.132482 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 11 08:15:57 crc kubenswrapper[4860]: I1211 08:15:57.149779 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 11 08:15:57 crc kubenswrapper[4860]: I1211 08:15:57.185173 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 11 08:15:57 crc kubenswrapper[4860]: I1211 08:15:57.264068 4860 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 11 08:15:57 crc kubenswrapper[4860]: I1211 08:15:57.315427 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 11 08:15:57 crc kubenswrapper[4860]: I1211 08:15:57.431264 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 11 08:15:57 crc kubenswrapper[4860]: I1211 08:15:57.441340 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 11 08:15:57 crc kubenswrapper[4860]: I1211 08:15:57.461967 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 11 08:15:57 crc kubenswrapper[4860]: I1211 08:15:57.476982 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 11 08:15:57 crc kubenswrapper[4860]: I1211 08:15:57.632006 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 11 08:15:57 crc kubenswrapper[4860]: I1211 08:15:57.659271 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 11 08:15:57 crc kubenswrapper[4860]: I1211 08:15:57.877616 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 11 08:15:57 crc kubenswrapper[4860]: I1211 08:15:57.901961 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 11 08:15:58 crc kubenswrapper[4860]: I1211 08:15:58.080327 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 11 08:15:58 crc kubenswrapper[4860]: I1211 08:15:58.082411 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 11 08:15:58 crc kubenswrapper[4860]: I1211 08:15:58.107880 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 11 08:15:58 crc kubenswrapper[4860]: I1211 08:15:58.197474 4860 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 11 08:15:58 crc kubenswrapper[4860]: I1211 08:15:58.197738 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://7c8d32000cd3750bd5ea7b6fd6bc17c8db195e315089cd5a1ed0023342f049aa" gracePeriod=5 Dec 11 08:15:58 crc kubenswrapper[4860]: I1211 08:15:58.198133 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 11 08:15:58 crc kubenswrapper[4860]: I1211 08:15:58.200496 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 11 08:15:58 crc kubenswrapper[4860]: I1211 08:15:58.274470 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 11 08:15:58 crc kubenswrapper[4860]: I1211 08:15:58.416366 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 11 08:15:58 crc kubenswrapper[4860]: I1211 08:15:58.476174 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 11 08:15:58 crc kubenswrapper[4860]: I1211 08:15:58.622076 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 11 08:15:58 crc kubenswrapper[4860]: I1211 08:15:58.650657 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 11 08:15:58 crc kubenswrapper[4860]: I1211 08:15:58.663652 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 11 08:15:58 crc kubenswrapper[4860]: I1211 08:15:58.666402 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 11 08:15:58 crc kubenswrapper[4860]: I1211 08:15:58.695205 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 11 08:15:58 crc kubenswrapper[4860]: I1211 08:15:58.801436 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 11 08:15:58 crc kubenswrapper[4860]: I1211 08:15:58.991935 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 11 08:15:59 crc kubenswrapper[4860]: I1211 08:15:59.303948 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 11 08:15:59 crc kubenswrapper[4860]: I1211 08:15:59.337404 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 11 08:15:59 crc kubenswrapper[4860]: I1211 08:15:59.354579 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 11 08:15:59 crc kubenswrapper[4860]: I1211 08:15:59.354781 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 11 08:15:59 crc kubenswrapper[4860]: I1211 08:15:59.407729 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 11 08:15:59 crc kubenswrapper[4860]: I1211 08:15:59.461572 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 11 08:15:59 crc kubenswrapper[4860]: I1211 08:15:59.554280 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 11 08:15:59 crc kubenswrapper[4860]: I1211 08:15:59.602958 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 11 08:15:59 crc kubenswrapper[4860]: I1211 08:15:59.676121 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 11 08:15:59 crc kubenswrapper[4860]: I1211 08:15:59.714846 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 11 08:15:59 crc kubenswrapper[4860]: I1211 08:15:59.796705 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 11 08:15:59 crc kubenswrapper[4860]: I1211 08:15:59.904569 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 11 08:16:00 crc kubenswrapper[4860]: I1211 08:16:00.041416 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 11 08:16:00 crc kubenswrapper[4860]: I1211 08:16:00.252941 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 11 08:16:00 crc kubenswrapper[4860]: I1211 08:16:00.322288 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 11 08:16:00 crc kubenswrapper[4860]: I1211 08:16:00.350536 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 11 08:16:00 crc kubenswrapper[4860]: I1211 08:16:00.455005 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 11 08:16:00 crc kubenswrapper[4860]: I1211 08:16:00.479422 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 11 08:16:00 crc kubenswrapper[4860]: I1211 08:16:00.572105 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 11 08:16:00 crc kubenswrapper[4860]: I1211 08:16:00.639860 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 11 08:16:00 crc kubenswrapper[4860]: I1211 08:16:00.714403 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 11 08:16:00 crc kubenswrapper[4860]: I1211 08:16:00.778858 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 11 08:16:00 crc kubenswrapper[4860]: I1211 08:16:00.788528 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 11 08:16:00 crc kubenswrapper[4860]: I1211 08:16:00.955960 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 11 08:16:00 crc kubenswrapper[4860]: I1211 08:16:00.956176 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 11 08:16:01 crc kubenswrapper[4860]: I1211 08:16:01.037063 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 11 08:16:01 crc kubenswrapper[4860]: I1211 08:16:01.081191 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 11 08:16:01 crc kubenswrapper[4860]: I1211 08:16:01.141320 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 11 08:16:01 crc kubenswrapper[4860]: I1211 08:16:01.147837 4860 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 11 08:16:01 crc kubenswrapper[4860]: I1211 08:16:01.150976 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 11 08:16:01 crc kubenswrapper[4860]: I1211 08:16:01.212320 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 11 08:16:01 crc kubenswrapper[4860]: I1211 08:16:01.385247 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 11 08:16:01 crc kubenswrapper[4860]: I1211 08:16:01.531921 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 11 08:16:01 crc kubenswrapper[4860]: I1211 08:16:01.646500 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 11 08:16:01 crc kubenswrapper[4860]: I1211 08:16:01.756170 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 11 08:16:01 crc kubenswrapper[4860]: I1211 08:16:01.969435 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 11 08:16:02 crc kubenswrapper[4860]: I1211 08:16:02.184606 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 11 08:16:02 crc kubenswrapper[4860]: I1211 08:16:02.515611 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 11 08:16:02 crc kubenswrapper[4860]: I1211 08:16:02.878910 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 11 08:16:03 crc kubenswrapper[4860]: I1211 08:16:03.253107 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 11 08:16:03 crc kubenswrapper[4860]: I1211 08:16:03.380010 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 11 08:16:03 crc kubenswrapper[4860]: I1211 08:16:03.775811 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 11 08:16:03 crc kubenswrapper[4860]: I1211 08:16:03.775905 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 11 08:16:03 crc kubenswrapper[4860]: I1211 08:16:03.871425 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 11 08:16:03 crc kubenswrapper[4860]: I1211 08:16:03.871755 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 11 08:16:03 crc kubenswrapper[4860]: I1211 08:16:03.871808 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 11 08:16:03 crc kubenswrapper[4860]: I1211 08:16:03.871872 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 11 08:16:03 crc kubenswrapper[4860]: I1211 08:16:03.871907 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 11 08:16:03 crc kubenswrapper[4860]: I1211 08:16:03.871572 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 08:16:03 crc kubenswrapper[4860]: I1211 08:16:03.872128 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 08:16:03 crc kubenswrapper[4860]: I1211 08:16:03.872736 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 08:16:03 crc kubenswrapper[4860]: I1211 08:16:03.872790 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 08:16:03 crc kubenswrapper[4860]: I1211 08:16:03.880891 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 08:16:03 crc kubenswrapper[4860]: I1211 08:16:03.973049 4860 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Dec 11 08:16:03 crc kubenswrapper[4860]: I1211 08:16:03.973098 4860 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 11 08:16:03 crc kubenswrapper[4860]: I1211 08:16:03.973118 4860 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Dec 11 08:16:03 crc kubenswrapper[4860]: I1211 08:16:03.973134 4860 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Dec 11 08:16:03 crc kubenswrapper[4860]: I1211 08:16:03.973151 4860 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 11 08:16:04 crc kubenswrapper[4860]: I1211 08:16:04.167072 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 11 08:16:04 crc kubenswrapper[4860]: I1211 08:16:04.167177 4860 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="7c8d32000cd3750bd5ea7b6fd6bc17c8db195e315089cd5a1ed0023342f049aa" exitCode=137 Dec 11 08:16:04 crc kubenswrapper[4860]: I1211 08:16:04.167271 4860 scope.go:117] "RemoveContainer" containerID="7c8d32000cd3750bd5ea7b6fd6bc17c8db195e315089cd5a1ed0023342f049aa" Dec 11 08:16:04 crc kubenswrapper[4860]: I1211 08:16:04.167267 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 11 08:16:04 crc kubenswrapper[4860]: I1211 08:16:04.187182 4860 scope.go:117] "RemoveContainer" containerID="7c8d32000cd3750bd5ea7b6fd6bc17c8db195e315089cd5a1ed0023342f049aa" Dec 11 08:16:04 crc kubenswrapper[4860]: E1211 08:16:04.188298 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7c8d32000cd3750bd5ea7b6fd6bc17c8db195e315089cd5a1ed0023342f049aa\": container with ID starting with 7c8d32000cd3750bd5ea7b6fd6bc17c8db195e315089cd5a1ed0023342f049aa not found: ID does not exist" containerID="7c8d32000cd3750bd5ea7b6fd6bc17c8db195e315089cd5a1ed0023342f049aa" Dec 11 08:16:04 crc kubenswrapper[4860]: I1211 08:16:04.188380 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7c8d32000cd3750bd5ea7b6fd6bc17c8db195e315089cd5a1ed0023342f049aa"} err="failed to get container status \"7c8d32000cd3750bd5ea7b6fd6bc17c8db195e315089cd5a1ed0023342f049aa\": rpc error: code = NotFound desc = could not find container \"7c8d32000cd3750bd5ea7b6fd6bc17c8db195e315089cd5a1ed0023342f049aa\": container with ID starting with 7c8d32000cd3750bd5ea7b6fd6bc17c8db195e315089cd5a1ed0023342f049aa not found: ID does not exist" Dec 11 08:16:05 crc kubenswrapper[4860]: I1211 08:16:05.585718 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Dec 11 08:16:07 crc kubenswrapper[4860]: I1211 08:16:07.429292 4860 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials Dec 11 08:16:18 crc kubenswrapper[4860]: I1211 08:16:18.561358 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-24dmg"] Dec 11 08:16:18 crc kubenswrapper[4860]: I1211 08:16:18.562247 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-24dmg" podUID="b7feb2dc-f1f2-4228-9974-8a771fe98f3c" containerName="controller-manager" containerID="cri-o://5b3be03c9a9ede15a987ac40ff9f744b5f05266925533e6a30991b110196cf7e" gracePeriod=30 Dec 11 08:16:18 crc kubenswrapper[4860]: I1211 08:16:18.665440 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-87b7k"] Dec 11 08:16:18 crc kubenswrapper[4860]: I1211 08:16:18.666421 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-87b7k" podUID="73ab8adf-cf47-4482-bbdb-8fb6c3edc87a" containerName="route-controller-manager" containerID="cri-o://ed658c6c2c4c4a661b4f95ad3cdd1ddf32c8e29375acb39b8f4467ba79fb6113" gracePeriod=30 Dec 11 08:16:18 crc kubenswrapper[4860]: I1211 08:16:18.997294 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-87b7k" Dec 11 08:16:19 crc kubenswrapper[4860]: I1211 08:16:19.187176 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vmgfp\" (UniqueName: \"kubernetes.io/projected/73ab8adf-cf47-4482-bbdb-8fb6c3edc87a-kube-api-access-vmgfp\") pod \"73ab8adf-cf47-4482-bbdb-8fb6c3edc87a\" (UID: \"73ab8adf-cf47-4482-bbdb-8fb6c3edc87a\") " Dec 11 08:16:19 crc kubenswrapper[4860]: I1211 08:16:19.187252 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/73ab8adf-cf47-4482-bbdb-8fb6c3edc87a-config\") pod \"73ab8adf-cf47-4482-bbdb-8fb6c3edc87a\" (UID: \"73ab8adf-cf47-4482-bbdb-8fb6c3edc87a\") " Dec 11 08:16:19 crc kubenswrapper[4860]: I1211 08:16:19.187288 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/73ab8adf-cf47-4482-bbdb-8fb6c3edc87a-serving-cert\") pod \"73ab8adf-cf47-4482-bbdb-8fb6c3edc87a\" (UID: \"73ab8adf-cf47-4482-bbdb-8fb6c3edc87a\") " Dec 11 08:16:19 crc kubenswrapper[4860]: I1211 08:16:19.187405 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/73ab8adf-cf47-4482-bbdb-8fb6c3edc87a-client-ca\") pod \"73ab8adf-cf47-4482-bbdb-8fb6c3edc87a\" (UID: \"73ab8adf-cf47-4482-bbdb-8fb6c3edc87a\") " Dec 11 08:16:19 crc kubenswrapper[4860]: I1211 08:16:19.188172 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/73ab8adf-cf47-4482-bbdb-8fb6c3edc87a-config" (OuterVolumeSpecName: "config") pod "73ab8adf-cf47-4482-bbdb-8fb6c3edc87a" (UID: "73ab8adf-cf47-4482-bbdb-8fb6c3edc87a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:16:19 crc kubenswrapper[4860]: I1211 08:16:19.188225 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/73ab8adf-cf47-4482-bbdb-8fb6c3edc87a-client-ca" (OuterVolumeSpecName: "client-ca") pod "73ab8adf-cf47-4482-bbdb-8fb6c3edc87a" (UID: "73ab8adf-cf47-4482-bbdb-8fb6c3edc87a"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:16:19 crc kubenswrapper[4860]: I1211 08:16:19.193026 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73ab8adf-cf47-4482-bbdb-8fb6c3edc87a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "73ab8adf-cf47-4482-bbdb-8fb6c3edc87a" (UID: "73ab8adf-cf47-4482-bbdb-8fb6c3edc87a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:16:19 crc kubenswrapper[4860]: I1211 08:16:19.195684 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/73ab8adf-cf47-4482-bbdb-8fb6c3edc87a-kube-api-access-vmgfp" (OuterVolumeSpecName: "kube-api-access-vmgfp") pod "73ab8adf-cf47-4482-bbdb-8fb6c3edc87a" (UID: "73ab8adf-cf47-4482-bbdb-8fb6c3edc87a"). InnerVolumeSpecName "kube-api-access-vmgfp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:16:19 crc kubenswrapper[4860]: I1211 08:16:19.249500 4860 generic.go:334] "Generic (PLEG): container finished" podID="73ab8adf-cf47-4482-bbdb-8fb6c3edc87a" containerID="ed658c6c2c4c4a661b4f95ad3cdd1ddf32c8e29375acb39b8f4467ba79fb6113" exitCode=0 Dec 11 08:16:19 crc kubenswrapper[4860]: I1211 08:16:19.249563 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-87b7k" event={"ID":"73ab8adf-cf47-4482-bbdb-8fb6c3edc87a","Type":"ContainerDied","Data":"ed658c6c2c4c4a661b4f95ad3cdd1ddf32c8e29375acb39b8f4467ba79fb6113"} Dec 11 08:16:19 crc kubenswrapper[4860]: I1211 08:16:19.249589 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-87b7k" event={"ID":"73ab8adf-cf47-4482-bbdb-8fb6c3edc87a","Type":"ContainerDied","Data":"8f5aadd283eb92b3ccba02c57fb38a8038de74ed2fe0d3b892fdfe51df4e0a69"} Dec 11 08:16:19 crc kubenswrapper[4860]: I1211 08:16:19.249590 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-87b7k" Dec 11 08:16:19 crc kubenswrapper[4860]: I1211 08:16:19.249607 4860 scope.go:117] "RemoveContainer" containerID="ed658c6c2c4c4a661b4f95ad3cdd1ddf32c8e29375acb39b8f4467ba79fb6113" Dec 11 08:16:19 crc kubenswrapper[4860]: I1211 08:16:19.251031 4860 generic.go:334] "Generic (PLEG): container finished" podID="b7feb2dc-f1f2-4228-9974-8a771fe98f3c" containerID="5b3be03c9a9ede15a987ac40ff9f744b5f05266925533e6a30991b110196cf7e" exitCode=0 Dec 11 08:16:19 crc kubenswrapper[4860]: I1211 08:16:19.251049 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-24dmg" event={"ID":"b7feb2dc-f1f2-4228-9974-8a771fe98f3c","Type":"ContainerDied","Data":"5b3be03c9a9ede15a987ac40ff9f744b5f05266925533e6a30991b110196cf7e"} Dec 11 08:16:19 crc kubenswrapper[4860]: I1211 08:16:19.270619 4860 scope.go:117] "RemoveContainer" containerID="ed658c6c2c4c4a661b4f95ad3cdd1ddf32c8e29375acb39b8f4467ba79fb6113" Dec 11 08:16:19 crc kubenswrapper[4860]: E1211 08:16:19.271071 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ed658c6c2c4c4a661b4f95ad3cdd1ddf32c8e29375acb39b8f4467ba79fb6113\": container with ID starting with ed658c6c2c4c4a661b4f95ad3cdd1ddf32c8e29375acb39b8f4467ba79fb6113 not found: ID does not exist" containerID="ed658c6c2c4c4a661b4f95ad3cdd1ddf32c8e29375acb39b8f4467ba79fb6113" Dec 11 08:16:19 crc kubenswrapper[4860]: I1211 08:16:19.271112 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed658c6c2c4c4a661b4f95ad3cdd1ddf32c8e29375acb39b8f4467ba79fb6113"} err="failed to get container status \"ed658c6c2c4c4a661b4f95ad3cdd1ddf32c8e29375acb39b8f4467ba79fb6113\": rpc error: code = NotFound desc = could not find container \"ed658c6c2c4c4a661b4f95ad3cdd1ddf32c8e29375acb39b8f4467ba79fb6113\": container with ID starting with ed658c6c2c4c4a661b4f95ad3cdd1ddf32c8e29375acb39b8f4467ba79fb6113 not found: ID does not exist" Dec 11 08:16:19 crc kubenswrapper[4860]: I1211 08:16:19.284630 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-87b7k"] Dec 11 08:16:19 crc kubenswrapper[4860]: I1211 08:16:19.290376 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vmgfp\" (UniqueName: \"kubernetes.io/projected/73ab8adf-cf47-4482-bbdb-8fb6c3edc87a-kube-api-access-vmgfp\") on node \"crc\" DevicePath \"\"" Dec 11 08:16:19 crc kubenswrapper[4860]: I1211 08:16:19.290411 4860 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/73ab8adf-cf47-4482-bbdb-8fb6c3edc87a-config\") on node \"crc\" DevicePath \"\"" Dec 11 08:16:19 crc kubenswrapper[4860]: I1211 08:16:19.290426 4860 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/73ab8adf-cf47-4482-bbdb-8fb6c3edc87a-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 08:16:19 crc kubenswrapper[4860]: I1211 08:16:19.290439 4860 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/73ab8adf-cf47-4482-bbdb-8fb6c3edc87a-client-ca\") on node \"crc\" DevicePath \"\"" Dec 11 08:16:19 crc kubenswrapper[4860]: I1211 08:16:19.290731 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-87b7k"] Dec 11 08:16:19 crc kubenswrapper[4860]: I1211 08:16:19.364540 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-24dmg" Dec 11 08:16:19 crc kubenswrapper[4860]: I1211 08:16:19.494771 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b7feb2dc-f1f2-4228-9974-8a771fe98f3c-proxy-ca-bundles\") pod \"b7feb2dc-f1f2-4228-9974-8a771fe98f3c\" (UID: \"b7feb2dc-f1f2-4228-9974-8a771fe98f3c\") " Dec 11 08:16:19 crc kubenswrapper[4860]: I1211 08:16:19.494895 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hkdsn\" (UniqueName: \"kubernetes.io/projected/b7feb2dc-f1f2-4228-9974-8a771fe98f3c-kube-api-access-hkdsn\") pod \"b7feb2dc-f1f2-4228-9974-8a771fe98f3c\" (UID: \"b7feb2dc-f1f2-4228-9974-8a771fe98f3c\") " Dec 11 08:16:19 crc kubenswrapper[4860]: I1211 08:16:19.494992 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b7feb2dc-f1f2-4228-9974-8a771fe98f3c-serving-cert\") pod \"b7feb2dc-f1f2-4228-9974-8a771fe98f3c\" (UID: \"b7feb2dc-f1f2-4228-9974-8a771fe98f3c\") " Dec 11 08:16:19 crc kubenswrapper[4860]: I1211 08:16:19.495057 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b7feb2dc-f1f2-4228-9974-8a771fe98f3c-client-ca\") pod \"b7feb2dc-f1f2-4228-9974-8a771fe98f3c\" (UID: \"b7feb2dc-f1f2-4228-9974-8a771fe98f3c\") " Dec 11 08:16:19 crc kubenswrapper[4860]: I1211 08:16:19.495104 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b7feb2dc-f1f2-4228-9974-8a771fe98f3c-config\") pod \"b7feb2dc-f1f2-4228-9974-8a771fe98f3c\" (UID: \"b7feb2dc-f1f2-4228-9974-8a771fe98f3c\") " Dec 11 08:16:19 crc kubenswrapper[4860]: I1211 08:16:19.496072 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b7feb2dc-f1f2-4228-9974-8a771fe98f3c-client-ca" (OuterVolumeSpecName: "client-ca") pod "b7feb2dc-f1f2-4228-9974-8a771fe98f3c" (UID: "b7feb2dc-f1f2-4228-9974-8a771fe98f3c"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:16:19 crc kubenswrapper[4860]: I1211 08:16:19.496198 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b7feb2dc-f1f2-4228-9974-8a771fe98f3c-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "b7feb2dc-f1f2-4228-9974-8a771fe98f3c" (UID: "b7feb2dc-f1f2-4228-9974-8a771fe98f3c"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:16:19 crc kubenswrapper[4860]: I1211 08:16:19.496208 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b7feb2dc-f1f2-4228-9974-8a771fe98f3c-config" (OuterVolumeSpecName: "config") pod "b7feb2dc-f1f2-4228-9974-8a771fe98f3c" (UID: "b7feb2dc-f1f2-4228-9974-8a771fe98f3c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:16:19 crc kubenswrapper[4860]: I1211 08:16:19.499413 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b7feb2dc-f1f2-4228-9974-8a771fe98f3c-kube-api-access-hkdsn" (OuterVolumeSpecName: "kube-api-access-hkdsn") pod "b7feb2dc-f1f2-4228-9974-8a771fe98f3c" (UID: "b7feb2dc-f1f2-4228-9974-8a771fe98f3c"). InnerVolumeSpecName "kube-api-access-hkdsn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:16:19 crc kubenswrapper[4860]: I1211 08:16:19.499769 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b7feb2dc-f1f2-4228-9974-8a771fe98f3c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "b7feb2dc-f1f2-4228-9974-8a771fe98f3c" (UID: "b7feb2dc-f1f2-4228-9974-8a771fe98f3c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:16:19 crc kubenswrapper[4860]: I1211 08:16:19.588307 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="73ab8adf-cf47-4482-bbdb-8fb6c3edc87a" path="/var/lib/kubelet/pods/73ab8adf-cf47-4482-bbdb-8fb6c3edc87a/volumes" Dec 11 08:16:19 crc kubenswrapper[4860]: I1211 08:16:19.596698 4860 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b7feb2dc-f1f2-4228-9974-8a771fe98f3c-client-ca\") on node \"crc\" DevicePath \"\"" Dec 11 08:16:19 crc kubenswrapper[4860]: I1211 08:16:19.596773 4860 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b7feb2dc-f1f2-4228-9974-8a771fe98f3c-config\") on node \"crc\" DevicePath \"\"" Dec 11 08:16:19 crc kubenswrapper[4860]: I1211 08:16:19.596800 4860 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b7feb2dc-f1f2-4228-9974-8a771fe98f3c-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 11 08:16:19 crc kubenswrapper[4860]: I1211 08:16:19.596828 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hkdsn\" (UniqueName: \"kubernetes.io/projected/b7feb2dc-f1f2-4228-9974-8a771fe98f3c-kube-api-access-hkdsn\") on node \"crc\" DevicePath \"\"" Dec 11 08:16:19 crc kubenswrapper[4860]: I1211 08:16:19.596853 4860 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b7feb2dc-f1f2-4228-9974-8a771fe98f3c-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 08:16:19 crc kubenswrapper[4860]: I1211 08:16:19.630928 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-5c997b8cc9-q7sqj"] Dec 11 08:16:19 crc kubenswrapper[4860]: E1211 08:16:19.631296 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ac89442-1754-45d3-b67a-aa6cc31a1235" containerName="extract-utilities" Dec 11 08:16:19 crc kubenswrapper[4860]: I1211 08:16:19.631329 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ac89442-1754-45d3-b67a-aa6cc31a1235" containerName="extract-utilities" Dec 11 08:16:19 crc kubenswrapper[4860]: E1211 08:16:19.631344 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e052d7c-59d1-43a4-ac7a-e9daf2a88c65" containerName="extract-content" Dec 11 08:16:19 crc kubenswrapper[4860]: I1211 08:16:19.631357 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e052d7c-59d1-43a4-ac7a-e9daf2a88c65" containerName="extract-content" Dec 11 08:16:19 crc kubenswrapper[4860]: E1211 08:16:19.631374 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73ab8adf-cf47-4482-bbdb-8fb6c3edc87a" containerName="route-controller-manager" Dec 11 08:16:19 crc kubenswrapper[4860]: I1211 08:16:19.631386 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="73ab8adf-cf47-4482-bbdb-8fb6c3edc87a" containerName="route-controller-manager" Dec 11 08:16:19 crc kubenswrapper[4860]: E1211 08:16:19.631397 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30107273-766a-4575-8cb1-3de387248437" containerName="extract-content" Dec 11 08:16:19 crc kubenswrapper[4860]: I1211 08:16:19.631406 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="30107273-766a-4575-8cb1-3de387248437" containerName="extract-content" Dec 11 08:16:19 crc kubenswrapper[4860]: E1211 08:16:19.631417 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30107273-766a-4575-8cb1-3de387248437" containerName="registry-server" Dec 11 08:16:19 crc kubenswrapper[4860]: I1211 08:16:19.631425 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="30107273-766a-4575-8cb1-3de387248437" containerName="registry-server" Dec 11 08:16:19 crc kubenswrapper[4860]: E1211 08:16:19.631437 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01481957-a690-40f4-8c89-352b7350d327" containerName="registry-server" Dec 11 08:16:19 crc kubenswrapper[4860]: I1211 08:16:19.631448 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="01481957-a690-40f4-8c89-352b7350d327" containerName="registry-server" Dec 11 08:16:19 crc kubenswrapper[4860]: E1211 08:16:19.631467 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="294792dd-5104-4308-a10f-f3ab0c4504e9" containerName="registry-server" Dec 11 08:16:19 crc kubenswrapper[4860]: I1211 08:16:19.631478 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="294792dd-5104-4308-a10f-f3ab0c4504e9" containerName="registry-server" Dec 11 08:16:19 crc kubenswrapper[4860]: E1211 08:16:19.631496 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7feb2dc-f1f2-4228-9974-8a771fe98f3c" containerName="controller-manager" Dec 11 08:16:19 crc kubenswrapper[4860]: I1211 08:16:19.631506 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7feb2dc-f1f2-4228-9974-8a771fe98f3c" containerName="controller-manager" Dec 11 08:16:19 crc kubenswrapper[4860]: E1211 08:16:19.631517 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad724960-18d1-4e83-944c-4cdeb033a436" containerName="marketplace-operator" Dec 11 08:16:19 crc kubenswrapper[4860]: I1211 08:16:19.631528 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad724960-18d1-4e83-944c-4cdeb033a436" containerName="marketplace-operator" Dec 11 08:16:19 crc kubenswrapper[4860]: E1211 08:16:19.631542 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="294792dd-5104-4308-a10f-f3ab0c4504e9" containerName="extract-content" Dec 11 08:16:19 crc kubenswrapper[4860]: I1211 08:16:19.631553 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="294792dd-5104-4308-a10f-f3ab0c4504e9" containerName="extract-content" Dec 11 08:16:19 crc kubenswrapper[4860]: E1211 08:16:19.631568 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="093a9903-7041-410b-bfb2-c74ef741e0be" containerName="extract-content" Dec 11 08:16:19 crc kubenswrapper[4860]: I1211 08:16:19.631581 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="093a9903-7041-410b-bfb2-c74ef741e0be" containerName="extract-content" Dec 11 08:16:19 crc kubenswrapper[4860]: E1211 08:16:19.631600 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01481957-a690-40f4-8c89-352b7350d327" containerName="extract-utilities" Dec 11 08:16:19 crc kubenswrapper[4860]: I1211 08:16:19.631612 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="01481957-a690-40f4-8c89-352b7350d327" containerName="extract-utilities" Dec 11 08:16:19 crc kubenswrapper[4860]: E1211 08:16:19.631626 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01481957-a690-40f4-8c89-352b7350d327" containerName="extract-content" Dec 11 08:16:19 crc kubenswrapper[4860]: I1211 08:16:19.631637 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="01481957-a690-40f4-8c89-352b7350d327" containerName="extract-content" Dec 11 08:16:19 crc kubenswrapper[4860]: E1211 08:16:19.631670 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ac89442-1754-45d3-b67a-aa6cc31a1235" containerName="extract-content" Dec 11 08:16:19 crc kubenswrapper[4860]: I1211 08:16:19.631679 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ac89442-1754-45d3-b67a-aa6cc31a1235" containerName="extract-content" Dec 11 08:16:19 crc kubenswrapper[4860]: E1211 08:16:19.631690 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e052d7c-59d1-43a4-ac7a-e9daf2a88c65" containerName="extract-utilities" Dec 11 08:16:19 crc kubenswrapper[4860]: I1211 08:16:19.631699 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e052d7c-59d1-43a4-ac7a-e9daf2a88c65" containerName="extract-utilities" Dec 11 08:16:19 crc kubenswrapper[4860]: E1211 08:16:19.631711 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="294792dd-5104-4308-a10f-f3ab0c4504e9" containerName="extract-utilities" Dec 11 08:16:19 crc kubenswrapper[4860]: I1211 08:16:19.631721 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="294792dd-5104-4308-a10f-f3ab0c4504e9" containerName="extract-utilities" Dec 11 08:16:19 crc kubenswrapper[4860]: E1211 08:16:19.631733 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 11 08:16:19 crc kubenswrapper[4860]: I1211 08:16:19.631741 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 11 08:16:19 crc kubenswrapper[4860]: E1211 08:16:19.631753 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e052d7c-59d1-43a4-ac7a-e9daf2a88c65" containerName="registry-server" Dec 11 08:16:19 crc kubenswrapper[4860]: I1211 08:16:19.631761 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e052d7c-59d1-43a4-ac7a-e9daf2a88c65" containerName="registry-server" Dec 11 08:16:19 crc kubenswrapper[4860]: E1211 08:16:19.631774 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="093a9903-7041-410b-bfb2-c74ef741e0be" containerName="registry-server" Dec 11 08:16:19 crc kubenswrapper[4860]: I1211 08:16:19.631783 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="093a9903-7041-410b-bfb2-c74ef741e0be" containerName="registry-server" Dec 11 08:16:19 crc kubenswrapper[4860]: E1211 08:16:19.631793 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30107273-766a-4575-8cb1-3de387248437" containerName="extract-utilities" Dec 11 08:16:19 crc kubenswrapper[4860]: I1211 08:16:19.631801 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="30107273-766a-4575-8cb1-3de387248437" containerName="extract-utilities" Dec 11 08:16:19 crc kubenswrapper[4860]: E1211 08:16:19.631812 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="093a9903-7041-410b-bfb2-c74ef741e0be" containerName="extract-utilities" Dec 11 08:16:19 crc kubenswrapper[4860]: I1211 08:16:19.631820 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="093a9903-7041-410b-bfb2-c74ef741e0be" containerName="extract-utilities" Dec 11 08:16:19 crc kubenswrapper[4860]: E1211 08:16:19.631833 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ac89442-1754-45d3-b67a-aa6cc31a1235" containerName="registry-server" Dec 11 08:16:19 crc kubenswrapper[4860]: I1211 08:16:19.631841 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ac89442-1754-45d3-b67a-aa6cc31a1235" containerName="registry-server" Dec 11 08:16:19 crc kubenswrapper[4860]: I1211 08:16:19.631958 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad724960-18d1-4e83-944c-4cdeb033a436" containerName="marketplace-operator" Dec 11 08:16:19 crc kubenswrapper[4860]: I1211 08:16:19.631970 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="73ab8adf-cf47-4482-bbdb-8fb6c3edc87a" containerName="route-controller-manager" Dec 11 08:16:19 crc kubenswrapper[4860]: I1211 08:16:19.631983 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 11 08:16:19 crc kubenswrapper[4860]: I1211 08:16:19.631997 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e052d7c-59d1-43a4-ac7a-e9daf2a88c65" containerName="registry-server" Dec 11 08:16:19 crc kubenswrapper[4860]: I1211 08:16:19.632009 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="30107273-766a-4575-8cb1-3de387248437" containerName="registry-server" Dec 11 08:16:19 crc kubenswrapper[4860]: I1211 08:16:19.632019 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="01481957-a690-40f4-8c89-352b7350d327" containerName="registry-server" Dec 11 08:16:19 crc kubenswrapper[4860]: I1211 08:16:19.632029 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="294792dd-5104-4308-a10f-f3ab0c4504e9" containerName="registry-server" Dec 11 08:16:19 crc kubenswrapper[4860]: I1211 08:16:19.632044 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="093a9903-7041-410b-bfb2-c74ef741e0be" containerName="registry-server" Dec 11 08:16:19 crc kubenswrapper[4860]: I1211 08:16:19.632052 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ac89442-1754-45d3-b67a-aa6cc31a1235" containerName="registry-server" Dec 11 08:16:19 crc kubenswrapper[4860]: I1211 08:16:19.632066 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7feb2dc-f1f2-4228-9974-8a771fe98f3c" containerName="controller-manager" Dec 11 08:16:19 crc kubenswrapper[4860]: I1211 08:16:19.632545 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5c997b8cc9-q7sqj" Dec 11 08:16:19 crc kubenswrapper[4860]: I1211 08:16:19.639587 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5c997b8cc9-q7sqj"] Dec 11 08:16:19 crc kubenswrapper[4860]: I1211 08:16:19.799701 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7bc93e1d-1531-472b-974f-201cc58dbbf5-serving-cert\") pod \"controller-manager-5c997b8cc9-q7sqj\" (UID: \"7bc93e1d-1531-472b-974f-201cc58dbbf5\") " pod="openshift-controller-manager/controller-manager-5c997b8cc9-q7sqj" Dec 11 08:16:19 crc kubenswrapper[4860]: I1211 08:16:19.799804 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7bc93e1d-1531-472b-974f-201cc58dbbf5-config\") pod \"controller-manager-5c997b8cc9-q7sqj\" (UID: \"7bc93e1d-1531-472b-974f-201cc58dbbf5\") " pod="openshift-controller-manager/controller-manager-5c997b8cc9-q7sqj" Dec 11 08:16:19 crc kubenswrapper[4860]: I1211 08:16:19.800386 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zwrwx\" (UniqueName: \"kubernetes.io/projected/7bc93e1d-1531-472b-974f-201cc58dbbf5-kube-api-access-zwrwx\") pod \"controller-manager-5c997b8cc9-q7sqj\" (UID: \"7bc93e1d-1531-472b-974f-201cc58dbbf5\") " pod="openshift-controller-manager/controller-manager-5c997b8cc9-q7sqj" Dec 11 08:16:19 crc kubenswrapper[4860]: I1211 08:16:19.800465 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7bc93e1d-1531-472b-974f-201cc58dbbf5-client-ca\") pod \"controller-manager-5c997b8cc9-q7sqj\" (UID: \"7bc93e1d-1531-472b-974f-201cc58dbbf5\") " pod="openshift-controller-manager/controller-manager-5c997b8cc9-q7sqj" Dec 11 08:16:19 crc kubenswrapper[4860]: I1211 08:16:19.800813 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7bc93e1d-1531-472b-974f-201cc58dbbf5-proxy-ca-bundles\") pod \"controller-manager-5c997b8cc9-q7sqj\" (UID: \"7bc93e1d-1531-472b-974f-201cc58dbbf5\") " pod="openshift-controller-manager/controller-manager-5c997b8cc9-q7sqj" Dec 11 08:16:19 crc kubenswrapper[4860]: I1211 08:16:19.902522 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7bc93e1d-1531-472b-974f-201cc58dbbf5-proxy-ca-bundles\") pod \"controller-manager-5c997b8cc9-q7sqj\" (UID: \"7bc93e1d-1531-472b-974f-201cc58dbbf5\") " pod="openshift-controller-manager/controller-manager-5c997b8cc9-q7sqj" Dec 11 08:16:19 crc kubenswrapper[4860]: I1211 08:16:19.902632 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7bc93e1d-1531-472b-974f-201cc58dbbf5-serving-cert\") pod \"controller-manager-5c997b8cc9-q7sqj\" (UID: \"7bc93e1d-1531-472b-974f-201cc58dbbf5\") " pod="openshift-controller-manager/controller-manager-5c997b8cc9-q7sqj" Dec 11 08:16:19 crc kubenswrapper[4860]: I1211 08:16:19.902705 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7bc93e1d-1531-472b-974f-201cc58dbbf5-config\") pod \"controller-manager-5c997b8cc9-q7sqj\" (UID: \"7bc93e1d-1531-472b-974f-201cc58dbbf5\") " pod="openshift-controller-manager/controller-manager-5c997b8cc9-q7sqj" Dec 11 08:16:19 crc kubenswrapper[4860]: I1211 08:16:19.902794 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zwrwx\" (UniqueName: \"kubernetes.io/projected/7bc93e1d-1531-472b-974f-201cc58dbbf5-kube-api-access-zwrwx\") pod \"controller-manager-5c997b8cc9-q7sqj\" (UID: \"7bc93e1d-1531-472b-974f-201cc58dbbf5\") " pod="openshift-controller-manager/controller-manager-5c997b8cc9-q7sqj" Dec 11 08:16:19 crc kubenswrapper[4860]: I1211 08:16:19.902842 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7bc93e1d-1531-472b-974f-201cc58dbbf5-client-ca\") pod \"controller-manager-5c997b8cc9-q7sqj\" (UID: \"7bc93e1d-1531-472b-974f-201cc58dbbf5\") " pod="openshift-controller-manager/controller-manager-5c997b8cc9-q7sqj" Dec 11 08:16:19 crc kubenswrapper[4860]: I1211 08:16:19.904882 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7bc93e1d-1531-472b-974f-201cc58dbbf5-client-ca\") pod \"controller-manager-5c997b8cc9-q7sqj\" (UID: \"7bc93e1d-1531-472b-974f-201cc58dbbf5\") " pod="openshift-controller-manager/controller-manager-5c997b8cc9-q7sqj" Dec 11 08:16:19 crc kubenswrapper[4860]: I1211 08:16:19.904945 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7bc93e1d-1531-472b-974f-201cc58dbbf5-proxy-ca-bundles\") pod \"controller-manager-5c997b8cc9-q7sqj\" (UID: \"7bc93e1d-1531-472b-974f-201cc58dbbf5\") " pod="openshift-controller-manager/controller-manager-5c997b8cc9-q7sqj" Dec 11 08:16:19 crc kubenswrapper[4860]: I1211 08:16:19.906486 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7bc93e1d-1531-472b-974f-201cc58dbbf5-config\") pod \"controller-manager-5c997b8cc9-q7sqj\" (UID: \"7bc93e1d-1531-472b-974f-201cc58dbbf5\") " pod="openshift-controller-manager/controller-manager-5c997b8cc9-q7sqj" Dec 11 08:16:19 crc kubenswrapper[4860]: I1211 08:16:19.908867 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7bc93e1d-1531-472b-974f-201cc58dbbf5-serving-cert\") pod \"controller-manager-5c997b8cc9-q7sqj\" (UID: \"7bc93e1d-1531-472b-974f-201cc58dbbf5\") " pod="openshift-controller-manager/controller-manager-5c997b8cc9-q7sqj" Dec 11 08:16:19 crc kubenswrapper[4860]: I1211 08:16:19.935620 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zwrwx\" (UniqueName: \"kubernetes.io/projected/7bc93e1d-1531-472b-974f-201cc58dbbf5-kube-api-access-zwrwx\") pod \"controller-manager-5c997b8cc9-q7sqj\" (UID: \"7bc93e1d-1531-472b-974f-201cc58dbbf5\") " pod="openshift-controller-manager/controller-manager-5c997b8cc9-q7sqj" Dec 11 08:16:19 crc kubenswrapper[4860]: I1211 08:16:19.970302 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5c997b8cc9-q7sqj" Dec 11 08:16:20 crc kubenswrapper[4860]: I1211 08:16:20.192733 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5c997b8cc9-q7sqj"] Dec 11 08:16:20 crc kubenswrapper[4860]: I1211 08:16:20.258978 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-24dmg" event={"ID":"b7feb2dc-f1f2-4228-9974-8a771fe98f3c","Type":"ContainerDied","Data":"9ba7a684a61fa08b505109d09f08a0e2f9b26c3043f14821e92cbf51299f2cd0"} Dec 11 08:16:20 crc kubenswrapper[4860]: I1211 08:16:20.258988 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-24dmg" Dec 11 08:16:20 crc kubenswrapper[4860]: I1211 08:16:20.259887 4860 scope.go:117] "RemoveContainer" containerID="5b3be03c9a9ede15a987ac40ff9f744b5f05266925533e6a30991b110196cf7e" Dec 11 08:16:20 crc kubenswrapper[4860]: I1211 08:16:20.263385 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5c997b8cc9-q7sqj" event={"ID":"7bc93e1d-1531-472b-974f-201cc58dbbf5","Type":"ContainerStarted","Data":"b0041fae7923eb1fc7917a55d48525bbc8250d7df6582b04d2cabfbed35326d7"} Dec 11 08:16:20 crc kubenswrapper[4860]: I1211 08:16:20.281764 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-24dmg"] Dec 11 08:16:20 crc kubenswrapper[4860]: I1211 08:16:20.287100 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-24dmg"] Dec 11 08:16:20 crc kubenswrapper[4860]: I1211 08:16:20.631057 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7688f75b54-fvbqm"] Dec 11 08:16:20 crc kubenswrapper[4860]: I1211 08:16:20.631854 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7688f75b54-fvbqm" Dec 11 08:16:20 crc kubenswrapper[4860]: I1211 08:16:20.634705 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 11 08:16:20 crc kubenswrapper[4860]: I1211 08:16:20.634995 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 11 08:16:20 crc kubenswrapper[4860]: I1211 08:16:20.635383 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 11 08:16:20 crc kubenswrapper[4860]: I1211 08:16:20.635468 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 11 08:16:20 crc kubenswrapper[4860]: I1211 08:16:20.635829 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 11 08:16:20 crc kubenswrapper[4860]: I1211 08:16:20.636029 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 11 08:16:20 crc kubenswrapper[4860]: I1211 08:16:20.647128 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7688f75b54-fvbqm"] Dec 11 08:16:20 crc kubenswrapper[4860]: I1211 08:16:20.815848 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/274516e1-d34a-4045-9a79-f8bf5912f442-serving-cert\") pod \"route-controller-manager-7688f75b54-fvbqm\" (UID: \"274516e1-d34a-4045-9a79-f8bf5912f442\") " pod="openshift-route-controller-manager/route-controller-manager-7688f75b54-fvbqm" Dec 11 08:16:20 crc kubenswrapper[4860]: I1211 08:16:20.815899 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/274516e1-d34a-4045-9a79-f8bf5912f442-client-ca\") pod \"route-controller-manager-7688f75b54-fvbqm\" (UID: \"274516e1-d34a-4045-9a79-f8bf5912f442\") " pod="openshift-route-controller-manager/route-controller-manager-7688f75b54-fvbqm" Dec 11 08:16:20 crc kubenswrapper[4860]: I1211 08:16:20.816048 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fz56g\" (UniqueName: \"kubernetes.io/projected/274516e1-d34a-4045-9a79-f8bf5912f442-kube-api-access-fz56g\") pod \"route-controller-manager-7688f75b54-fvbqm\" (UID: \"274516e1-d34a-4045-9a79-f8bf5912f442\") " pod="openshift-route-controller-manager/route-controller-manager-7688f75b54-fvbqm" Dec 11 08:16:20 crc kubenswrapper[4860]: I1211 08:16:20.816309 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/274516e1-d34a-4045-9a79-f8bf5912f442-config\") pod \"route-controller-manager-7688f75b54-fvbqm\" (UID: \"274516e1-d34a-4045-9a79-f8bf5912f442\") " pod="openshift-route-controller-manager/route-controller-manager-7688f75b54-fvbqm" Dec 11 08:16:20 crc kubenswrapper[4860]: I1211 08:16:20.918033 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/274516e1-d34a-4045-9a79-f8bf5912f442-serving-cert\") pod \"route-controller-manager-7688f75b54-fvbqm\" (UID: \"274516e1-d34a-4045-9a79-f8bf5912f442\") " pod="openshift-route-controller-manager/route-controller-manager-7688f75b54-fvbqm" Dec 11 08:16:20 crc kubenswrapper[4860]: I1211 08:16:20.918076 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/274516e1-d34a-4045-9a79-f8bf5912f442-client-ca\") pod \"route-controller-manager-7688f75b54-fvbqm\" (UID: \"274516e1-d34a-4045-9a79-f8bf5912f442\") " pod="openshift-route-controller-manager/route-controller-manager-7688f75b54-fvbqm" Dec 11 08:16:20 crc kubenswrapper[4860]: I1211 08:16:20.918098 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fz56g\" (UniqueName: \"kubernetes.io/projected/274516e1-d34a-4045-9a79-f8bf5912f442-kube-api-access-fz56g\") pod \"route-controller-manager-7688f75b54-fvbqm\" (UID: \"274516e1-d34a-4045-9a79-f8bf5912f442\") " pod="openshift-route-controller-manager/route-controller-manager-7688f75b54-fvbqm" Dec 11 08:16:20 crc kubenswrapper[4860]: I1211 08:16:20.918139 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/274516e1-d34a-4045-9a79-f8bf5912f442-config\") pod \"route-controller-manager-7688f75b54-fvbqm\" (UID: \"274516e1-d34a-4045-9a79-f8bf5912f442\") " pod="openshift-route-controller-manager/route-controller-manager-7688f75b54-fvbqm" Dec 11 08:16:20 crc kubenswrapper[4860]: I1211 08:16:20.919424 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/274516e1-d34a-4045-9a79-f8bf5912f442-config\") pod \"route-controller-manager-7688f75b54-fvbqm\" (UID: \"274516e1-d34a-4045-9a79-f8bf5912f442\") " pod="openshift-route-controller-manager/route-controller-manager-7688f75b54-fvbqm" Dec 11 08:16:20 crc kubenswrapper[4860]: I1211 08:16:20.920169 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/274516e1-d34a-4045-9a79-f8bf5912f442-client-ca\") pod \"route-controller-manager-7688f75b54-fvbqm\" (UID: \"274516e1-d34a-4045-9a79-f8bf5912f442\") " pod="openshift-route-controller-manager/route-controller-manager-7688f75b54-fvbqm" Dec 11 08:16:20 crc kubenswrapper[4860]: I1211 08:16:20.924565 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/274516e1-d34a-4045-9a79-f8bf5912f442-serving-cert\") pod \"route-controller-manager-7688f75b54-fvbqm\" (UID: \"274516e1-d34a-4045-9a79-f8bf5912f442\") " pod="openshift-route-controller-manager/route-controller-manager-7688f75b54-fvbqm" Dec 11 08:16:20 crc kubenswrapper[4860]: I1211 08:16:20.939256 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fz56g\" (UniqueName: \"kubernetes.io/projected/274516e1-d34a-4045-9a79-f8bf5912f442-kube-api-access-fz56g\") pod \"route-controller-manager-7688f75b54-fvbqm\" (UID: \"274516e1-d34a-4045-9a79-f8bf5912f442\") " pod="openshift-route-controller-manager/route-controller-manager-7688f75b54-fvbqm" Dec 11 08:16:20 crc kubenswrapper[4860]: I1211 08:16:20.947993 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7688f75b54-fvbqm" Dec 11 08:16:21 crc kubenswrapper[4860]: I1211 08:16:21.164855 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7688f75b54-fvbqm"] Dec 11 08:16:21 crc kubenswrapper[4860]: I1211 08:16:21.268295 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5c997b8cc9-q7sqj" event={"ID":"7bc93e1d-1531-472b-974f-201cc58dbbf5","Type":"ContainerStarted","Data":"7a02a44278e305b7f34d2bd50034c6045f87bd4c19ef7373a9e8559c9892630c"} Dec 11 08:16:21 crc kubenswrapper[4860]: I1211 08:16:21.268874 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-5c997b8cc9-q7sqj" Dec 11 08:16:21 crc kubenswrapper[4860]: I1211 08:16:21.273242 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7688f75b54-fvbqm" event={"ID":"274516e1-d34a-4045-9a79-f8bf5912f442","Type":"ContainerStarted","Data":"ffc739e0b47303a8ea78cb11d38dd9d86a80a6297178e0f56a694476ffbc1c19"} Dec 11 08:16:21 crc kubenswrapper[4860]: I1211 08:16:21.274240 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-5c997b8cc9-q7sqj" Dec 11 08:16:21 crc kubenswrapper[4860]: I1211 08:16:21.288141 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-5c997b8cc9-q7sqj" podStartSLOduration=3.288123737 podStartE2EDuration="3.288123737s" podCreationTimestamp="2025-12-11 08:16:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:16:21.284520411 +0000 UTC m=+314.013039486" watchObservedRunningTime="2025-12-11 08:16:21.288123737 +0000 UTC m=+314.016642792" Dec 11 08:16:21 crc kubenswrapper[4860]: I1211 08:16:21.585128 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b7feb2dc-f1f2-4228-9974-8a771fe98f3c" path="/var/lib/kubelet/pods/b7feb2dc-f1f2-4228-9974-8a771fe98f3c/volumes" Dec 11 08:16:21 crc kubenswrapper[4860]: I1211 08:16:21.782235 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-5c997b8cc9-q7sqj"] Dec 11 08:16:21 crc kubenswrapper[4860]: I1211 08:16:21.792716 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7688f75b54-fvbqm"] Dec 11 08:16:22 crc kubenswrapper[4860]: I1211 08:16:22.278804 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7688f75b54-fvbqm" event={"ID":"274516e1-d34a-4045-9a79-f8bf5912f442","Type":"ContainerStarted","Data":"3744356c8afe6a48bcd3e42f6e7079ecf96beb792792cfda8374bd7fdfbc5965"} Dec 11 08:16:22 crc kubenswrapper[4860]: I1211 08:16:22.279205 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-7688f75b54-fvbqm" Dec 11 08:16:22 crc kubenswrapper[4860]: I1211 08:16:22.287050 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-7688f75b54-fvbqm" Dec 11 08:16:22 crc kubenswrapper[4860]: I1211 08:16:22.295634 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-7688f75b54-fvbqm" podStartSLOduration=4.295616318 podStartE2EDuration="4.295616318s" podCreationTimestamp="2025-12-11 08:16:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:16:22.291106702 +0000 UTC m=+315.019625757" watchObservedRunningTime="2025-12-11 08:16:22.295616318 +0000 UTC m=+315.024135373" Dec 11 08:16:23 crc kubenswrapper[4860]: I1211 08:16:23.283510 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-7688f75b54-fvbqm" podUID="274516e1-d34a-4045-9a79-f8bf5912f442" containerName="route-controller-manager" containerID="cri-o://3744356c8afe6a48bcd3e42f6e7079ecf96beb792792cfda8374bd7fdfbc5965" gracePeriod=30 Dec 11 08:16:23 crc kubenswrapper[4860]: I1211 08:16:23.286019 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-5c997b8cc9-q7sqj" podUID="7bc93e1d-1531-472b-974f-201cc58dbbf5" containerName="controller-manager" containerID="cri-o://7a02a44278e305b7f34d2bd50034c6045f87bd4c19ef7373a9e8559c9892630c" gracePeriod=30 Dec 11 08:16:23 crc kubenswrapper[4860]: I1211 08:16:23.763932 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7688f75b54-fvbqm" Dec 11 08:16:23 crc kubenswrapper[4860]: I1211 08:16:23.768852 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5c997b8cc9-q7sqj" Dec 11 08:16:23 crc kubenswrapper[4860]: I1211 08:16:23.796768 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7bb5f84cb5-c5br2"] Dec 11 08:16:23 crc kubenswrapper[4860]: E1211 08:16:23.796974 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7bc93e1d-1531-472b-974f-201cc58dbbf5" containerName="controller-manager" Dec 11 08:16:23 crc kubenswrapper[4860]: I1211 08:16:23.796986 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="7bc93e1d-1531-472b-974f-201cc58dbbf5" containerName="controller-manager" Dec 11 08:16:23 crc kubenswrapper[4860]: E1211 08:16:23.797010 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="274516e1-d34a-4045-9a79-f8bf5912f442" containerName="route-controller-manager" Dec 11 08:16:23 crc kubenswrapper[4860]: I1211 08:16:23.797018 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="274516e1-d34a-4045-9a79-f8bf5912f442" containerName="route-controller-manager" Dec 11 08:16:23 crc kubenswrapper[4860]: I1211 08:16:23.797102 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="7bc93e1d-1531-472b-974f-201cc58dbbf5" containerName="controller-manager" Dec 11 08:16:23 crc kubenswrapper[4860]: I1211 08:16:23.797112 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="274516e1-d34a-4045-9a79-f8bf5912f442" containerName="route-controller-manager" Dec 11 08:16:23 crc kubenswrapper[4860]: I1211 08:16:23.797471 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7bb5f84cb5-c5br2" Dec 11 08:16:23 crc kubenswrapper[4860]: I1211 08:16:23.818967 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7bb5f84cb5-c5br2"] Dec 11 08:16:23 crc kubenswrapper[4860]: I1211 08:16:23.855699 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/274516e1-d34a-4045-9a79-f8bf5912f442-client-ca\") pod \"274516e1-d34a-4045-9a79-f8bf5912f442\" (UID: \"274516e1-d34a-4045-9a79-f8bf5912f442\") " Dec 11 08:16:23 crc kubenswrapper[4860]: I1211 08:16:23.855736 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/274516e1-d34a-4045-9a79-f8bf5912f442-serving-cert\") pod \"274516e1-d34a-4045-9a79-f8bf5912f442\" (UID: \"274516e1-d34a-4045-9a79-f8bf5912f442\") " Dec 11 08:16:23 crc kubenswrapper[4860]: I1211 08:16:23.855805 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/274516e1-d34a-4045-9a79-f8bf5912f442-config\") pod \"274516e1-d34a-4045-9a79-f8bf5912f442\" (UID: \"274516e1-d34a-4045-9a79-f8bf5912f442\") " Dec 11 08:16:23 crc kubenswrapper[4860]: I1211 08:16:23.855834 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fz56g\" (UniqueName: \"kubernetes.io/projected/274516e1-d34a-4045-9a79-f8bf5912f442-kube-api-access-fz56g\") pod \"274516e1-d34a-4045-9a79-f8bf5912f442\" (UID: \"274516e1-d34a-4045-9a79-f8bf5912f442\") " Dec 11 08:16:23 crc kubenswrapper[4860]: I1211 08:16:23.856500 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/274516e1-d34a-4045-9a79-f8bf5912f442-client-ca" (OuterVolumeSpecName: "client-ca") pod "274516e1-d34a-4045-9a79-f8bf5912f442" (UID: "274516e1-d34a-4045-9a79-f8bf5912f442"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:16:23 crc kubenswrapper[4860]: I1211 08:16:23.856629 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/274516e1-d34a-4045-9a79-f8bf5912f442-config" (OuterVolumeSpecName: "config") pod "274516e1-d34a-4045-9a79-f8bf5912f442" (UID: "274516e1-d34a-4045-9a79-f8bf5912f442"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:16:23 crc kubenswrapper[4860]: I1211 08:16:23.861084 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/274516e1-d34a-4045-9a79-f8bf5912f442-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "274516e1-d34a-4045-9a79-f8bf5912f442" (UID: "274516e1-d34a-4045-9a79-f8bf5912f442"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:16:23 crc kubenswrapper[4860]: I1211 08:16:23.861301 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/274516e1-d34a-4045-9a79-f8bf5912f442-kube-api-access-fz56g" (OuterVolumeSpecName: "kube-api-access-fz56g") pod "274516e1-d34a-4045-9a79-f8bf5912f442" (UID: "274516e1-d34a-4045-9a79-f8bf5912f442"). InnerVolumeSpecName "kube-api-access-fz56g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:16:23 crc kubenswrapper[4860]: I1211 08:16:23.956994 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7bc93e1d-1531-472b-974f-201cc58dbbf5-serving-cert\") pod \"7bc93e1d-1531-472b-974f-201cc58dbbf5\" (UID: \"7bc93e1d-1531-472b-974f-201cc58dbbf5\") " Dec 11 08:16:23 crc kubenswrapper[4860]: I1211 08:16:23.957044 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7bc93e1d-1531-472b-974f-201cc58dbbf5-client-ca\") pod \"7bc93e1d-1531-472b-974f-201cc58dbbf5\" (UID: \"7bc93e1d-1531-472b-974f-201cc58dbbf5\") " Dec 11 08:16:23 crc kubenswrapper[4860]: I1211 08:16:23.957207 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7bc93e1d-1531-472b-974f-201cc58dbbf5-config\") pod \"7bc93e1d-1531-472b-974f-201cc58dbbf5\" (UID: \"7bc93e1d-1531-472b-974f-201cc58dbbf5\") " Dec 11 08:16:23 crc kubenswrapper[4860]: I1211 08:16:23.957261 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7bc93e1d-1531-472b-974f-201cc58dbbf5-proxy-ca-bundles\") pod \"7bc93e1d-1531-472b-974f-201cc58dbbf5\" (UID: \"7bc93e1d-1531-472b-974f-201cc58dbbf5\") " Dec 11 08:16:23 crc kubenswrapper[4860]: I1211 08:16:23.957286 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zwrwx\" (UniqueName: \"kubernetes.io/projected/7bc93e1d-1531-472b-974f-201cc58dbbf5-kube-api-access-zwrwx\") pod \"7bc93e1d-1531-472b-974f-201cc58dbbf5\" (UID: \"7bc93e1d-1531-472b-974f-201cc58dbbf5\") " Dec 11 08:16:23 crc kubenswrapper[4860]: I1211 08:16:23.957463 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09f55a0a-4641-4c18-a783-e17f641aa0b0-serving-cert\") pod \"route-controller-manager-7bb5f84cb5-c5br2\" (UID: \"09f55a0a-4641-4c18-a783-e17f641aa0b0\") " pod="openshift-route-controller-manager/route-controller-manager-7bb5f84cb5-c5br2" Dec 11 08:16:23 crc kubenswrapper[4860]: I1211 08:16:23.957564 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/09f55a0a-4641-4c18-a783-e17f641aa0b0-client-ca\") pod \"route-controller-manager-7bb5f84cb5-c5br2\" (UID: \"09f55a0a-4641-4c18-a783-e17f641aa0b0\") " pod="openshift-route-controller-manager/route-controller-manager-7bb5f84cb5-c5br2" Dec 11 08:16:23 crc kubenswrapper[4860]: I1211 08:16:23.957605 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hhshd\" (UniqueName: \"kubernetes.io/projected/09f55a0a-4641-4c18-a783-e17f641aa0b0-kube-api-access-hhshd\") pod \"route-controller-manager-7bb5f84cb5-c5br2\" (UID: \"09f55a0a-4641-4c18-a783-e17f641aa0b0\") " pod="openshift-route-controller-manager/route-controller-manager-7bb5f84cb5-c5br2" Dec 11 08:16:23 crc kubenswrapper[4860]: I1211 08:16:23.957632 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09f55a0a-4641-4c18-a783-e17f641aa0b0-config\") pod \"route-controller-manager-7bb5f84cb5-c5br2\" (UID: \"09f55a0a-4641-4c18-a783-e17f641aa0b0\") " pod="openshift-route-controller-manager/route-controller-manager-7bb5f84cb5-c5br2" Dec 11 08:16:23 crc kubenswrapper[4860]: I1211 08:16:23.957684 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bc93e1d-1531-472b-974f-201cc58dbbf5-client-ca" (OuterVolumeSpecName: "client-ca") pod "7bc93e1d-1531-472b-974f-201cc58dbbf5" (UID: "7bc93e1d-1531-472b-974f-201cc58dbbf5"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:16:23 crc kubenswrapper[4860]: I1211 08:16:23.957817 4860 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/274516e1-d34a-4045-9a79-f8bf5912f442-client-ca\") on node \"crc\" DevicePath \"\"" Dec 11 08:16:23 crc kubenswrapper[4860]: I1211 08:16:23.957840 4860 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/274516e1-d34a-4045-9a79-f8bf5912f442-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 08:16:23 crc kubenswrapper[4860]: I1211 08:16:23.957854 4860 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/274516e1-d34a-4045-9a79-f8bf5912f442-config\") on node \"crc\" DevicePath \"\"" Dec 11 08:16:23 crc kubenswrapper[4860]: I1211 08:16:23.957865 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fz56g\" (UniqueName: \"kubernetes.io/projected/274516e1-d34a-4045-9a79-f8bf5912f442-kube-api-access-fz56g\") on node \"crc\" DevicePath \"\"" Dec 11 08:16:23 crc kubenswrapper[4860]: I1211 08:16:23.957919 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bc93e1d-1531-472b-974f-201cc58dbbf5-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7bc93e1d-1531-472b-974f-201cc58dbbf5" (UID: "7bc93e1d-1531-472b-974f-201cc58dbbf5"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:16:23 crc kubenswrapper[4860]: I1211 08:16:23.958365 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bc93e1d-1531-472b-974f-201cc58dbbf5-config" (OuterVolumeSpecName: "config") pod "7bc93e1d-1531-472b-974f-201cc58dbbf5" (UID: "7bc93e1d-1531-472b-974f-201cc58dbbf5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:16:23 crc kubenswrapper[4860]: I1211 08:16:23.960149 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bc93e1d-1531-472b-974f-201cc58dbbf5-kube-api-access-zwrwx" (OuterVolumeSpecName: "kube-api-access-zwrwx") pod "7bc93e1d-1531-472b-974f-201cc58dbbf5" (UID: "7bc93e1d-1531-472b-974f-201cc58dbbf5"). InnerVolumeSpecName "kube-api-access-zwrwx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:16:23 crc kubenswrapper[4860]: I1211 08:16:23.960192 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7bc93e1d-1531-472b-974f-201cc58dbbf5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7bc93e1d-1531-472b-974f-201cc58dbbf5" (UID: "7bc93e1d-1531-472b-974f-201cc58dbbf5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:16:24 crc kubenswrapper[4860]: I1211 08:16:24.059540 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09f55a0a-4641-4c18-a783-e17f641aa0b0-serving-cert\") pod \"route-controller-manager-7bb5f84cb5-c5br2\" (UID: \"09f55a0a-4641-4c18-a783-e17f641aa0b0\") " pod="openshift-route-controller-manager/route-controller-manager-7bb5f84cb5-c5br2" Dec 11 08:16:24 crc kubenswrapper[4860]: I1211 08:16:24.059622 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/09f55a0a-4641-4c18-a783-e17f641aa0b0-client-ca\") pod \"route-controller-manager-7bb5f84cb5-c5br2\" (UID: \"09f55a0a-4641-4c18-a783-e17f641aa0b0\") " pod="openshift-route-controller-manager/route-controller-manager-7bb5f84cb5-c5br2" Dec 11 08:16:24 crc kubenswrapper[4860]: I1211 08:16:24.059665 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hhshd\" (UniqueName: \"kubernetes.io/projected/09f55a0a-4641-4c18-a783-e17f641aa0b0-kube-api-access-hhshd\") pod \"route-controller-manager-7bb5f84cb5-c5br2\" (UID: \"09f55a0a-4641-4c18-a783-e17f641aa0b0\") " pod="openshift-route-controller-manager/route-controller-manager-7bb5f84cb5-c5br2" Dec 11 08:16:24 crc kubenswrapper[4860]: I1211 08:16:24.059698 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09f55a0a-4641-4c18-a783-e17f641aa0b0-config\") pod \"route-controller-manager-7bb5f84cb5-c5br2\" (UID: \"09f55a0a-4641-4c18-a783-e17f641aa0b0\") " pod="openshift-route-controller-manager/route-controller-manager-7bb5f84cb5-c5br2" Dec 11 08:16:24 crc kubenswrapper[4860]: I1211 08:16:24.059744 4860 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7bc93e1d-1531-472b-974f-201cc58dbbf5-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 11 08:16:24 crc kubenswrapper[4860]: I1211 08:16:24.059755 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zwrwx\" (UniqueName: \"kubernetes.io/projected/7bc93e1d-1531-472b-974f-201cc58dbbf5-kube-api-access-zwrwx\") on node \"crc\" DevicePath \"\"" Dec 11 08:16:24 crc kubenswrapper[4860]: I1211 08:16:24.059767 4860 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7bc93e1d-1531-472b-974f-201cc58dbbf5-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 08:16:24 crc kubenswrapper[4860]: I1211 08:16:24.059775 4860 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7bc93e1d-1531-472b-974f-201cc58dbbf5-client-ca\") on node \"crc\" DevicePath \"\"" Dec 11 08:16:24 crc kubenswrapper[4860]: I1211 08:16:24.059787 4860 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7bc93e1d-1531-472b-974f-201cc58dbbf5-config\") on node \"crc\" DevicePath \"\"" Dec 11 08:16:24 crc kubenswrapper[4860]: I1211 08:16:24.060678 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/09f55a0a-4641-4c18-a783-e17f641aa0b0-client-ca\") pod \"route-controller-manager-7bb5f84cb5-c5br2\" (UID: \"09f55a0a-4641-4c18-a783-e17f641aa0b0\") " pod="openshift-route-controller-manager/route-controller-manager-7bb5f84cb5-c5br2" Dec 11 08:16:24 crc kubenswrapper[4860]: I1211 08:16:24.060930 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09f55a0a-4641-4c18-a783-e17f641aa0b0-config\") pod \"route-controller-manager-7bb5f84cb5-c5br2\" (UID: \"09f55a0a-4641-4c18-a783-e17f641aa0b0\") " pod="openshift-route-controller-manager/route-controller-manager-7bb5f84cb5-c5br2" Dec 11 08:16:24 crc kubenswrapper[4860]: I1211 08:16:24.064942 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09f55a0a-4641-4c18-a783-e17f641aa0b0-serving-cert\") pod \"route-controller-manager-7bb5f84cb5-c5br2\" (UID: \"09f55a0a-4641-4c18-a783-e17f641aa0b0\") " pod="openshift-route-controller-manager/route-controller-manager-7bb5f84cb5-c5br2" Dec 11 08:16:24 crc kubenswrapper[4860]: I1211 08:16:24.081950 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hhshd\" (UniqueName: \"kubernetes.io/projected/09f55a0a-4641-4c18-a783-e17f641aa0b0-kube-api-access-hhshd\") pod \"route-controller-manager-7bb5f84cb5-c5br2\" (UID: \"09f55a0a-4641-4c18-a783-e17f641aa0b0\") " pod="openshift-route-controller-manager/route-controller-manager-7bb5f84cb5-c5br2" Dec 11 08:16:24 crc kubenswrapper[4860]: I1211 08:16:24.120961 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7bb5f84cb5-c5br2" Dec 11 08:16:24 crc kubenswrapper[4860]: I1211 08:16:24.289810 4860 generic.go:334] "Generic (PLEG): container finished" podID="274516e1-d34a-4045-9a79-f8bf5912f442" containerID="3744356c8afe6a48bcd3e42f6e7079ecf96beb792792cfda8374bd7fdfbc5965" exitCode=0 Dec 11 08:16:24 crc kubenswrapper[4860]: I1211 08:16:24.289932 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7688f75b54-fvbqm" event={"ID":"274516e1-d34a-4045-9a79-f8bf5912f442","Type":"ContainerDied","Data":"3744356c8afe6a48bcd3e42f6e7079ecf96beb792792cfda8374bd7fdfbc5965"} Dec 11 08:16:24 crc kubenswrapper[4860]: I1211 08:16:24.289942 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7688f75b54-fvbqm" Dec 11 08:16:24 crc kubenswrapper[4860]: I1211 08:16:24.289963 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7688f75b54-fvbqm" event={"ID":"274516e1-d34a-4045-9a79-f8bf5912f442","Type":"ContainerDied","Data":"ffc739e0b47303a8ea78cb11d38dd9d86a80a6297178e0f56a694476ffbc1c19"} Dec 11 08:16:24 crc kubenswrapper[4860]: I1211 08:16:24.289983 4860 scope.go:117] "RemoveContainer" containerID="3744356c8afe6a48bcd3e42f6e7079ecf96beb792792cfda8374bd7fdfbc5965" Dec 11 08:16:24 crc kubenswrapper[4860]: I1211 08:16:24.292366 4860 generic.go:334] "Generic (PLEG): container finished" podID="7bc93e1d-1531-472b-974f-201cc58dbbf5" containerID="7a02a44278e305b7f34d2bd50034c6045f87bd4c19ef7373a9e8559c9892630c" exitCode=0 Dec 11 08:16:24 crc kubenswrapper[4860]: I1211 08:16:24.292406 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5c997b8cc9-q7sqj" event={"ID":"7bc93e1d-1531-472b-974f-201cc58dbbf5","Type":"ContainerDied","Data":"7a02a44278e305b7f34d2bd50034c6045f87bd4c19ef7373a9e8559c9892630c"} Dec 11 08:16:24 crc kubenswrapper[4860]: I1211 08:16:24.292437 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5c997b8cc9-q7sqj" event={"ID":"7bc93e1d-1531-472b-974f-201cc58dbbf5","Type":"ContainerDied","Data":"b0041fae7923eb1fc7917a55d48525bbc8250d7df6582b04d2cabfbed35326d7"} Dec 11 08:16:24 crc kubenswrapper[4860]: I1211 08:16:24.292488 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5c997b8cc9-q7sqj" Dec 11 08:16:24 crc kubenswrapper[4860]: I1211 08:16:24.329878 4860 scope.go:117] "RemoveContainer" containerID="3744356c8afe6a48bcd3e42f6e7079ecf96beb792792cfda8374bd7fdfbc5965" Dec 11 08:16:24 crc kubenswrapper[4860]: I1211 08:16:24.330019 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7688f75b54-fvbqm"] Dec 11 08:16:24 crc kubenswrapper[4860]: E1211 08:16:24.330385 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3744356c8afe6a48bcd3e42f6e7079ecf96beb792792cfda8374bd7fdfbc5965\": container with ID starting with 3744356c8afe6a48bcd3e42f6e7079ecf96beb792792cfda8374bd7fdfbc5965 not found: ID does not exist" containerID="3744356c8afe6a48bcd3e42f6e7079ecf96beb792792cfda8374bd7fdfbc5965" Dec 11 08:16:24 crc kubenswrapper[4860]: I1211 08:16:24.330409 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3744356c8afe6a48bcd3e42f6e7079ecf96beb792792cfda8374bd7fdfbc5965"} err="failed to get container status \"3744356c8afe6a48bcd3e42f6e7079ecf96beb792792cfda8374bd7fdfbc5965\": rpc error: code = NotFound desc = could not find container \"3744356c8afe6a48bcd3e42f6e7079ecf96beb792792cfda8374bd7fdfbc5965\": container with ID starting with 3744356c8afe6a48bcd3e42f6e7079ecf96beb792792cfda8374bd7fdfbc5965 not found: ID does not exist" Dec 11 08:16:24 crc kubenswrapper[4860]: I1211 08:16:24.330430 4860 scope.go:117] "RemoveContainer" containerID="7a02a44278e305b7f34d2bd50034c6045f87bd4c19ef7373a9e8559c9892630c" Dec 11 08:16:24 crc kubenswrapper[4860]: I1211 08:16:24.335267 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7688f75b54-fvbqm"] Dec 11 08:16:24 crc kubenswrapper[4860]: I1211 08:16:24.343236 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-5c997b8cc9-q7sqj"] Dec 11 08:16:24 crc kubenswrapper[4860]: I1211 08:16:24.347474 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-5c997b8cc9-q7sqj"] Dec 11 08:16:24 crc kubenswrapper[4860]: I1211 08:16:24.350601 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7bb5f84cb5-c5br2"] Dec 11 08:16:24 crc kubenswrapper[4860]: I1211 08:16:24.353087 4860 scope.go:117] "RemoveContainer" containerID="7a02a44278e305b7f34d2bd50034c6045f87bd4c19ef7373a9e8559c9892630c" Dec 11 08:16:24 crc kubenswrapper[4860]: E1211 08:16:24.353512 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7a02a44278e305b7f34d2bd50034c6045f87bd4c19ef7373a9e8559c9892630c\": container with ID starting with 7a02a44278e305b7f34d2bd50034c6045f87bd4c19ef7373a9e8559c9892630c not found: ID does not exist" containerID="7a02a44278e305b7f34d2bd50034c6045f87bd4c19ef7373a9e8559c9892630c" Dec 11 08:16:24 crc kubenswrapper[4860]: I1211 08:16:24.353554 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a02a44278e305b7f34d2bd50034c6045f87bd4c19ef7373a9e8559c9892630c"} err="failed to get container status \"7a02a44278e305b7f34d2bd50034c6045f87bd4c19ef7373a9e8559c9892630c\": rpc error: code = NotFound desc = could not find container \"7a02a44278e305b7f34d2bd50034c6045f87bd4c19ef7373a9e8559c9892630c\": container with ID starting with 7a02a44278e305b7f34d2bd50034c6045f87bd4c19ef7373a9e8559c9892630c not found: ID does not exist" Dec 11 08:16:24 crc kubenswrapper[4860]: W1211 08:16:24.356288 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod09f55a0a_4641_4c18_a783_e17f641aa0b0.slice/crio-bcc1371dfec677f832a151fb5508a67ca07260ecb82477b3d4113e1ada5ac232 WatchSource:0}: Error finding container bcc1371dfec677f832a151fb5508a67ca07260ecb82477b3d4113e1ada5ac232: Status 404 returned error can't find the container with id bcc1371dfec677f832a151fb5508a67ca07260ecb82477b3d4113e1ada5ac232 Dec 11 08:16:25 crc kubenswrapper[4860]: I1211 08:16:25.302420 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7bb5f84cb5-c5br2" event={"ID":"09f55a0a-4641-4c18-a783-e17f641aa0b0","Type":"ContainerStarted","Data":"3d45b89afaf85a0f687957be1a96fe00b702a679586fb5e7c5c08a7c4d350c30"} Dec 11 08:16:25 crc kubenswrapper[4860]: I1211 08:16:25.302714 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7bb5f84cb5-c5br2" event={"ID":"09f55a0a-4641-4c18-a783-e17f641aa0b0","Type":"ContainerStarted","Data":"bcc1371dfec677f832a151fb5508a67ca07260ecb82477b3d4113e1ada5ac232"} Dec 11 08:16:25 crc kubenswrapper[4860]: I1211 08:16:25.302851 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-7bb5f84cb5-c5br2" Dec 11 08:16:25 crc kubenswrapper[4860]: I1211 08:16:25.311396 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-7bb5f84cb5-c5br2" Dec 11 08:16:25 crc kubenswrapper[4860]: I1211 08:16:25.324423 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-7bb5f84cb5-c5br2" podStartSLOduration=4.324400642 podStartE2EDuration="4.324400642s" podCreationTimestamp="2025-12-11 08:16:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:16:25.322672362 +0000 UTC m=+318.051191417" watchObservedRunningTime="2025-12-11 08:16:25.324400642 +0000 UTC m=+318.052919707" Dec 11 08:16:25 crc kubenswrapper[4860]: I1211 08:16:25.590637 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="274516e1-d34a-4045-9a79-f8bf5912f442" path="/var/lib/kubelet/pods/274516e1-d34a-4045-9a79-f8bf5912f442/volumes" Dec 11 08:16:25 crc kubenswrapper[4860]: I1211 08:16:25.591839 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bc93e1d-1531-472b-974f-201cc58dbbf5" path="/var/lib/kubelet/pods/7bc93e1d-1531-472b-974f-201cc58dbbf5/volumes" Dec 11 08:16:26 crc kubenswrapper[4860]: I1211 08:16:26.183074 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7bb5f84cb5-c5br2"] Dec 11 08:16:26 crc kubenswrapper[4860]: I1211 08:16:26.641797 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-f4b45fc4f-ttzzw"] Dec 11 08:16:26 crc kubenswrapper[4860]: I1211 08:16:26.642958 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-f4b45fc4f-ttzzw" Dec 11 08:16:26 crc kubenswrapper[4860]: I1211 08:16:26.647046 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 11 08:16:26 crc kubenswrapper[4860]: I1211 08:16:26.647453 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 11 08:16:26 crc kubenswrapper[4860]: I1211 08:16:26.650552 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 11 08:16:26 crc kubenswrapper[4860]: I1211 08:16:26.650559 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 11 08:16:26 crc kubenswrapper[4860]: I1211 08:16:26.650688 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 11 08:16:26 crc kubenswrapper[4860]: I1211 08:16:26.651556 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 11 08:16:26 crc kubenswrapper[4860]: I1211 08:16:26.659750 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-f4b45fc4f-ttzzw"] Dec 11 08:16:26 crc kubenswrapper[4860]: I1211 08:16:26.673120 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 11 08:16:26 crc kubenswrapper[4860]: I1211 08:16:26.803501 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7de45232-1210-4276-b62e-b7693c2c4360-proxy-ca-bundles\") pod \"controller-manager-f4b45fc4f-ttzzw\" (UID: \"7de45232-1210-4276-b62e-b7693c2c4360\") " pod="openshift-controller-manager/controller-manager-f4b45fc4f-ttzzw" Dec 11 08:16:26 crc kubenswrapper[4860]: I1211 08:16:26.803632 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7de45232-1210-4276-b62e-b7693c2c4360-config\") pod \"controller-manager-f4b45fc4f-ttzzw\" (UID: \"7de45232-1210-4276-b62e-b7693c2c4360\") " pod="openshift-controller-manager/controller-manager-f4b45fc4f-ttzzw" Dec 11 08:16:26 crc kubenswrapper[4860]: I1211 08:16:26.803767 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7de45232-1210-4276-b62e-b7693c2c4360-serving-cert\") pod \"controller-manager-f4b45fc4f-ttzzw\" (UID: \"7de45232-1210-4276-b62e-b7693c2c4360\") " pod="openshift-controller-manager/controller-manager-f4b45fc4f-ttzzw" Dec 11 08:16:26 crc kubenswrapper[4860]: I1211 08:16:26.803851 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2vwmj\" (UniqueName: \"kubernetes.io/projected/7de45232-1210-4276-b62e-b7693c2c4360-kube-api-access-2vwmj\") pod \"controller-manager-f4b45fc4f-ttzzw\" (UID: \"7de45232-1210-4276-b62e-b7693c2c4360\") " pod="openshift-controller-manager/controller-manager-f4b45fc4f-ttzzw" Dec 11 08:16:26 crc kubenswrapper[4860]: I1211 08:16:26.803932 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7de45232-1210-4276-b62e-b7693c2c4360-client-ca\") pod \"controller-manager-f4b45fc4f-ttzzw\" (UID: \"7de45232-1210-4276-b62e-b7693c2c4360\") " pod="openshift-controller-manager/controller-manager-f4b45fc4f-ttzzw" Dec 11 08:16:26 crc kubenswrapper[4860]: I1211 08:16:26.904920 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7de45232-1210-4276-b62e-b7693c2c4360-proxy-ca-bundles\") pod \"controller-manager-f4b45fc4f-ttzzw\" (UID: \"7de45232-1210-4276-b62e-b7693c2c4360\") " pod="openshift-controller-manager/controller-manager-f4b45fc4f-ttzzw" Dec 11 08:16:26 crc kubenswrapper[4860]: I1211 08:16:26.905009 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7de45232-1210-4276-b62e-b7693c2c4360-config\") pod \"controller-manager-f4b45fc4f-ttzzw\" (UID: \"7de45232-1210-4276-b62e-b7693c2c4360\") " pod="openshift-controller-manager/controller-manager-f4b45fc4f-ttzzw" Dec 11 08:16:26 crc kubenswrapper[4860]: I1211 08:16:26.905049 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7de45232-1210-4276-b62e-b7693c2c4360-serving-cert\") pod \"controller-manager-f4b45fc4f-ttzzw\" (UID: \"7de45232-1210-4276-b62e-b7693c2c4360\") " pod="openshift-controller-manager/controller-manager-f4b45fc4f-ttzzw" Dec 11 08:16:26 crc kubenswrapper[4860]: I1211 08:16:26.905079 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2vwmj\" (UniqueName: \"kubernetes.io/projected/7de45232-1210-4276-b62e-b7693c2c4360-kube-api-access-2vwmj\") pod \"controller-manager-f4b45fc4f-ttzzw\" (UID: \"7de45232-1210-4276-b62e-b7693c2c4360\") " pod="openshift-controller-manager/controller-manager-f4b45fc4f-ttzzw" Dec 11 08:16:26 crc kubenswrapper[4860]: I1211 08:16:26.905118 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7de45232-1210-4276-b62e-b7693c2c4360-client-ca\") pod \"controller-manager-f4b45fc4f-ttzzw\" (UID: \"7de45232-1210-4276-b62e-b7693c2c4360\") " pod="openshift-controller-manager/controller-manager-f4b45fc4f-ttzzw" Dec 11 08:16:26 crc kubenswrapper[4860]: I1211 08:16:26.906234 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7de45232-1210-4276-b62e-b7693c2c4360-client-ca\") pod \"controller-manager-f4b45fc4f-ttzzw\" (UID: \"7de45232-1210-4276-b62e-b7693c2c4360\") " pod="openshift-controller-manager/controller-manager-f4b45fc4f-ttzzw" Dec 11 08:16:26 crc kubenswrapper[4860]: I1211 08:16:26.907528 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7de45232-1210-4276-b62e-b7693c2c4360-proxy-ca-bundles\") pod \"controller-manager-f4b45fc4f-ttzzw\" (UID: \"7de45232-1210-4276-b62e-b7693c2c4360\") " pod="openshift-controller-manager/controller-manager-f4b45fc4f-ttzzw" Dec 11 08:16:26 crc kubenswrapper[4860]: I1211 08:16:26.907932 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7de45232-1210-4276-b62e-b7693c2c4360-config\") pod \"controller-manager-f4b45fc4f-ttzzw\" (UID: \"7de45232-1210-4276-b62e-b7693c2c4360\") " pod="openshift-controller-manager/controller-manager-f4b45fc4f-ttzzw" Dec 11 08:16:26 crc kubenswrapper[4860]: I1211 08:16:26.917050 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7de45232-1210-4276-b62e-b7693c2c4360-serving-cert\") pod \"controller-manager-f4b45fc4f-ttzzw\" (UID: \"7de45232-1210-4276-b62e-b7693c2c4360\") " pod="openshift-controller-manager/controller-manager-f4b45fc4f-ttzzw" Dec 11 08:16:26 crc kubenswrapper[4860]: I1211 08:16:26.932594 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2vwmj\" (UniqueName: \"kubernetes.io/projected/7de45232-1210-4276-b62e-b7693c2c4360-kube-api-access-2vwmj\") pod \"controller-manager-f4b45fc4f-ttzzw\" (UID: \"7de45232-1210-4276-b62e-b7693c2c4360\") " pod="openshift-controller-manager/controller-manager-f4b45fc4f-ttzzw" Dec 11 08:16:26 crc kubenswrapper[4860]: I1211 08:16:26.970473 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-f4b45fc4f-ttzzw" Dec 11 08:16:27 crc kubenswrapper[4860]: I1211 08:16:27.174769 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-f4b45fc4f-ttzzw"] Dec 11 08:16:27 crc kubenswrapper[4860]: I1211 08:16:27.318553 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-f4b45fc4f-ttzzw" event={"ID":"7de45232-1210-4276-b62e-b7693c2c4360","Type":"ContainerStarted","Data":"71e3b9cd2f8fa1202446b0a9896e2268ff21c20961fbceadad5f3ec40cf9b4e2"} Dec 11 08:16:27 crc kubenswrapper[4860]: I1211 08:16:27.319004 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-f4b45fc4f-ttzzw" event={"ID":"7de45232-1210-4276-b62e-b7693c2c4360","Type":"ContainerStarted","Data":"4fbcfc8702bc5af8426ca7c2f2049993c40a49edb8792d984f1486ac2f62fe8a"} Dec 11 08:16:27 crc kubenswrapper[4860]: I1211 08:16:27.318801 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-7bb5f84cb5-c5br2" podUID="09f55a0a-4641-4c18-a783-e17f641aa0b0" containerName="route-controller-manager" containerID="cri-o://3d45b89afaf85a0f687957be1a96fe00b702a679586fb5e7c5c08a7c4d350c30" gracePeriod=30 Dec 11 08:16:27 crc kubenswrapper[4860]: I1211 08:16:27.339322 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-f4b45fc4f-ttzzw" podStartSLOduration=6.339295963 podStartE2EDuration="6.339295963s" podCreationTimestamp="2025-12-11 08:16:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:16:27.333939527 +0000 UTC m=+320.062458612" watchObservedRunningTime="2025-12-11 08:16:27.339295963 +0000 UTC m=+320.067815018" Dec 11 08:16:27 crc kubenswrapper[4860]: I1211 08:16:27.676778 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7bb5f84cb5-c5br2" Dec 11 08:16:27 crc kubenswrapper[4860]: I1211 08:16:27.698588 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6b5d4bcb95-jqnl2"] Dec 11 08:16:27 crc kubenswrapper[4860]: E1211 08:16:27.698851 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09f55a0a-4641-4c18-a783-e17f641aa0b0" containerName="route-controller-manager" Dec 11 08:16:27 crc kubenswrapper[4860]: I1211 08:16:27.698870 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="09f55a0a-4641-4c18-a783-e17f641aa0b0" containerName="route-controller-manager" Dec 11 08:16:27 crc kubenswrapper[4860]: I1211 08:16:27.698979 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="09f55a0a-4641-4c18-a783-e17f641aa0b0" containerName="route-controller-manager" Dec 11 08:16:27 crc kubenswrapper[4860]: I1211 08:16:27.699406 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6b5d4bcb95-jqnl2" Dec 11 08:16:27 crc kubenswrapper[4860]: I1211 08:16:27.718876 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jhbsf\" (UniqueName: \"kubernetes.io/projected/477e8461-916b-4a89-9b1b-0414cbe1d4d6-kube-api-access-jhbsf\") pod \"route-controller-manager-6b5d4bcb95-jqnl2\" (UID: \"477e8461-916b-4a89-9b1b-0414cbe1d4d6\") " pod="openshift-route-controller-manager/route-controller-manager-6b5d4bcb95-jqnl2" Dec 11 08:16:27 crc kubenswrapper[4860]: I1211 08:16:27.718957 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/477e8461-916b-4a89-9b1b-0414cbe1d4d6-client-ca\") pod \"route-controller-manager-6b5d4bcb95-jqnl2\" (UID: \"477e8461-916b-4a89-9b1b-0414cbe1d4d6\") " pod="openshift-route-controller-manager/route-controller-manager-6b5d4bcb95-jqnl2" Dec 11 08:16:27 crc kubenswrapper[4860]: I1211 08:16:27.719068 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/477e8461-916b-4a89-9b1b-0414cbe1d4d6-serving-cert\") pod \"route-controller-manager-6b5d4bcb95-jqnl2\" (UID: \"477e8461-916b-4a89-9b1b-0414cbe1d4d6\") " pod="openshift-route-controller-manager/route-controller-manager-6b5d4bcb95-jqnl2" Dec 11 08:16:27 crc kubenswrapper[4860]: I1211 08:16:27.719111 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/477e8461-916b-4a89-9b1b-0414cbe1d4d6-config\") pod \"route-controller-manager-6b5d4bcb95-jqnl2\" (UID: \"477e8461-916b-4a89-9b1b-0414cbe1d4d6\") " pod="openshift-route-controller-manager/route-controller-manager-6b5d4bcb95-jqnl2" Dec 11 08:16:27 crc kubenswrapper[4860]: I1211 08:16:27.743825 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6b5d4bcb95-jqnl2"] Dec 11 08:16:27 crc kubenswrapper[4860]: I1211 08:16:27.819689 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/09f55a0a-4641-4c18-a783-e17f641aa0b0-client-ca\") pod \"09f55a0a-4641-4c18-a783-e17f641aa0b0\" (UID: \"09f55a0a-4641-4c18-a783-e17f641aa0b0\") " Dec 11 08:16:27 crc kubenswrapper[4860]: I1211 08:16:27.819759 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09f55a0a-4641-4c18-a783-e17f641aa0b0-config\") pod \"09f55a0a-4641-4c18-a783-e17f641aa0b0\" (UID: \"09f55a0a-4641-4c18-a783-e17f641aa0b0\") " Dec 11 08:16:27 crc kubenswrapper[4860]: I1211 08:16:27.819787 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hhshd\" (UniqueName: \"kubernetes.io/projected/09f55a0a-4641-4c18-a783-e17f641aa0b0-kube-api-access-hhshd\") pod \"09f55a0a-4641-4c18-a783-e17f641aa0b0\" (UID: \"09f55a0a-4641-4c18-a783-e17f641aa0b0\") " Dec 11 08:16:27 crc kubenswrapper[4860]: I1211 08:16:27.819811 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09f55a0a-4641-4c18-a783-e17f641aa0b0-serving-cert\") pod \"09f55a0a-4641-4c18-a783-e17f641aa0b0\" (UID: \"09f55a0a-4641-4c18-a783-e17f641aa0b0\") " Dec 11 08:16:27 crc kubenswrapper[4860]: I1211 08:16:27.819897 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jhbsf\" (UniqueName: \"kubernetes.io/projected/477e8461-916b-4a89-9b1b-0414cbe1d4d6-kube-api-access-jhbsf\") pod \"route-controller-manager-6b5d4bcb95-jqnl2\" (UID: \"477e8461-916b-4a89-9b1b-0414cbe1d4d6\") " pod="openshift-route-controller-manager/route-controller-manager-6b5d4bcb95-jqnl2" Dec 11 08:16:27 crc kubenswrapper[4860]: I1211 08:16:27.819924 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/477e8461-916b-4a89-9b1b-0414cbe1d4d6-client-ca\") pod \"route-controller-manager-6b5d4bcb95-jqnl2\" (UID: \"477e8461-916b-4a89-9b1b-0414cbe1d4d6\") " pod="openshift-route-controller-manager/route-controller-manager-6b5d4bcb95-jqnl2" Dec 11 08:16:27 crc kubenswrapper[4860]: I1211 08:16:27.819952 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/477e8461-916b-4a89-9b1b-0414cbe1d4d6-serving-cert\") pod \"route-controller-manager-6b5d4bcb95-jqnl2\" (UID: \"477e8461-916b-4a89-9b1b-0414cbe1d4d6\") " pod="openshift-route-controller-manager/route-controller-manager-6b5d4bcb95-jqnl2" Dec 11 08:16:27 crc kubenswrapper[4860]: I1211 08:16:27.819976 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/477e8461-916b-4a89-9b1b-0414cbe1d4d6-config\") pod \"route-controller-manager-6b5d4bcb95-jqnl2\" (UID: \"477e8461-916b-4a89-9b1b-0414cbe1d4d6\") " pod="openshift-route-controller-manager/route-controller-manager-6b5d4bcb95-jqnl2" Dec 11 08:16:27 crc kubenswrapper[4860]: I1211 08:16:27.821238 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09f55a0a-4641-4c18-a783-e17f641aa0b0-client-ca" (OuterVolumeSpecName: "client-ca") pod "09f55a0a-4641-4c18-a783-e17f641aa0b0" (UID: "09f55a0a-4641-4c18-a783-e17f641aa0b0"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:16:27 crc kubenswrapper[4860]: I1211 08:16:27.821757 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/477e8461-916b-4a89-9b1b-0414cbe1d4d6-client-ca\") pod \"route-controller-manager-6b5d4bcb95-jqnl2\" (UID: \"477e8461-916b-4a89-9b1b-0414cbe1d4d6\") " pod="openshift-route-controller-manager/route-controller-manager-6b5d4bcb95-jqnl2" Dec 11 08:16:27 crc kubenswrapper[4860]: I1211 08:16:27.821826 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09f55a0a-4641-4c18-a783-e17f641aa0b0-config" (OuterVolumeSpecName: "config") pod "09f55a0a-4641-4c18-a783-e17f641aa0b0" (UID: "09f55a0a-4641-4c18-a783-e17f641aa0b0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:16:27 crc kubenswrapper[4860]: I1211 08:16:27.823539 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/477e8461-916b-4a89-9b1b-0414cbe1d4d6-config\") pod \"route-controller-manager-6b5d4bcb95-jqnl2\" (UID: \"477e8461-916b-4a89-9b1b-0414cbe1d4d6\") " pod="openshift-route-controller-manager/route-controller-manager-6b5d4bcb95-jqnl2" Dec 11 08:16:27 crc kubenswrapper[4860]: I1211 08:16:27.825090 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09f55a0a-4641-4c18-a783-e17f641aa0b0-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09f55a0a-4641-4c18-a783-e17f641aa0b0" (UID: "09f55a0a-4641-4c18-a783-e17f641aa0b0"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:16:27 crc kubenswrapper[4860]: I1211 08:16:27.825225 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09f55a0a-4641-4c18-a783-e17f641aa0b0-kube-api-access-hhshd" (OuterVolumeSpecName: "kube-api-access-hhshd") pod "09f55a0a-4641-4c18-a783-e17f641aa0b0" (UID: "09f55a0a-4641-4c18-a783-e17f641aa0b0"). InnerVolumeSpecName "kube-api-access-hhshd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:16:27 crc kubenswrapper[4860]: I1211 08:16:27.825664 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/477e8461-916b-4a89-9b1b-0414cbe1d4d6-serving-cert\") pod \"route-controller-manager-6b5d4bcb95-jqnl2\" (UID: \"477e8461-916b-4a89-9b1b-0414cbe1d4d6\") " pod="openshift-route-controller-manager/route-controller-manager-6b5d4bcb95-jqnl2" Dec 11 08:16:27 crc kubenswrapper[4860]: I1211 08:16:27.839142 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jhbsf\" (UniqueName: \"kubernetes.io/projected/477e8461-916b-4a89-9b1b-0414cbe1d4d6-kube-api-access-jhbsf\") pod \"route-controller-manager-6b5d4bcb95-jqnl2\" (UID: \"477e8461-916b-4a89-9b1b-0414cbe1d4d6\") " pod="openshift-route-controller-manager/route-controller-manager-6b5d4bcb95-jqnl2" Dec 11 08:16:27 crc kubenswrapper[4860]: I1211 08:16:27.921747 4860 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/09f55a0a-4641-4c18-a783-e17f641aa0b0-client-ca\") on node \"crc\" DevicePath \"\"" Dec 11 08:16:27 crc kubenswrapper[4860]: I1211 08:16:27.922147 4860 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09f55a0a-4641-4c18-a783-e17f641aa0b0-config\") on node \"crc\" DevicePath \"\"" Dec 11 08:16:27 crc kubenswrapper[4860]: I1211 08:16:27.922210 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hhshd\" (UniqueName: \"kubernetes.io/projected/09f55a0a-4641-4c18-a783-e17f641aa0b0-kube-api-access-hhshd\") on node \"crc\" DevicePath \"\"" Dec 11 08:16:27 crc kubenswrapper[4860]: I1211 08:16:27.922332 4860 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09f55a0a-4641-4c18-a783-e17f641aa0b0-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 08:16:28 crc kubenswrapper[4860]: I1211 08:16:28.035341 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6b5d4bcb95-jqnl2" Dec 11 08:16:28 crc kubenswrapper[4860]: I1211 08:16:28.263610 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6b5d4bcb95-jqnl2"] Dec 11 08:16:28 crc kubenswrapper[4860]: I1211 08:16:28.327322 4860 generic.go:334] "Generic (PLEG): container finished" podID="09f55a0a-4641-4c18-a783-e17f641aa0b0" containerID="3d45b89afaf85a0f687957be1a96fe00b702a679586fb5e7c5c08a7c4d350c30" exitCode=0 Dec 11 08:16:28 crc kubenswrapper[4860]: I1211 08:16:28.327528 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7bb5f84cb5-c5br2" Dec 11 08:16:28 crc kubenswrapper[4860]: I1211 08:16:28.327875 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7bb5f84cb5-c5br2" event={"ID":"09f55a0a-4641-4c18-a783-e17f641aa0b0","Type":"ContainerDied","Data":"3d45b89afaf85a0f687957be1a96fe00b702a679586fb5e7c5c08a7c4d350c30"} Dec 11 08:16:28 crc kubenswrapper[4860]: I1211 08:16:28.327931 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7bb5f84cb5-c5br2" event={"ID":"09f55a0a-4641-4c18-a783-e17f641aa0b0","Type":"ContainerDied","Data":"bcc1371dfec677f832a151fb5508a67ca07260ecb82477b3d4113e1ada5ac232"} Dec 11 08:16:28 crc kubenswrapper[4860]: I1211 08:16:28.327953 4860 scope.go:117] "RemoveContainer" containerID="3d45b89afaf85a0f687957be1a96fe00b702a679586fb5e7c5c08a7c4d350c30" Dec 11 08:16:28 crc kubenswrapper[4860]: I1211 08:16:28.330630 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6b5d4bcb95-jqnl2" event={"ID":"477e8461-916b-4a89-9b1b-0414cbe1d4d6","Type":"ContainerStarted","Data":"e312fc2b7208920a37da4708212e91ebf208a741fc1b5f6154fa0c57496457e1"} Dec 11 08:16:28 crc kubenswrapper[4860]: I1211 08:16:28.330998 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-f4b45fc4f-ttzzw" Dec 11 08:16:28 crc kubenswrapper[4860]: I1211 08:16:28.335788 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-f4b45fc4f-ttzzw" Dec 11 08:16:28 crc kubenswrapper[4860]: I1211 08:16:28.357280 4860 scope.go:117] "RemoveContainer" containerID="3d45b89afaf85a0f687957be1a96fe00b702a679586fb5e7c5c08a7c4d350c30" Dec 11 08:16:28 crc kubenswrapper[4860]: E1211 08:16:28.357913 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3d45b89afaf85a0f687957be1a96fe00b702a679586fb5e7c5c08a7c4d350c30\": container with ID starting with 3d45b89afaf85a0f687957be1a96fe00b702a679586fb5e7c5c08a7c4d350c30 not found: ID does not exist" containerID="3d45b89afaf85a0f687957be1a96fe00b702a679586fb5e7c5c08a7c4d350c30" Dec 11 08:16:28 crc kubenswrapper[4860]: I1211 08:16:28.358056 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d45b89afaf85a0f687957be1a96fe00b702a679586fb5e7c5c08a7c4d350c30"} err="failed to get container status \"3d45b89afaf85a0f687957be1a96fe00b702a679586fb5e7c5c08a7c4d350c30\": rpc error: code = NotFound desc = could not find container \"3d45b89afaf85a0f687957be1a96fe00b702a679586fb5e7c5c08a7c4d350c30\": container with ID starting with 3d45b89afaf85a0f687957be1a96fe00b702a679586fb5e7c5c08a7c4d350c30 not found: ID does not exist" Dec 11 08:16:28 crc kubenswrapper[4860]: I1211 08:16:28.396754 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7bb5f84cb5-c5br2"] Dec 11 08:16:28 crc kubenswrapper[4860]: I1211 08:16:28.409155 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7bb5f84cb5-c5br2"] Dec 11 08:16:29 crc kubenswrapper[4860]: I1211 08:16:29.338737 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6b5d4bcb95-jqnl2" event={"ID":"477e8461-916b-4a89-9b1b-0414cbe1d4d6","Type":"ContainerStarted","Data":"219ba1a832ab31c229f0a9cd07f03dec5403ecc0d385284499e9f73c2daf92c1"} Dec 11 08:16:29 crc kubenswrapper[4860]: I1211 08:16:29.359713 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6b5d4bcb95-jqnl2" podStartSLOduration=3.359692434 podStartE2EDuration="3.359692434s" podCreationTimestamp="2025-12-11 08:16:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:16:29.356513533 +0000 UTC m=+322.085032618" watchObservedRunningTime="2025-12-11 08:16:29.359692434 +0000 UTC m=+322.088211499" Dec 11 08:16:29 crc kubenswrapper[4860]: I1211 08:16:29.585238 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09f55a0a-4641-4c18-a783-e17f641aa0b0" path="/var/lib/kubelet/pods/09f55a0a-4641-4c18-a783-e17f641aa0b0/volumes" Dec 11 08:16:29 crc kubenswrapper[4860]: I1211 08:16:29.723392 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-dz7kr"] Dec 11 08:16:30 crc kubenswrapper[4860]: I1211 08:16:30.344684 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6b5d4bcb95-jqnl2" Dec 11 08:16:30 crc kubenswrapper[4860]: I1211 08:16:30.351973 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6b5d4bcb95-jqnl2" Dec 11 08:16:37 crc kubenswrapper[4860]: I1211 08:16:37.596353 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-kzkbs"] Dec 11 08:16:37 crc kubenswrapper[4860]: I1211 08:16:37.597978 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kzkbs" Dec 11 08:16:37 crc kubenswrapper[4860]: I1211 08:16:37.600594 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 11 08:16:37 crc kubenswrapper[4860]: I1211 08:16:37.606893 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-kzkbs"] Dec 11 08:16:37 crc kubenswrapper[4860]: I1211 08:16:37.746850 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f1bc94aa-96ad-4e1f-a407-9968d42069aa-catalog-content\") pod \"community-operators-kzkbs\" (UID: \"f1bc94aa-96ad-4e1f-a407-9968d42069aa\") " pod="openshift-marketplace/community-operators-kzkbs" Dec 11 08:16:37 crc kubenswrapper[4860]: I1211 08:16:37.746946 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f1bc94aa-96ad-4e1f-a407-9968d42069aa-utilities\") pod \"community-operators-kzkbs\" (UID: \"f1bc94aa-96ad-4e1f-a407-9968d42069aa\") " pod="openshift-marketplace/community-operators-kzkbs" Dec 11 08:16:37 crc kubenswrapper[4860]: I1211 08:16:37.747230 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hg8h2\" (UniqueName: \"kubernetes.io/projected/f1bc94aa-96ad-4e1f-a407-9968d42069aa-kube-api-access-hg8h2\") pod \"community-operators-kzkbs\" (UID: \"f1bc94aa-96ad-4e1f-a407-9968d42069aa\") " pod="openshift-marketplace/community-operators-kzkbs" Dec 11 08:16:37 crc kubenswrapper[4860]: I1211 08:16:37.847945 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hg8h2\" (UniqueName: \"kubernetes.io/projected/f1bc94aa-96ad-4e1f-a407-9968d42069aa-kube-api-access-hg8h2\") pod \"community-operators-kzkbs\" (UID: \"f1bc94aa-96ad-4e1f-a407-9968d42069aa\") " pod="openshift-marketplace/community-operators-kzkbs" Dec 11 08:16:37 crc kubenswrapper[4860]: I1211 08:16:37.848000 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f1bc94aa-96ad-4e1f-a407-9968d42069aa-catalog-content\") pod \"community-operators-kzkbs\" (UID: \"f1bc94aa-96ad-4e1f-a407-9968d42069aa\") " pod="openshift-marketplace/community-operators-kzkbs" Dec 11 08:16:37 crc kubenswrapper[4860]: I1211 08:16:37.848027 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f1bc94aa-96ad-4e1f-a407-9968d42069aa-utilities\") pod \"community-operators-kzkbs\" (UID: \"f1bc94aa-96ad-4e1f-a407-9968d42069aa\") " pod="openshift-marketplace/community-operators-kzkbs" Dec 11 08:16:37 crc kubenswrapper[4860]: I1211 08:16:37.848483 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f1bc94aa-96ad-4e1f-a407-9968d42069aa-utilities\") pod \"community-operators-kzkbs\" (UID: \"f1bc94aa-96ad-4e1f-a407-9968d42069aa\") " pod="openshift-marketplace/community-operators-kzkbs" Dec 11 08:16:37 crc kubenswrapper[4860]: I1211 08:16:37.848632 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f1bc94aa-96ad-4e1f-a407-9968d42069aa-catalog-content\") pod \"community-operators-kzkbs\" (UID: \"f1bc94aa-96ad-4e1f-a407-9968d42069aa\") " pod="openshift-marketplace/community-operators-kzkbs" Dec 11 08:16:37 crc kubenswrapper[4860]: I1211 08:16:37.879697 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hg8h2\" (UniqueName: \"kubernetes.io/projected/f1bc94aa-96ad-4e1f-a407-9968d42069aa-kube-api-access-hg8h2\") pod \"community-operators-kzkbs\" (UID: \"f1bc94aa-96ad-4e1f-a407-9968d42069aa\") " pod="openshift-marketplace/community-operators-kzkbs" Dec 11 08:16:37 crc kubenswrapper[4860]: I1211 08:16:37.917993 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kzkbs" Dec 11 08:16:38 crc kubenswrapper[4860]: I1211 08:16:38.363366 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-kzkbs"] Dec 11 08:16:38 crc kubenswrapper[4860]: I1211 08:16:38.391618 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kzkbs" event={"ID":"f1bc94aa-96ad-4e1f-a407-9968d42069aa","Type":"ContainerStarted","Data":"77964d7f53fc02ff2e30b6b0254dc39258c8e1754e611a205ad1f7297ffbd522"} Dec 11 08:16:38 crc kubenswrapper[4860]: I1211 08:16:38.540897 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6b5d4bcb95-jqnl2"] Dec 11 08:16:38 crc kubenswrapper[4860]: I1211 08:16:38.541100 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6b5d4bcb95-jqnl2" podUID="477e8461-916b-4a89-9b1b-0414cbe1d4d6" containerName="route-controller-manager" containerID="cri-o://219ba1a832ab31c229f0a9cd07f03dec5403ecc0d385284499e9f73c2daf92c1" gracePeriod=30 Dec 11 08:16:38 crc kubenswrapper[4860]: I1211 08:16:38.584076 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-nkvw8"] Dec 11 08:16:38 crc kubenswrapper[4860]: I1211 08:16:38.585053 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nkvw8" Dec 11 08:16:38 crc kubenswrapper[4860]: I1211 08:16:38.587202 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 11 08:16:38 crc kubenswrapper[4860]: I1211 08:16:38.593757 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nkvw8"] Dec 11 08:16:38 crc kubenswrapper[4860]: I1211 08:16:38.760541 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ccc9q\" (UniqueName: \"kubernetes.io/projected/74e80c39-0b0c-42f6-88b6-a800ec910606-kube-api-access-ccc9q\") pod \"redhat-marketplace-nkvw8\" (UID: \"74e80c39-0b0c-42f6-88b6-a800ec910606\") " pod="openshift-marketplace/redhat-marketplace-nkvw8" Dec 11 08:16:38 crc kubenswrapper[4860]: I1211 08:16:38.760876 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74e80c39-0b0c-42f6-88b6-a800ec910606-utilities\") pod \"redhat-marketplace-nkvw8\" (UID: \"74e80c39-0b0c-42f6-88b6-a800ec910606\") " pod="openshift-marketplace/redhat-marketplace-nkvw8" Dec 11 08:16:38 crc kubenswrapper[4860]: I1211 08:16:38.760949 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74e80c39-0b0c-42f6-88b6-a800ec910606-catalog-content\") pod \"redhat-marketplace-nkvw8\" (UID: \"74e80c39-0b0c-42f6-88b6-a800ec910606\") " pod="openshift-marketplace/redhat-marketplace-nkvw8" Dec 11 08:16:38 crc kubenswrapper[4860]: I1211 08:16:38.861863 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74e80c39-0b0c-42f6-88b6-a800ec910606-utilities\") pod \"redhat-marketplace-nkvw8\" (UID: \"74e80c39-0b0c-42f6-88b6-a800ec910606\") " pod="openshift-marketplace/redhat-marketplace-nkvw8" Dec 11 08:16:38 crc kubenswrapper[4860]: I1211 08:16:38.861925 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74e80c39-0b0c-42f6-88b6-a800ec910606-catalog-content\") pod \"redhat-marketplace-nkvw8\" (UID: \"74e80c39-0b0c-42f6-88b6-a800ec910606\") " pod="openshift-marketplace/redhat-marketplace-nkvw8" Dec 11 08:16:38 crc kubenswrapper[4860]: I1211 08:16:38.861988 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ccc9q\" (UniqueName: \"kubernetes.io/projected/74e80c39-0b0c-42f6-88b6-a800ec910606-kube-api-access-ccc9q\") pod \"redhat-marketplace-nkvw8\" (UID: \"74e80c39-0b0c-42f6-88b6-a800ec910606\") " pod="openshift-marketplace/redhat-marketplace-nkvw8" Dec 11 08:16:38 crc kubenswrapper[4860]: I1211 08:16:38.862455 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74e80c39-0b0c-42f6-88b6-a800ec910606-utilities\") pod \"redhat-marketplace-nkvw8\" (UID: \"74e80c39-0b0c-42f6-88b6-a800ec910606\") " pod="openshift-marketplace/redhat-marketplace-nkvw8" Dec 11 08:16:38 crc kubenswrapper[4860]: I1211 08:16:38.862570 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74e80c39-0b0c-42f6-88b6-a800ec910606-catalog-content\") pod \"redhat-marketplace-nkvw8\" (UID: \"74e80c39-0b0c-42f6-88b6-a800ec910606\") " pod="openshift-marketplace/redhat-marketplace-nkvw8" Dec 11 08:16:38 crc kubenswrapper[4860]: I1211 08:16:38.894489 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ccc9q\" (UniqueName: \"kubernetes.io/projected/74e80c39-0b0c-42f6-88b6-a800ec910606-kube-api-access-ccc9q\") pod \"redhat-marketplace-nkvw8\" (UID: \"74e80c39-0b0c-42f6-88b6-a800ec910606\") " pod="openshift-marketplace/redhat-marketplace-nkvw8" Dec 11 08:16:38 crc kubenswrapper[4860]: I1211 08:16:38.934499 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6b5d4bcb95-jqnl2" Dec 11 08:16:38 crc kubenswrapper[4860]: I1211 08:16:38.943881 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nkvw8" Dec 11 08:16:39 crc kubenswrapper[4860]: I1211 08:16:39.064247 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/477e8461-916b-4a89-9b1b-0414cbe1d4d6-serving-cert\") pod \"477e8461-916b-4a89-9b1b-0414cbe1d4d6\" (UID: \"477e8461-916b-4a89-9b1b-0414cbe1d4d6\") " Dec 11 08:16:39 crc kubenswrapper[4860]: I1211 08:16:39.064720 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/477e8461-916b-4a89-9b1b-0414cbe1d4d6-client-ca\") pod \"477e8461-916b-4a89-9b1b-0414cbe1d4d6\" (UID: \"477e8461-916b-4a89-9b1b-0414cbe1d4d6\") " Dec 11 08:16:39 crc kubenswrapper[4860]: I1211 08:16:39.064788 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/477e8461-916b-4a89-9b1b-0414cbe1d4d6-config\") pod \"477e8461-916b-4a89-9b1b-0414cbe1d4d6\" (UID: \"477e8461-916b-4a89-9b1b-0414cbe1d4d6\") " Dec 11 08:16:39 crc kubenswrapper[4860]: I1211 08:16:39.064883 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbsf\" (UniqueName: \"kubernetes.io/projected/477e8461-916b-4a89-9b1b-0414cbe1d4d6-kube-api-access-jhbsf\") pod \"477e8461-916b-4a89-9b1b-0414cbe1d4d6\" (UID: \"477e8461-916b-4a89-9b1b-0414cbe1d4d6\") " Dec 11 08:16:39 crc kubenswrapper[4860]: I1211 08:16:39.066246 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/477e8461-916b-4a89-9b1b-0414cbe1d4d6-client-ca" (OuterVolumeSpecName: "client-ca") pod "477e8461-916b-4a89-9b1b-0414cbe1d4d6" (UID: "477e8461-916b-4a89-9b1b-0414cbe1d4d6"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:16:39 crc kubenswrapper[4860]: I1211 08:16:39.067116 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/477e8461-916b-4a89-9b1b-0414cbe1d4d6-config" (OuterVolumeSpecName: "config") pod "477e8461-916b-4a89-9b1b-0414cbe1d4d6" (UID: "477e8461-916b-4a89-9b1b-0414cbe1d4d6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:16:39 crc kubenswrapper[4860]: I1211 08:16:39.071760 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/477e8461-916b-4a89-9b1b-0414cbe1d4d6-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "477e8461-916b-4a89-9b1b-0414cbe1d4d6" (UID: "477e8461-916b-4a89-9b1b-0414cbe1d4d6"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:16:39 crc kubenswrapper[4860]: I1211 08:16:39.072846 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/477e8461-916b-4a89-9b1b-0414cbe1d4d6-kube-api-access-jhbsf" (OuterVolumeSpecName: "kube-api-access-jhbsf") pod "477e8461-916b-4a89-9b1b-0414cbe1d4d6" (UID: "477e8461-916b-4a89-9b1b-0414cbe1d4d6"). InnerVolumeSpecName "kube-api-access-jhbsf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:16:39 crc kubenswrapper[4860]: I1211 08:16:39.166224 4860 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/477e8461-916b-4a89-9b1b-0414cbe1d4d6-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 08:16:39 crc kubenswrapper[4860]: I1211 08:16:39.166278 4860 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/477e8461-916b-4a89-9b1b-0414cbe1d4d6-client-ca\") on node \"crc\" DevicePath \"\"" Dec 11 08:16:39 crc kubenswrapper[4860]: I1211 08:16:39.166291 4860 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/477e8461-916b-4a89-9b1b-0414cbe1d4d6-config\") on node \"crc\" DevicePath \"\"" Dec 11 08:16:39 crc kubenswrapper[4860]: I1211 08:16:39.166304 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbsf\" (UniqueName: \"kubernetes.io/projected/477e8461-916b-4a89-9b1b-0414cbe1d4d6-kube-api-access-jhbsf\") on node \"crc\" DevicePath \"\"" Dec 11 08:16:39 crc kubenswrapper[4860]: I1211 08:16:39.340719 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nkvw8"] Dec 11 08:16:39 crc kubenswrapper[4860]: W1211 08:16:39.346501 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod74e80c39_0b0c_42f6_88b6_a800ec910606.slice/crio-d9402b3dfee76233153634df67bcf13d5cfdcbbcd2e78ea9ce34c45cc2411d6d WatchSource:0}: Error finding container d9402b3dfee76233153634df67bcf13d5cfdcbbcd2e78ea9ce34c45cc2411d6d: Status 404 returned error can't find the container with id d9402b3dfee76233153634df67bcf13d5cfdcbbcd2e78ea9ce34c45cc2411d6d Dec 11 08:16:39 crc kubenswrapper[4860]: I1211 08:16:39.398632 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nkvw8" event={"ID":"74e80c39-0b0c-42f6-88b6-a800ec910606","Type":"ContainerStarted","Data":"d9402b3dfee76233153634df67bcf13d5cfdcbbcd2e78ea9ce34c45cc2411d6d"} Dec 11 08:16:39 crc kubenswrapper[4860]: I1211 08:16:39.400125 4860 generic.go:334] "Generic (PLEG): container finished" podID="477e8461-916b-4a89-9b1b-0414cbe1d4d6" containerID="219ba1a832ab31c229f0a9cd07f03dec5403ecc0d385284499e9f73c2daf92c1" exitCode=0 Dec 11 08:16:39 crc kubenswrapper[4860]: I1211 08:16:39.400181 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6b5d4bcb95-jqnl2" event={"ID":"477e8461-916b-4a89-9b1b-0414cbe1d4d6","Type":"ContainerDied","Data":"219ba1a832ab31c229f0a9cd07f03dec5403ecc0d385284499e9f73c2daf92c1"} Dec 11 08:16:39 crc kubenswrapper[4860]: I1211 08:16:39.400202 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6b5d4bcb95-jqnl2" event={"ID":"477e8461-916b-4a89-9b1b-0414cbe1d4d6","Type":"ContainerDied","Data":"e312fc2b7208920a37da4708212e91ebf208a741fc1b5f6154fa0c57496457e1"} Dec 11 08:16:39 crc kubenswrapper[4860]: I1211 08:16:39.400218 4860 scope.go:117] "RemoveContainer" containerID="219ba1a832ab31c229f0a9cd07f03dec5403ecc0d385284499e9f73c2daf92c1" Dec 11 08:16:39 crc kubenswrapper[4860]: I1211 08:16:39.400246 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6b5d4bcb95-jqnl2" Dec 11 08:16:39 crc kubenswrapper[4860]: I1211 08:16:39.401663 4860 generic.go:334] "Generic (PLEG): container finished" podID="f1bc94aa-96ad-4e1f-a407-9968d42069aa" containerID="e1d538356ed81690df0ba1420e84cf871a973ee2319617b69487f7afcb3e136d" exitCode=0 Dec 11 08:16:39 crc kubenswrapper[4860]: I1211 08:16:39.401686 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kzkbs" event={"ID":"f1bc94aa-96ad-4e1f-a407-9968d42069aa","Type":"ContainerDied","Data":"e1d538356ed81690df0ba1420e84cf871a973ee2319617b69487f7afcb3e136d"} Dec 11 08:16:39 crc kubenswrapper[4860]: I1211 08:16:39.416104 4860 scope.go:117] "RemoveContainer" containerID="219ba1a832ab31c229f0a9cd07f03dec5403ecc0d385284499e9f73c2daf92c1" Dec 11 08:16:39 crc kubenswrapper[4860]: E1211 08:16:39.416448 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"219ba1a832ab31c229f0a9cd07f03dec5403ecc0d385284499e9f73c2daf92c1\": container with ID starting with 219ba1a832ab31c229f0a9cd07f03dec5403ecc0d385284499e9f73c2daf92c1 not found: ID does not exist" containerID="219ba1a832ab31c229f0a9cd07f03dec5403ecc0d385284499e9f73c2daf92c1" Dec 11 08:16:39 crc kubenswrapper[4860]: I1211 08:16:39.416479 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"219ba1a832ab31c229f0a9cd07f03dec5403ecc0d385284499e9f73c2daf92c1"} err="failed to get container status \"219ba1a832ab31c229f0a9cd07f03dec5403ecc0d385284499e9f73c2daf92c1\": rpc error: code = NotFound desc = could not find container \"219ba1a832ab31c229f0a9cd07f03dec5403ecc0d385284499e9f73c2daf92c1\": container with ID starting with 219ba1a832ab31c229f0a9cd07f03dec5403ecc0d385284499e9f73c2daf92c1 not found: ID does not exist" Dec 11 08:16:39 crc kubenswrapper[4860]: I1211 08:16:39.463266 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6b5d4bcb95-jqnl2"] Dec 11 08:16:39 crc kubenswrapper[4860]: I1211 08:16:39.466296 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6b5d4bcb95-jqnl2"] Dec 11 08:16:39 crc kubenswrapper[4860]: I1211 08:16:39.587765 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="477e8461-916b-4a89-9b1b-0414cbe1d4d6" path="/var/lib/kubelet/pods/477e8461-916b-4a89-9b1b-0414cbe1d4d6/volumes" Dec 11 08:16:39 crc kubenswrapper[4860]: I1211 08:16:39.647517 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7bb5f84cb5-cjnbd"] Dec 11 08:16:39 crc kubenswrapper[4860]: E1211 08:16:39.648143 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="477e8461-916b-4a89-9b1b-0414cbe1d4d6" containerName="route-controller-manager" Dec 11 08:16:39 crc kubenswrapper[4860]: I1211 08:16:39.648162 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="477e8461-916b-4a89-9b1b-0414cbe1d4d6" containerName="route-controller-manager" Dec 11 08:16:39 crc kubenswrapper[4860]: I1211 08:16:39.648339 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="477e8461-916b-4a89-9b1b-0414cbe1d4d6" containerName="route-controller-manager" Dec 11 08:16:39 crc kubenswrapper[4860]: I1211 08:16:39.649090 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7bb5f84cb5-cjnbd" Dec 11 08:16:39 crc kubenswrapper[4860]: I1211 08:16:39.654103 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 11 08:16:39 crc kubenswrapper[4860]: I1211 08:16:39.654244 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 11 08:16:39 crc kubenswrapper[4860]: I1211 08:16:39.654379 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 11 08:16:39 crc kubenswrapper[4860]: I1211 08:16:39.654543 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 11 08:16:39 crc kubenswrapper[4860]: I1211 08:16:39.654678 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 11 08:16:39 crc kubenswrapper[4860]: I1211 08:16:39.654766 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 11 08:16:39 crc kubenswrapper[4860]: I1211 08:16:39.668156 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7bb5f84cb5-cjnbd"] Dec 11 08:16:39 crc kubenswrapper[4860]: I1211 08:16:39.676761 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pl6bb\" (UniqueName: \"kubernetes.io/projected/4554553e-34a3-42a0-8829-3a5937e82558-kube-api-access-pl6bb\") pod \"route-controller-manager-7bb5f84cb5-cjnbd\" (UID: \"4554553e-34a3-42a0-8829-3a5937e82558\") " pod="openshift-route-controller-manager/route-controller-manager-7bb5f84cb5-cjnbd" Dec 11 08:16:39 crc kubenswrapper[4860]: I1211 08:16:39.677112 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4554553e-34a3-42a0-8829-3a5937e82558-serving-cert\") pod \"route-controller-manager-7bb5f84cb5-cjnbd\" (UID: \"4554553e-34a3-42a0-8829-3a5937e82558\") " pod="openshift-route-controller-manager/route-controller-manager-7bb5f84cb5-cjnbd" Dec 11 08:16:39 crc kubenswrapper[4860]: I1211 08:16:39.677246 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4554553e-34a3-42a0-8829-3a5937e82558-config\") pod \"route-controller-manager-7bb5f84cb5-cjnbd\" (UID: \"4554553e-34a3-42a0-8829-3a5937e82558\") " pod="openshift-route-controller-manager/route-controller-manager-7bb5f84cb5-cjnbd" Dec 11 08:16:39 crc kubenswrapper[4860]: I1211 08:16:39.677352 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4554553e-34a3-42a0-8829-3a5937e82558-client-ca\") pod \"route-controller-manager-7bb5f84cb5-cjnbd\" (UID: \"4554553e-34a3-42a0-8829-3a5937e82558\") " pod="openshift-route-controller-manager/route-controller-manager-7bb5f84cb5-cjnbd" Dec 11 08:16:39 crc kubenswrapper[4860]: I1211 08:16:39.778495 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4554553e-34a3-42a0-8829-3a5937e82558-client-ca\") pod \"route-controller-manager-7bb5f84cb5-cjnbd\" (UID: \"4554553e-34a3-42a0-8829-3a5937e82558\") " pod="openshift-route-controller-manager/route-controller-manager-7bb5f84cb5-cjnbd" Dec 11 08:16:39 crc kubenswrapper[4860]: I1211 08:16:39.778608 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pl6bb\" (UniqueName: \"kubernetes.io/projected/4554553e-34a3-42a0-8829-3a5937e82558-kube-api-access-pl6bb\") pod \"route-controller-manager-7bb5f84cb5-cjnbd\" (UID: \"4554553e-34a3-42a0-8829-3a5937e82558\") " pod="openshift-route-controller-manager/route-controller-manager-7bb5f84cb5-cjnbd" Dec 11 08:16:39 crc kubenswrapper[4860]: I1211 08:16:39.778665 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4554553e-34a3-42a0-8829-3a5937e82558-serving-cert\") pod \"route-controller-manager-7bb5f84cb5-cjnbd\" (UID: \"4554553e-34a3-42a0-8829-3a5937e82558\") " pod="openshift-route-controller-manager/route-controller-manager-7bb5f84cb5-cjnbd" Dec 11 08:16:39 crc kubenswrapper[4860]: I1211 08:16:39.778730 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4554553e-34a3-42a0-8829-3a5937e82558-config\") pod \"route-controller-manager-7bb5f84cb5-cjnbd\" (UID: \"4554553e-34a3-42a0-8829-3a5937e82558\") " pod="openshift-route-controller-manager/route-controller-manager-7bb5f84cb5-cjnbd" Dec 11 08:16:39 crc kubenswrapper[4860]: I1211 08:16:39.779882 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4554553e-34a3-42a0-8829-3a5937e82558-client-ca\") pod \"route-controller-manager-7bb5f84cb5-cjnbd\" (UID: \"4554553e-34a3-42a0-8829-3a5937e82558\") " pod="openshift-route-controller-manager/route-controller-manager-7bb5f84cb5-cjnbd" Dec 11 08:16:39 crc kubenswrapper[4860]: I1211 08:16:39.780386 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4554553e-34a3-42a0-8829-3a5937e82558-config\") pod \"route-controller-manager-7bb5f84cb5-cjnbd\" (UID: \"4554553e-34a3-42a0-8829-3a5937e82558\") " pod="openshift-route-controller-manager/route-controller-manager-7bb5f84cb5-cjnbd" Dec 11 08:16:39 crc kubenswrapper[4860]: I1211 08:16:39.791142 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4554553e-34a3-42a0-8829-3a5937e82558-serving-cert\") pod \"route-controller-manager-7bb5f84cb5-cjnbd\" (UID: \"4554553e-34a3-42a0-8829-3a5937e82558\") " pod="openshift-route-controller-manager/route-controller-manager-7bb5f84cb5-cjnbd" Dec 11 08:16:39 crc kubenswrapper[4860]: I1211 08:16:39.812810 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pl6bb\" (UniqueName: \"kubernetes.io/projected/4554553e-34a3-42a0-8829-3a5937e82558-kube-api-access-pl6bb\") pod \"route-controller-manager-7bb5f84cb5-cjnbd\" (UID: \"4554553e-34a3-42a0-8829-3a5937e82558\") " pod="openshift-route-controller-manager/route-controller-manager-7bb5f84cb5-cjnbd" Dec 11 08:16:39 crc kubenswrapper[4860]: I1211 08:16:39.966430 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7bb5f84cb5-cjnbd" Dec 11 08:16:39 crc kubenswrapper[4860]: I1211 08:16:39.980754 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-7vj8m"] Dec 11 08:16:39 crc kubenswrapper[4860]: I1211 08:16:39.982070 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7vj8m" Dec 11 08:16:39 crc kubenswrapper[4860]: I1211 08:16:39.984064 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 11 08:16:39 crc kubenswrapper[4860]: I1211 08:16:39.991900 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7vj8m"] Dec 11 08:16:40 crc kubenswrapper[4860]: I1211 08:16:40.082733 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-245cv\" (UniqueName: \"kubernetes.io/projected/7d8a48c0-183c-4699-86d1-f6d0ef767cdf-kube-api-access-245cv\") pod \"redhat-operators-7vj8m\" (UID: \"7d8a48c0-183c-4699-86d1-f6d0ef767cdf\") " pod="openshift-marketplace/redhat-operators-7vj8m" Dec 11 08:16:40 crc kubenswrapper[4860]: I1211 08:16:40.082806 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d8a48c0-183c-4699-86d1-f6d0ef767cdf-utilities\") pod \"redhat-operators-7vj8m\" (UID: \"7d8a48c0-183c-4699-86d1-f6d0ef767cdf\") " pod="openshift-marketplace/redhat-operators-7vj8m" Dec 11 08:16:40 crc kubenswrapper[4860]: I1211 08:16:40.082964 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d8a48c0-183c-4699-86d1-f6d0ef767cdf-catalog-content\") pod \"redhat-operators-7vj8m\" (UID: \"7d8a48c0-183c-4699-86d1-f6d0ef767cdf\") " pod="openshift-marketplace/redhat-operators-7vj8m" Dec 11 08:16:40 crc kubenswrapper[4860]: I1211 08:16:40.184198 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d8a48c0-183c-4699-86d1-f6d0ef767cdf-catalog-content\") pod \"redhat-operators-7vj8m\" (UID: \"7d8a48c0-183c-4699-86d1-f6d0ef767cdf\") " pod="openshift-marketplace/redhat-operators-7vj8m" Dec 11 08:16:40 crc kubenswrapper[4860]: I1211 08:16:40.184464 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-245cv\" (UniqueName: \"kubernetes.io/projected/7d8a48c0-183c-4699-86d1-f6d0ef767cdf-kube-api-access-245cv\") pod \"redhat-operators-7vj8m\" (UID: \"7d8a48c0-183c-4699-86d1-f6d0ef767cdf\") " pod="openshift-marketplace/redhat-operators-7vj8m" Dec 11 08:16:40 crc kubenswrapper[4860]: I1211 08:16:40.184489 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d8a48c0-183c-4699-86d1-f6d0ef767cdf-utilities\") pod \"redhat-operators-7vj8m\" (UID: \"7d8a48c0-183c-4699-86d1-f6d0ef767cdf\") " pod="openshift-marketplace/redhat-operators-7vj8m" Dec 11 08:16:40 crc kubenswrapper[4860]: I1211 08:16:40.184734 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d8a48c0-183c-4699-86d1-f6d0ef767cdf-catalog-content\") pod \"redhat-operators-7vj8m\" (UID: \"7d8a48c0-183c-4699-86d1-f6d0ef767cdf\") " pod="openshift-marketplace/redhat-operators-7vj8m" Dec 11 08:16:40 crc kubenswrapper[4860]: I1211 08:16:40.185203 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d8a48c0-183c-4699-86d1-f6d0ef767cdf-utilities\") pod \"redhat-operators-7vj8m\" (UID: \"7d8a48c0-183c-4699-86d1-f6d0ef767cdf\") " pod="openshift-marketplace/redhat-operators-7vj8m" Dec 11 08:16:40 crc kubenswrapper[4860]: I1211 08:16:40.204772 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-245cv\" (UniqueName: \"kubernetes.io/projected/7d8a48c0-183c-4699-86d1-f6d0ef767cdf-kube-api-access-245cv\") pod \"redhat-operators-7vj8m\" (UID: \"7d8a48c0-183c-4699-86d1-f6d0ef767cdf\") " pod="openshift-marketplace/redhat-operators-7vj8m" Dec 11 08:16:40 crc kubenswrapper[4860]: I1211 08:16:40.331265 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7vj8m" Dec 11 08:16:40 crc kubenswrapper[4860]: I1211 08:16:40.394613 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7bb5f84cb5-cjnbd"] Dec 11 08:16:40 crc kubenswrapper[4860]: I1211 08:16:40.412034 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kzkbs" event={"ID":"f1bc94aa-96ad-4e1f-a407-9968d42069aa","Type":"ContainerStarted","Data":"8aea4c1bcaebc0d43f6d63cd032c6d3051ede272192315bb12fd5915143bbe14"} Dec 11 08:16:40 crc kubenswrapper[4860]: I1211 08:16:40.415630 4860 generic.go:334] "Generic (PLEG): container finished" podID="74e80c39-0b0c-42f6-88b6-a800ec910606" containerID="c21f19d3510a5fe71fa89259260b88a9b3f2cafed75964f28c3dbc8c3558ba62" exitCode=0 Dec 11 08:16:40 crc kubenswrapper[4860]: I1211 08:16:40.415758 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nkvw8" event={"ID":"74e80c39-0b0c-42f6-88b6-a800ec910606","Type":"ContainerDied","Data":"c21f19d3510a5fe71fa89259260b88a9b3f2cafed75964f28c3dbc8c3558ba62"} Dec 11 08:16:40 crc kubenswrapper[4860]: I1211 08:16:40.417470 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7bb5f84cb5-cjnbd" event={"ID":"4554553e-34a3-42a0-8829-3a5937e82558","Type":"ContainerStarted","Data":"73e0d1229e4918f1349ac8b1f6813c3542a131de6eba70fba8765a87e870bdd2"} Dec 11 08:16:40 crc kubenswrapper[4860]: I1211 08:16:40.833141 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7vj8m"] Dec 11 08:16:40 crc kubenswrapper[4860]: W1211 08:16:40.838821 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7d8a48c0_183c_4699_86d1_f6d0ef767cdf.slice/crio-4305010f17c935a4adae7200b47379619ead7cb09b259dce9c2246a84f90190f WatchSource:0}: Error finding container 4305010f17c935a4adae7200b47379619ead7cb09b259dce9c2246a84f90190f: Status 404 returned error can't find the container with id 4305010f17c935a4adae7200b47379619ead7cb09b259dce9c2246a84f90190f Dec 11 08:16:41 crc kubenswrapper[4860]: I1211 08:16:41.187015 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-bg2t9"] Dec 11 08:16:41 crc kubenswrapper[4860]: I1211 08:16:41.189789 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bg2t9" Dec 11 08:16:41 crc kubenswrapper[4860]: I1211 08:16:41.192511 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 11 08:16:41 crc kubenswrapper[4860]: I1211 08:16:41.210009 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/79181f73-fb9e-40c5-96b8-0ad95545f1f7-catalog-content\") pod \"certified-operators-bg2t9\" (UID: \"79181f73-fb9e-40c5-96b8-0ad95545f1f7\") " pod="openshift-marketplace/certified-operators-bg2t9" Dec 11 08:16:41 crc kubenswrapper[4860]: I1211 08:16:41.210062 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/79181f73-fb9e-40c5-96b8-0ad95545f1f7-utilities\") pod \"certified-operators-bg2t9\" (UID: \"79181f73-fb9e-40c5-96b8-0ad95545f1f7\") " pod="openshift-marketplace/certified-operators-bg2t9" Dec 11 08:16:41 crc kubenswrapper[4860]: I1211 08:16:41.210114 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-snsww\" (UniqueName: \"kubernetes.io/projected/79181f73-fb9e-40c5-96b8-0ad95545f1f7-kube-api-access-snsww\") pod \"certified-operators-bg2t9\" (UID: \"79181f73-fb9e-40c5-96b8-0ad95545f1f7\") " pod="openshift-marketplace/certified-operators-bg2t9" Dec 11 08:16:41 crc kubenswrapper[4860]: I1211 08:16:41.214545 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-bg2t9"] Dec 11 08:16:41 crc kubenswrapper[4860]: I1211 08:16:41.311333 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-snsww\" (UniqueName: \"kubernetes.io/projected/79181f73-fb9e-40c5-96b8-0ad95545f1f7-kube-api-access-snsww\") pod \"certified-operators-bg2t9\" (UID: \"79181f73-fb9e-40c5-96b8-0ad95545f1f7\") " pod="openshift-marketplace/certified-operators-bg2t9" Dec 11 08:16:41 crc kubenswrapper[4860]: I1211 08:16:41.311425 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/79181f73-fb9e-40c5-96b8-0ad95545f1f7-catalog-content\") pod \"certified-operators-bg2t9\" (UID: \"79181f73-fb9e-40c5-96b8-0ad95545f1f7\") " pod="openshift-marketplace/certified-operators-bg2t9" Dec 11 08:16:41 crc kubenswrapper[4860]: I1211 08:16:41.311460 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/79181f73-fb9e-40c5-96b8-0ad95545f1f7-utilities\") pod \"certified-operators-bg2t9\" (UID: \"79181f73-fb9e-40c5-96b8-0ad95545f1f7\") " pod="openshift-marketplace/certified-operators-bg2t9" Dec 11 08:16:41 crc kubenswrapper[4860]: I1211 08:16:41.312061 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/79181f73-fb9e-40c5-96b8-0ad95545f1f7-utilities\") pod \"certified-operators-bg2t9\" (UID: \"79181f73-fb9e-40c5-96b8-0ad95545f1f7\") " pod="openshift-marketplace/certified-operators-bg2t9" Dec 11 08:16:41 crc kubenswrapper[4860]: I1211 08:16:41.312361 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/79181f73-fb9e-40c5-96b8-0ad95545f1f7-catalog-content\") pod \"certified-operators-bg2t9\" (UID: \"79181f73-fb9e-40c5-96b8-0ad95545f1f7\") " pod="openshift-marketplace/certified-operators-bg2t9" Dec 11 08:16:41 crc kubenswrapper[4860]: I1211 08:16:41.335257 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-snsww\" (UniqueName: \"kubernetes.io/projected/79181f73-fb9e-40c5-96b8-0ad95545f1f7-kube-api-access-snsww\") pod \"certified-operators-bg2t9\" (UID: \"79181f73-fb9e-40c5-96b8-0ad95545f1f7\") " pod="openshift-marketplace/certified-operators-bg2t9" Dec 11 08:16:41 crc kubenswrapper[4860]: I1211 08:16:41.423280 4860 generic.go:334] "Generic (PLEG): container finished" podID="f1bc94aa-96ad-4e1f-a407-9968d42069aa" containerID="8aea4c1bcaebc0d43f6d63cd032c6d3051ede272192315bb12fd5915143bbe14" exitCode=0 Dec 11 08:16:41 crc kubenswrapper[4860]: I1211 08:16:41.423335 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kzkbs" event={"ID":"f1bc94aa-96ad-4e1f-a407-9968d42069aa","Type":"ContainerDied","Data":"8aea4c1bcaebc0d43f6d63cd032c6d3051ede272192315bb12fd5915143bbe14"} Dec 11 08:16:41 crc kubenswrapper[4860]: I1211 08:16:41.426378 4860 generic.go:334] "Generic (PLEG): container finished" podID="74e80c39-0b0c-42f6-88b6-a800ec910606" containerID="2aeffab23dcd4e5f80b6fdeced8b0a8bfd37d0bea9d5d18ebef4b173f5cc0376" exitCode=0 Dec 11 08:16:41 crc kubenswrapper[4860]: I1211 08:16:41.426455 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nkvw8" event={"ID":"74e80c39-0b0c-42f6-88b6-a800ec910606","Type":"ContainerDied","Data":"2aeffab23dcd4e5f80b6fdeced8b0a8bfd37d0bea9d5d18ebef4b173f5cc0376"} Dec 11 08:16:41 crc kubenswrapper[4860]: I1211 08:16:41.428397 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7bb5f84cb5-cjnbd" event={"ID":"4554553e-34a3-42a0-8829-3a5937e82558","Type":"ContainerStarted","Data":"c71fd082f58b2f0d2efe4da77627b5c49a784beb83e86ee318a0d54df7cddcb7"} Dec 11 08:16:41 crc kubenswrapper[4860]: I1211 08:16:41.428898 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-7bb5f84cb5-cjnbd" Dec 11 08:16:41 crc kubenswrapper[4860]: I1211 08:16:41.431160 4860 generic.go:334] "Generic (PLEG): container finished" podID="7d8a48c0-183c-4699-86d1-f6d0ef767cdf" containerID="c9ff932c73fb694c3cbcdedb8475fd4bbe3c2cba96e8055301f808e6b24130aa" exitCode=0 Dec 11 08:16:41 crc kubenswrapper[4860]: I1211 08:16:41.431193 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7vj8m" event={"ID":"7d8a48c0-183c-4699-86d1-f6d0ef767cdf","Type":"ContainerDied","Data":"c9ff932c73fb694c3cbcdedb8475fd4bbe3c2cba96e8055301f808e6b24130aa"} Dec 11 08:16:41 crc kubenswrapper[4860]: I1211 08:16:41.431219 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7vj8m" event={"ID":"7d8a48c0-183c-4699-86d1-f6d0ef767cdf","Type":"ContainerStarted","Data":"4305010f17c935a4adae7200b47379619ead7cb09b259dce9c2246a84f90190f"} Dec 11 08:16:41 crc kubenswrapper[4860]: I1211 08:16:41.434704 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-7bb5f84cb5-cjnbd" Dec 11 08:16:41 crc kubenswrapper[4860]: I1211 08:16:41.506537 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-7bb5f84cb5-cjnbd" podStartSLOduration=3.506510962 podStartE2EDuration="3.506510962s" podCreationTimestamp="2025-12-11 08:16:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:16:41.504092083 +0000 UTC m=+334.232611138" watchObservedRunningTime="2025-12-11 08:16:41.506510962 +0000 UTC m=+334.235030017" Dec 11 08:16:41 crc kubenswrapper[4860]: I1211 08:16:41.535193 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bg2t9" Dec 11 08:16:41 crc kubenswrapper[4860]: I1211 08:16:41.988014 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-bg2t9"] Dec 11 08:16:42 crc kubenswrapper[4860]: I1211 08:16:42.438140 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nkvw8" event={"ID":"74e80c39-0b0c-42f6-88b6-a800ec910606","Type":"ContainerStarted","Data":"209ec812c9463e169f70200c4271e34417c36021a7e653288ff7ca6599503031"} Dec 11 08:16:42 crc kubenswrapper[4860]: I1211 08:16:42.439675 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7vj8m" event={"ID":"7d8a48c0-183c-4699-86d1-f6d0ef767cdf","Type":"ContainerStarted","Data":"1c77d8a7b4ccde5fc359da5ae850dde6040f66cb085c51fbbe2b4d591d5573fe"} Dec 11 08:16:42 crc kubenswrapper[4860]: I1211 08:16:42.441027 4860 generic.go:334] "Generic (PLEG): container finished" podID="79181f73-fb9e-40c5-96b8-0ad95545f1f7" containerID="d1585bcd928ee3378d845750d6f7e2ccf054829034403190ffd43b2c41726c42" exitCode=0 Dec 11 08:16:42 crc kubenswrapper[4860]: I1211 08:16:42.441093 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bg2t9" event={"ID":"79181f73-fb9e-40c5-96b8-0ad95545f1f7","Type":"ContainerDied","Data":"d1585bcd928ee3378d845750d6f7e2ccf054829034403190ffd43b2c41726c42"} Dec 11 08:16:42 crc kubenswrapper[4860]: I1211 08:16:42.441152 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bg2t9" event={"ID":"79181f73-fb9e-40c5-96b8-0ad95545f1f7","Type":"ContainerStarted","Data":"5ae1ad5001c6639b6d383a7eb9286f413e7f0cdd18c8c3feba0845a9ef4a64ca"} Dec 11 08:16:42 crc kubenswrapper[4860]: I1211 08:16:42.443183 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kzkbs" event={"ID":"f1bc94aa-96ad-4e1f-a407-9968d42069aa","Type":"ContainerStarted","Data":"45ad7fbdfad4954799ec9369ef820e53904f07ea80d0a26faabf8f6730338b00"} Dec 11 08:16:42 crc kubenswrapper[4860]: I1211 08:16:42.459507 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-nkvw8" podStartSLOduration=3.010363536 podStartE2EDuration="4.459488173s" podCreationTimestamp="2025-12-11 08:16:38 +0000 UTC" firstStartedPulling="2025-12-11 08:16:40.417674122 +0000 UTC m=+333.146193197" lastFinishedPulling="2025-12-11 08:16:41.866798779 +0000 UTC m=+334.595317834" observedRunningTime="2025-12-11 08:16:42.455289104 +0000 UTC m=+335.183808159" watchObservedRunningTime="2025-12-11 08:16:42.459488173 +0000 UTC m=+335.188007228" Dec 11 08:16:42 crc kubenswrapper[4860]: I1211 08:16:42.472117 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-kzkbs" podStartSLOduration=2.815598478 podStartE2EDuration="5.472098409s" podCreationTimestamp="2025-12-11 08:16:37 +0000 UTC" firstStartedPulling="2025-12-11 08:16:39.405327473 +0000 UTC m=+332.133846528" lastFinishedPulling="2025-12-11 08:16:42.061827404 +0000 UTC m=+334.790346459" observedRunningTime="2025-12-11 08:16:42.469797523 +0000 UTC m=+335.198316568" watchObservedRunningTime="2025-12-11 08:16:42.472098409 +0000 UTC m=+335.200617464" Dec 11 08:16:43 crc kubenswrapper[4860]: I1211 08:16:43.101793 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-555fn"] Dec 11 08:16:43 crc kubenswrapper[4860]: I1211 08:16:43.102936 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-555fn" Dec 11 08:16:43 crc kubenswrapper[4860]: I1211 08:16:43.121225 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-555fn"] Dec 11 08:16:43 crc kubenswrapper[4860]: I1211 08:16:43.240512 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/5ef69e5e-c1ab-4d1b-9cf8-3ad23b9ff27f-installation-pull-secrets\") pod \"image-registry-66df7c8f76-555fn\" (UID: \"5ef69e5e-c1ab-4d1b-9cf8-3ad23b9ff27f\") " pod="openshift-image-registry/image-registry-66df7c8f76-555fn" Dec 11 08:16:43 crc kubenswrapper[4860]: I1211 08:16:43.240614 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/5ef69e5e-c1ab-4d1b-9cf8-3ad23b9ff27f-registry-certificates\") pod \"image-registry-66df7c8f76-555fn\" (UID: \"5ef69e5e-c1ab-4d1b-9cf8-3ad23b9ff27f\") " pod="openshift-image-registry/image-registry-66df7c8f76-555fn" Dec 11 08:16:43 crc kubenswrapper[4860]: I1211 08:16:43.240662 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/5ef69e5e-c1ab-4d1b-9cf8-3ad23b9ff27f-ca-trust-extracted\") pod \"image-registry-66df7c8f76-555fn\" (UID: \"5ef69e5e-c1ab-4d1b-9cf8-3ad23b9ff27f\") " pod="openshift-image-registry/image-registry-66df7c8f76-555fn" Dec 11 08:16:43 crc kubenswrapper[4860]: I1211 08:16:43.240695 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5ef69e5e-c1ab-4d1b-9cf8-3ad23b9ff27f-bound-sa-token\") pod \"image-registry-66df7c8f76-555fn\" (UID: \"5ef69e5e-c1ab-4d1b-9cf8-3ad23b9ff27f\") " pod="openshift-image-registry/image-registry-66df7c8f76-555fn" Dec 11 08:16:43 crc kubenswrapper[4860]: I1211 08:16:43.240781 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5ef69e5e-c1ab-4d1b-9cf8-3ad23b9ff27f-trusted-ca\") pod \"image-registry-66df7c8f76-555fn\" (UID: \"5ef69e5e-c1ab-4d1b-9cf8-3ad23b9ff27f\") " pod="openshift-image-registry/image-registry-66df7c8f76-555fn" Dec 11 08:16:43 crc kubenswrapper[4860]: I1211 08:16:43.240836 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-555fn\" (UID: \"5ef69e5e-c1ab-4d1b-9cf8-3ad23b9ff27f\") " pod="openshift-image-registry/image-registry-66df7c8f76-555fn" Dec 11 08:16:43 crc kubenswrapper[4860]: I1211 08:16:43.240889 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nfddd\" (UniqueName: \"kubernetes.io/projected/5ef69e5e-c1ab-4d1b-9cf8-3ad23b9ff27f-kube-api-access-nfddd\") pod \"image-registry-66df7c8f76-555fn\" (UID: \"5ef69e5e-c1ab-4d1b-9cf8-3ad23b9ff27f\") " pod="openshift-image-registry/image-registry-66df7c8f76-555fn" Dec 11 08:16:43 crc kubenswrapper[4860]: I1211 08:16:43.240942 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/5ef69e5e-c1ab-4d1b-9cf8-3ad23b9ff27f-registry-tls\") pod \"image-registry-66df7c8f76-555fn\" (UID: \"5ef69e5e-c1ab-4d1b-9cf8-3ad23b9ff27f\") " pod="openshift-image-registry/image-registry-66df7c8f76-555fn" Dec 11 08:16:43 crc kubenswrapper[4860]: I1211 08:16:43.260145 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-555fn\" (UID: \"5ef69e5e-c1ab-4d1b-9cf8-3ad23b9ff27f\") " pod="openshift-image-registry/image-registry-66df7c8f76-555fn" Dec 11 08:16:43 crc kubenswrapper[4860]: I1211 08:16:43.342169 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/5ef69e5e-c1ab-4d1b-9cf8-3ad23b9ff27f-installation-pull-secrets\") pod \"image-registry-66df7c8f76-555fn\" (UID: \"5ef69e5e-c1ab-4d1b-9cf8-3ad23b9ff27f\") " pod="openshift-image-registry/image-registry-66df7c8f76-555fn" Dec 11 08:16:43 crc kubenswrapper[4860]: I1211 08:16:43.342242 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/5ef69e5e-c1ab-4d1b-9cf8-3ad23b9ff27f-ca-trust-extracted\") pod \"image-registry-66df7c8f76-555fn\" (UID: \"5ef69e5e-c1ab-4d1b-9cf8-3ad23b9ff27f\") " pod="openshift-image-registry/image-registry-66df7c8f76-555fn" Dec 11 08:16:43 crc kubenswrapper[4860]: I1211 08:16:43.342264 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/5ef69e5e-c1ab-4d1b-9cf8-3ad23b9ff27f-registry-certificates\") pod \"image-registry-66df7c8f76-555fn\" (UID: \"5ef69e5e-c1ab-4d1b-9cf8-3ad23b9ff27f\") " pod="openshift-image-registry/image-registry-66df7c8f76-555fn" Dec 11 08:16:43 crc kubenswrapper[4860]: I1211 08:16:43.342282 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5ef69e5e-c1ab-4d1b-9cf8-3ad23b9ff27f-bound-sa-token\") pod \"image-registry-66df7c8f76-555fn\" (UID: \"5ef69e5e-c1ab-4d1b-9cf8-3ad23b9ff27f\") " pod="openshift-image-registry/image-registry-66df7c8f76-555fn" Dec 11 08:16:43 crc kubenswrapper[4860]: I1211 08:16:43.342302 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5ef69e5e-c1ab-4d1b-9cf8-3ad23b9ff27f-trusted-ca\") pod \"image-registry-66df7c8f76-555fn\" (UID: \"5ef69e5e-c1ab-4d1b-9cf8-3ad23b9ff27f\") " pod="openshift-image-registry/image-registry-66df7c8f76-555fn" Dec 11 08:16:43 crc kubenswrapper[4860]: I1211 08:16:43.342326 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nfddd\" (UniqueName: \"kubernetes.io/projected/5ef69e5e-c1ab-4d1b-9cf8-3ad23b9ff27f-kube-api-access-nfddd\") pod \"image-registry-66df7c8f76-555fn\" (UID: \"5ef69e5e-c1ab-4d1b-9cf8-3ad23b9ff27f\") " pod="openshift-image-registry/image-registry-66df7c8f76-555fn" Dec 11 08:16:43 crc kubenswrapper[4860]: I1211 08:16:43.342349 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/5ef69e5e-c1ab-4d1b-9cf8-3ad23b9ff27f-registry-tls\") pod \"image-registry-66df7c8f76-555fn\" (UID: \"5ef69e5e-c1ab-4d1b-9cf8-3ad23b9ff27f\") " pod="openshift-image-registry/image-registry-66df7c8f76-555fn" Dec 11 08:16:43 crc kubenswrapper[4860]: I1211 08:16:43.343048 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/5ef69e5e-c1ab-4d1b-9cf8-3ad23b9ff27f-ca-trust-extracted\") pod \"image-registry-66df7c8f76-555fn\" (UID: \"5ef69e5e-c1ab-4d1b-9cf8-3ad23b9ff27f\") " pod="openshift-image-registry/image-registry-66df7c8f76-555fn" Dec 11 08:16:43 crc kubenswrapper[4860]: I1211 08:16:43.343627 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/5ef69e5e-c1ab-4d1b-9cf8-3ad23b9ff27f-registry-certificates\") pod \"image-registry-66df7c8f76-555fn\" (UID: \"5ef69e5e-c1ab-4d1b-9cf8-3ad23b9ff27f\") " pod="openshift-image-registry/image-registry-66df7c8f76-555fn" Dec 11 08:16:43 crc kubenswrapper[4860]: I1211 08:16:43.343754 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5ef69e5e-c1ab-4d1b-9cf8-3ad23b9ff27f-trusted-ca\") pod \"image-registry-66df7c8f76-555fn\" (UID: \"5ef69e5e-c1ab-4d1b-9cf8-3ad23b9ff27f\") " pod="openshift-image-registry/image-registry-66df7c8f76-555fn" Dec 11 08:16:43 crc kubenswrapper[4860]: I1211 08:16:43.348700 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/5ef69e5e-c1ab-4d1b-9cf8-3ad23b9ff27f-registry-tls\") pod \"image-registry-66df7c8f76-555fn\" (UID: \"5ef69e5e-c1ab-4d1b-9cf8-3ad23b9ff27f\") " pod="openshift-image-registry/image-registry-66df7c8f76-555fn" Dec 11 08:16:43 crc kubenswrapper[4860]: I1211 08:16:43.364770 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/5ef69e5e-c1ab-4d1b-9cf8-3ad23b9ff27f-installation-pull-secrets\") pod \"image-registry-66df7c8f76-555fn\" (UID: \"5ef69e5e-c1ab-4d1b-9cf8-3ad23b9ff27f\") " pod="openshift-image-registry/image-registry-66df7c8f76-555fn" Dec 11 08:16:43 crc kubenswrapper[4860]: I1211 08:16:43.368255 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nfddd\" (UniqueName: \"kubernetes.io/projected/5ef69e5e-c1ab-4d1b-9cf8-3ad23b9ff27f-kube-api-access-nfddd\") pod \"image-registry-66df7c8f76-555fn\" (UID: \"5ef69e5e-c1ab-4d1b-9cf8-3ad23b9ff27f\") " pod="openshift-image-registry/image-registry-66df7c8f76-555fn" Dec 11 08:16:43 crc kubenswrapper[4860]: I1211 08:16:43.374221 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5ef69e5e-c1ab-4d1b-9cf8-3ad23b9ff27f-bound-sa-token\") pod \"image-registry-66df7c8f76-555fn\" (UID: \"5ef69e5e-c1ab-4d1b-9cf8-3ad23b9ff27f\") " pod="openshift-image-registry/image-registry-66df7c8f76-555fn" Dec 11 08:16:43 crc kubenswrapper[4860]: I1211 08:16:43.418260 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-555fn" Dec 11 08:16:43 crc kubenswrapper[4860]: I1211 08:16:43.452145 4860 generic.go:334] "Generic (PLEG): container finished" podID="7d8a48c0-183c-4699-86d1-f6d0ef767cdf" containerID="1c77d8a7b4ccde5fc359da5ae850dde6040f66cb085c51fbbe2b4d591d5573fe" exitCode=0 Dec 11 08:16:43 crc kubenswrapper[4860]: I1211 08:16:43.452357 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7vj8m" event={"ID":"7d8a48c0-183c-4699-86d1-f6d0ef767cdf","Type":"ContainerDied","Data":"1c77d8a7b4ccde5fc359da5ae850dde6040f66cb085c51fbbe2b4d591d5573fe"} Dec 11 08:16:43 crc kubenswrapper[4860]: I1211 08:16:43.458588 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bg2t9" event={"ID":"79181f73-fb9e-40c5-96b8-0ad95545f1f7","Type":"ContainerStarted","Data":"9623fc5e78f8e31d06b3812864d694cf37d0eb7e455db2054a6a52af186c218d"} Dec 11 08:16:43 crc kubenswrapper[4860]: I1211 08:16:43.833023 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-555fn"] Dec 11 08:16:44 crc kubenswrapper[4860]: I1211 08:16:44.465697 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7vj8m" event={"ID":"7d8a48c0-183c-4699-86d1-f6d0ef767cdf","Type":"ContainerStarted","Data":"a83babeca0e6a995445d2112ff747c39bcc3ffec06ab9493e6fcef28d6ff0eb4"} Dec 11 08:16:44 crc kubenswrapper[4860]: I1211 08:16:44.469633 4860 generic.go:334] "Generic (PLEG): container finished" podID="79181f73-fb9e-40c5-96b8-0ad95545f1f7" containerID="9623fc5e78f8e31d06b3812864d694cf37d0eb7e455db2054a6a52af186c218d" exitCode=0 Dec 11 08:16:44 crc kubenswrapper[4860]: I1211 08:16:44.469754 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bg2t9" event={"ID":"79181f73-fb9e-40c5-96b8-0ad95545f1f7","Type":"ContainerDied","Data":"9623fc5e78f8e31d06b3812864d694cf37d0eb7e455db2054a6a52af186c218d"} Dec 11 08:16:44 crc kubenswrapper[4860]: I1211 08:16:44.471210 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-555fn" event={"ID":"5ef69e5e-c1ab-4d1b-9cf8-3ad23b9ff27f","Type":"ContainerStarted","Data":"19acf64581b8cf69b22998a8671715c54e1b6372b7fc89091cf9c02d25af0912"} Dec 11 08:16:44 crc kubenswrapper[4860]: I1211 08:16:44.471248 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-555fn" event={"ID":"5ef69e5e-c1ab-4d1b-9cf8-3ad23b9ff27f","Type":"ContainerStarted","Data":"50e6d372514a264c300359e7347868fdb01fed2dfac00708b376e30c4792c828"} Dec 11 08:16:44 crc kubenswrapper[4860]: I1211 08:16:44.471953 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-555fn" Dec 11 08:16:44 crc kubenswrapper[4860]: I1211 08:16:44.484542 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-7vj8m" podStartSLOduration=2.986228108 podStartE2EDuration="5.48451581s" podCreationTimestamp="2025-12-11 08:16:39 +0000 UTC" firstStartedPulling="2025-12-11 08:16:41.432205971 +0000 UTC m=+334.160725036" lastFinishedPulling="2025-12-11 08:16:43.930493683 +0000 UTC m=+336.659012738" observedRunningTime="2025-12-11 08:16:44.483230608 +0000 UTC m=+337.211749753" watchObservedRunningTime="2025-12-11 08:16:44.48451581 +0000 UTC m=+337.213034905" Dec 11 08:16:44 crc kubenswrapper[4860]: I1211 08:16:44.523944 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-555fn" podStartSLOduration=1.5239090499999999 podStartE2EDuration="1.52390905s" podCreationTimestamp="2025-12-11 08:16:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:16:44.517625213 +0000 UTC m=+337.246144288" watchObservedRunningTime="2025-12-11 08:16:44.52390905 +0000 UTC m=+337.252428165" Dec 11 08:16:45 crc kubenswrapper[4860]: I1211 08:16:45.478504 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bg2t9" event={"ID":"79181f73-fb9e-40c5-96b8-0ad95545f1f7","Type":"ContainerStarted","Data":"bc043d881d41517d2fcf32a42dfb5e277d761dc57c3aff2a38b0ad8f709c94ae"} Dec 11 08:16:45 crc kubenswrapper[4860]: I1211 08:16:45.498600 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-bg2t9" podStartSLOduration=1.858321091 podStartE2EDuration="4.498581406s" podCreationTimestamp="2025-12-11 08:16:41 +0000 UTC" firstStartedPulling="2025-12-11 08:16:42.442394639 +0000 UTC m=+335.170913694" lastFinishedPulling="2025-12-11 08:16:45.082654954 +0000 UTC m=+337.811174009" observedRunningTime="2025-12-11 08:16:45.494303515 +0000 UTC m=+338.222822580" watchObservedRunningTime="2025-12-11 08:16:45.498581406 +0000 UTC m=+338.227100461" Dec 11 08:16:47 crc kubenswrapper[4860]: I1211 08:16:47.918759 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-kzkbs" Dec 11 08:16:47 crc kubenswrapper[4860]: I1211 08:16:47.919100 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-kzkbs" Dec 11 08:16:47 crc kubenswrapper[4860]: I1211 08:16:47.967592 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-kzkbs" Dec 11 08:16:48 crc kubenswrapper[4860]: I1211 08:16:48.572706 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-kzkbs" Dec 11 08:16:48 crc kubenswrapper[4860]: I1211 08:16:48.944739 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-nkvw8" Dec 11 08:16:48 crc kubenswrapper[4860]: I1211 08:16:48.945432 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-nkvw8" Dec 11 08:16:48 crc kubenswrapper[4860]: I1211 08:16:48.989601 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-nkvw8" Dec 11 08:16:49 crc kubenswrapper[4860]: I1211 08:16:49.547657 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-nkvw8" Dec 11 08:16:50 crc kubenswrapper[4860]: I1211 08:16:50.331789 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-7vj8m" Dec 11 08:16:50 crc kubenswrapper[4860]: I1211 08:16:50.332835 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-7vj8m" Dec 11 08:16:50 crc kubenswrapper[4860]: I1211 08:16:50.378475 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-7vj8m" Dec 11 08:16:50 crc kubenswrapper[4860]: I1211 08:16:50.547243 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-7vj8m" Dec 11 08:16:51 crc kubenswrapper[4860]: I1211 08:16:51.535659 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-bg2t9" Dec 11 08:16:51 crc kubenswrapper[4860]: I1211 08:16:51.536051 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-bg2t9" Dec 11 08:16:51 crc kubenswrapper[4860]: I1211 08:16:51.590814 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-bg2t9" Dec 11 08:16:52 crc kubenswrapper[4860]: I1211 08:16:52.573065 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-bg2t9" Dec 11 08:16:54 crc kubenswrapper[4860]: I1211 08:16:54.753702 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-dz7kr" podUID="ccbceeb2-58ab-4409-82df-5c99189569b8" containerName="oauth-openshift" containerID="cri-o://da81c95fa79e88485cc7ff78c8483d365da07d5e70bf98b6b224c2b59dea5747" gracePeriod=15 Dec 11 08:16:56 crc kubenswrapper[4860]: I1211 08:16:56.171348 4860 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-dz7kr container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.21:6443/healthz\": dial tcp 10.217.0.21:6443: connect: connection refused" start-of-body= Dec 11 08:16:56 crc kubenswrapper[4860]: I1211 08:16:56.171484 4860 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-dz7kr" podUID="ccbceeb2-58ab-4409-82df-5c99189569b8" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.21:6443/healthz\": dial tcp 10.217.0.21:6443: connect: connection refused" Dec 11 08:16:57 crc kubenswrapper[4860]: I1211 08:16:57.559056 4860 generic.go:334] "Generic (PLEG): container finished" podID="ccbceeb2-58ab-4409-82df-5c99189569b8" containerID="da81c95fa79e88485cc7ff78c8483d365da07d5e70bf98b6b224c2b59dea5747" exitCode=0 Dec 11 08:16:57 crc kubenswrapper[4860]: I1211 08:16:57.559106 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-dz7kr" event={"ID":"ccbceeb2-58ab-4409-82df-5c99189569b8","Type":"ContainerDied","Data":"da81c95fa79e88485cc7ff78c8483d365da07d5e70bf98b6b224c2b59dea5747"} Dec 11 08:16:57 crc kubenswrapper[4860]: I1211 08:16:57.990586 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-dz7kr" Dec 11 08:16:57 crc kubenswrapper[4860]: I1211 08:16:57.992154 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ccbceeb2-58ab-4409-82df-5c99189569b8-audit-policies\") pod \"ccbceeb2-58ab-4409-82df-5c99189569b8\" (UID: \"ccbceeb2-58ab-4409-82df-5c99189569b8\") " Dec 11 08:16:57 crc kubenswrapper[4860]: I1211 08:16:57.992209 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/ccbceeb2-58ab-4409-82df-5c99189569b8-v4-0-config-system-cliconfig\") pod \"ccbceeb2-58ab-4409-82df-5c99189569b8\" (UID: \"ccbceeb2-58ab-4409-82df-5c99189569b8\") " Dec 11 08:16:57 crc kubenswrapper[4860]: I1211 08:16:57.992481 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/ccbceeb2-58ab-4409-82df-5c99189569b8-v4-0-config-user-template-error\") pod \"ccbceeb2-58ab-4409-82df-5c99189569b8\" (UID: \"ccbceeb2-58ab-4409-82df-5c99189569b8\") " Dec 11 08:16:57 crc kubenswrapper[4860]: I1211 08:16:57.992514 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/ccbceeb2-58ab-4409-82df-5c99189569b8-v4-0-config-user-idp-0-file-data\") pod \"ccbceeb2-58ab-4409-82df-5c99189569b8\" (UID: \"ccbceeb2-58ab-4409-82df-5c99189569b8\") " Dec 11 08:16:57 crc kubenswrapper[4860]: I1211 08:16:57.992936 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ccbceeb2-58ab-4409-82df-5c99189569b8-v4-0-config-system-trusted-ca-bundle\") pod \"ccbceeb2-58ab-4409-82df-5c99189569b8\" (UID: \"ccbceeb2-58ab-4409-82df-5c99189569b8\") " Dec 11 08:16:57 crc kubenswrapper[4860]: I1211 08:16:57.992979 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ccbceeb2-58ab-4409-82df-5c99189569b8-audit-dir\") pod \"ccbceeb2-58ab-4409-82df-5c99189569b8\" (UID: \"ccbceeb2-58ab-4409-82df-5c99189569b8\") " Dec 11 08:16:57 crc kubenswrapper[4860]: I1211 08:16:57.993031 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/ccbceeb2-58ab-4409-82df-5c99189569b8-v4-0-config-user-template-provider-selection\") pod \"ccbceeb2-58ab-4409-82df-5c99189569b8\" (UID: \"ccbceeb2-58ab-4409-82df-5c99189569b8\") " Dec 11 08:16:57 crc kubenswrapper[4860]: I1211 08:16:57.993038 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ccbceeb2-58ab-4409-82df-5c99189569b8-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "ccbceeb2-58ab-4409-82df-5c99189569b8" (UID: "ccbceeb2-58ab-4409-82df-5c99189569b8"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:16:57 crc kubenswrapper[4860]: I1211 08:16:57.993064 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/ccbceeb2-58ab-4409-82df-5c99189569b8-v4-0-config-system-ocp-branding-template\") pod \"ccbceeb2-58ab-4409-82df-5c99189569b8\" (UID: \"ccbceeb2-58ab-4409-82df-5c99189569b8\") " Dec 11 08:16:57 crc kubenswrapper[4860]: I1211 08:16:57.993090 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/ccbceeb2-58ab-4409-82df-5c99189569b8-v4-0-config-system-serving-cert\") pod \"ccbceeb2-58ab-4409-82df-5c99189569b8\" (UID: \"ccbceeb2-58ab-4409-82df-5c99189569b8\") " Dec 11 08:16:57 crc kubenswrapper[4860]: I1211 08:16:57.993113 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/ccbceeb2-58ab-4409-82df-5c99189569b8-v4-0-config-user-template-login\") pod \"ccbceeb2-58ab-4409-82df-5c99189569b8\" (UID: \"ccbceeb2-58ab-4409-82df-5c99189569b8\") " Dec 11 08:16:57 crc kubenswrapper[4860]: I1211 08:16:57.993137 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/ccbceeb2-58ab-4409-82df-5c99189569b8-v4-0-config-system-session\") pod \"ccbceeb2-58ab-4409-82df-5c99189569b8\" (UID: \"ccbceeb2-58ab-4409-82df-5c99189569b8\") " Dec 11 08:16:57 crc kubenswrapper[4860]: I1211 08:16:57.993139 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ccbceeb2-58ab-4409-82df-5c99189569b8-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "ccbceeb2-58ab-4409-82df-5c99189569b8" (UID: "ccbceeb2-58ab-4409-82df-5c99189569b8"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 08:16:57 crc kubenswrapper[4860]: I1211 08:16:57.993168 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/ccbceeb2-58ab-4409-82df-5c99189569b8-v4-0-config-system-service-ca\") pod \"ccbceeb2-58ab-4409-82df-5c99189569b8\" (UID: \"ccbceeb2-58ab-4409-82df-5c99189569b8\") " Dec 11 08:16:57 crc kubenswrapper[4860]: I1211 08:16:57.993273 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8k4gt\" (UniqueName: \"kubernetes.io/projected/ccbceeb2-58ab-4409-82df-5c99189569b8-kube-api-access-8k4gt\") pod \"ccbceeb2-58ab-4409-82df-5c99189569b8\" (UID: \"ccbceeb2-58ab-4409-82df-5c99189569b8\") " Dec 11 08:16:57 crc kubenswrapper[4860]: I1211 08:16:57.993326 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/ccbceeb2-58ab-4409-82df-5c99189569b8-v4-0-config-system-router-certs\") pod \"ccbceeb2-58ab-4409-82df-5c99189569b8\" (UID: \"ccbceeb2-58ab-4409-82df-5c99189569b8\") " Dec 11 08:16:57 crc kubenswrapper[4860]: I1211 08:16:57.993766 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ccbceeb2-58ab-4409-82df-5c99189569b8-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "ccbceeb2-58ab-4409-82df-5c99189569b8" (UID: "ccbceeb2-58ab-4409-82df-5c99189569b8"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:16:57 crc kubenswrapper[4860]: I1211 08:16:57.993859 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ccbceeb2-58ab-4409-82df-5c99189569b8-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "ccbceeb2-58ab-4409-82df-5c99189569b8" (UID: "ccbceeb2-58ab-4409-82df-5c99189569b8"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:16:57 crc kubenswrapper[4860]: I1211 08:16:57.993918 4860 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ccbceeb2-58ab-4409-82df-5c99189569b8-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 11 08:16:57 crc kubenswrapper[4860]: I1211 08:16:57.993940 4860 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/ccbceeb2-58ab-4409-82df-5c99189569b8-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 11 08:16:57 crc kubenswrapper[4860]: I1211 08:16:57.993954 4860 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ccbceeb2-58ab-4409-82df-5c99189569b8-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 11 08:16:57 crc kubenswrapper[4860]: I1211 08:16:57.995013 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ccbceeb2-58ab-4409-82df-5c99189569b8-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "ccbceeb2-58ab-4409-82df-5c99189569b8" (UID: "ccbceeb2-58ab-4409-82df-5c99189569b8"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:16:57 crc kubenswrapper[4860]: I1211 08:16:57.998179 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ccbceeb2-58ab-4409-82df-5c99189569b8-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "ccbceeb2-58ab-4409-82df-5c99189569b8" (UID: "ccbceeb2-58ab-4409-82df-5c99189569b8"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:16:57 crc kubenswrapper[4860]: I1211 08:16:57.998712 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ccbceeb2-58ab-4409-82df-5c99189569b8-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "ccbceeb2-58ab-4409-82df-5c99189569b8" (UID: "ccbceeb2-58ab-4409-82df-5c99189569b8"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:16:57 crc kubenswrapper[4860]: I1211 08:16:57.998749 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ccbceeb2-58ab-4409-82df-5c99189569b8-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "ccbceeb2-58ab-4409-82df-5c99189569b8" (UID: "ccbceeb2-58ab-4409-82df-5c99189569b8"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:16:57 crc kubenswrapper[4860]: I1211 08:16:57.999225 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ccbceeb2-58ab-4409-82df-5c99189569b8-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "ccbceeb2-58ab-4409-82df-5c99189569b8" (UID: "ccbceeb2-58ab-4409-82df-5c99189569b8"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:16:57 crc kubenswrapper[4860]: I1211 08:16:57.999797 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ccbceeb2-58ab-4409-82df-5c99189569b8-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "ccbceeb2-58ab-4409-82df-5c99189569b8" (UID: "ccbceeb2-58ab-4409-82df-5c99189569b8"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:16:58 crc kubenswrapper[4860]: I1211 08:16:58.000005 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ccbceeb2-58ab-4409-82df-5c99189569b8-kube-api-access-8k4gt" (OuterVolumeSpecName: "kube-api-access-8k4gt") pod "ccbceeb2-58ab-4409-82df-5c99189569b8" (UID: "ccbceeb2-58ab-4409-82df-5c99189569b8"). InnerVolumeSpecName "kube-api-access-8k4gt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:16:58 crc kubenswrapper[4860]: I1211 08:16:58.000249 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ccbceeb2-58ab-4409-82df-5c99189569b8-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "ccbceeb2-58ab-4409-82df-5c99189569b8" (UID: "ccbceeb2-58ab-4409-82df-5c99189569b8"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:16:58 crc kubenswrapper[4860]: I1211 08:16:58.001116 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ccbceeb2-58ab-4409-82df-5c99189569b8-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "ccbceeb2-58ab-4409-82df-5c99189569b8" (UID: "ccbceeb2-58ab-4409-82df-5c99189569b8"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:16:58 crc kubenswrapper[4860]: I1211 08:16:58.002027 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ccbceeb2-58ab-4409-82df-5c99189569b8-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "ccbceeb2-58ab-4409-82df-5c99189569b8" (UID: "ccbceeb2-58ab-4409-82df-5c99189569b8"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:16:58 crc kubenswrapper[4860]: I1211 08:16:58.028240 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-66dbbf5c6f-x2xtm"] Dec 11 08:16:58 crc kubenswrapper[4860]: E1211 08:16:58.028795 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccbceeb2-58ab-4409-82df-5c99189569b8" containerName="oauth-openshift" Dec 11 08:16:58 crc kubenswrapper[4860]: I1211 08:16:58.028832 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccbceeb2-58ab-4409-82df-5c99189569b8" containerName="oauth-openshift" Dec 11 08:16:58 crc kubenswrapper[4860]: I1211 08:16:58.029133 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="ccbceeb2-58ab-4409-82df-5c99189569b8" containerName="oauth-openshift" Dec 11 08:16:58 crc kubenswrapper[4860]: I1211 08:16:58.029829 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-66dbbf5c6f-x2xtm" Dec 11 08:16:58 crc kubenswrapper[4860]: I1211 08:16:58.039525 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-66dbbf5c6f-x2xtm"] Dec 11 08:16:58 crc kubenswrapper[4860]: I1211 08:16:58.095039 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/4a4ee064-015a-46b3-890e-9028f4205905-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-66dbbf5c6f-x2xtm\" (UID: \"4a4ee064-015a-46b3-890e-9028f4205905\") " pod="openshift-authentication/oauth-openshift-66dbbf5c6f-x2xtm" Dec 11 08:16:58 crc kubenswrapper[4860]: I1211 08:16:58.095112 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/4a4ee064-015a-46b3-890e-9028f4205905-v4-0-config-system-service-ca\") pod \"oauth-openshift-66dbbf5c6f-x2xtm\" (UID: \"4a4ee064-015a-46b3-890e-9028f4205905\") " pod="openshift-authentication/oauth-openshift-66dbbf5c6f-x2xtm" Dec 11 08:16:58 crc kubenswrapper[4860]: I1211 08:16:58.095143 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/4a4ee064-015a-46b3-890e-9028f4205905-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-66dbbf5c6f-x2xtm\" (UID: \"4a4ee064-015a-46b3-890e-9028f4205905\") " pod="openshift-authentication/oauth-openshift-66dbbf5c6f-x2xtm" Dec 11 08:16:58 crc kubenswrapper[4860]: I1211 08:16:58.095190 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4a4ee064-015a-46b3-890e-9028f4205905-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-66dbbf5c6f-x2xtm\" (UID: \"4a4ee064-015a-46b3-890e-9028f4205905\") " pod="openshift-authentication/oauth-openshift-66dbbf5c6f-x2xtm" Dec 11 08:16:58 crc kubenswrapper[4860]: I1211 08:16:58.095226 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/4a4ee064-015a-46b3-890e-9028f4205905-v4-0-config-system-cliconfig\") pod \"oauth-openshift-66dbbf5c6f-x2xtm\" (UID: \"4a4ee064-015a-46b3-890e-9028f4205905\") " pod="openshift-authentication/oauth-openshift-66dbbf5c6f-x2xtm" Dec 11 08:16:58 crc kubenswrapper[4860]: I1211 08:16:58.095269 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/4a4ee064-015a-46b3-890e-9028f4205905-audit-policies\") pod \"oauth-openshift-66dbbf5c6f-x2xtm\" (UID: \"4a4ee064-015a-46b3-890e-9028f4205905\") " pod="openshift-authentication/oauth-openshift-66dbbf5c6f-x2xtm" Dec 11 08:16:58 crc kubenswrapper[4860]: I1211 08:16:58.095329 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/4a4ee064-015a-46b3-890e-9028f4205905-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-66dbbf5c6f-x2xtm\" (UID: \"4a4ee064-015a-46b3-890e-9028f4205905\") " pod="openshift-authentication/oauth-openshift-66dbbf5c6f-x2xtm" Dec 11 08:16:58 crc kubenswrapper[4860]: I1211 08:16:58.095391 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/4a4ee064-015a-46b3-890e-9028f4205905-v4-0-config-user-template-login\") pod \"oauth-openshift-66dbbf5c6f-x2xtm\" (UID: \"4a4ee064-015a-46b3-890e-9028f4205905\") " pod="openshift-authentication/oauth-openshift-66dbbf5c6f-x2xtm" Dec 11 08:16:58 crc kubenswrapper[4860]: I1211 08:16:58.095426 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/4a4ee064-015a-46b3-890e-9028f4205905-v4-0-config-user-template-error\") pod \"oauth-openshift-66dbbf5c6f-x2xtm\" (UID: \"4a4ee064-015a-46b3-890e-9028f4205905\") " pod="openshift-authentication/oauth-openshift-66dbbf5c6f-x2xtm" Dec 11 08:16:58 crc kubenswrapper[4860]: I1211 08:16:58.095458 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9m7vs\" (UniqueName: \"kubernetes.io/projected/4a4ee064-015a-46b3-890e-9028f4205905-kube-api-access-9m7vs\") pod \"oauth-openshift-66dbbf5c6f-x2xtm\" (UID: \"4a4ee064-015a-46b3-890e-9028f4205905\") " pod="openshift-authentication/oauth-openshift-66dbbf5c6f-x2xtm" Dec 11 08:16:58 crc kubenswrapper[4860]: I1211 08:16:58.095563 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/4a4ee064-015a-46b3-890e-9028f4205905-v4-0-config-system-serving-cert\") pod \"oauth-openshift-66dbbf5c6f-x2xtm\" (UID: \"4a4ee064-015a-46b3-890e-9028f4205905\") " pod="openshift-authentication/oauth-openshift-66dbbf5c6f-x2xtm" Dec 11 08:16:58 crc kubenswrapper[4860]: I1211 08:16:58.095706 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/4a4ee064-015a-46b3-890e-9028f4205905-v4-0-config-system-router-certs\") pod \"oauth-openshift-66dbbf5c6f-x2xtm\" (UID: \"4a4ee064-015a-46b3-890e-9028f4205905\") " pod="openshift-authentication/oauth-openshift-66dbbf5c6f-x2xtm" Dec 11 08:16:58 crc kubenswrapper[4860]: I1211 08:16:58.095789 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/4a4ee064-015a-46b3-890e-9028f4205905-audit-dir\") pod \"oauth-openshift-66dbbf5c6f-x2xtm\" (UID: \"4a4ee064-015a-46b3-890e-9028f4205905\") " pod="openshift-authentication/oauth-openshift-66dbbf5c6f-x2xtm" Dec 11 08:16:58 crc kubenswrapper[4860]: I1211 08:16:58.095870 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/4a4ee064-015a-46b3-890e-9028f4205905-v4-0-config-system-session\") pod \"oauth-openshift-66dbbf5c6f-x2xtm\" (UID: \"4a4ee064-015a-46b3-890e-9028f4205905\") " pod="openshift-authentication/oauth-openshift-66dbbf5c6f-x2xtm" Dec 11 08:16:58 crc kubenswrapper[4860]: I1211 08:16:58.096131 4860 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/ccbceeb2-58ab-4409-82df-5c99189569b8-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 11 08:16:58 crc kubenswrapper[4860]: I1211 08:16:58.096160 4860 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/ccbceeb2-58ab-4409-82df-5c99189569b8-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 11 08:16:58 crc kubenswrapper[4860]: I1211 08:16:58.096176 4860 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/ccbceeb2-58ab-4409-82df-5c99189569b8-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 08:16:58 crc kubenswrapper[4860]: I1211 08:16:58.096193 4860 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/ccbceeb2-58ab-4409-82df-5c99189569b8-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 11 08:16:58 crc kubenswrapper[4860]: I1211 08:16:58.096206 4860 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/ccbceeb2-58ab-4409-82df-5c99189569b8-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 11 08:16:58 crc kubenswrapper[4860]: I1211 08:16:58.096223 4860 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/ccbceeb2-58ab-4409-82df-5c99189569b8-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 11 08:16:58 crc kubenswrapper[4860]: I1211 08:16:58.096237 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8k4gt\" (UniqueName: \"kubernetes.io/projected/ccbceeb2-58ab-4409-82df-5c99189569b8-kube-api-access-8k4gt\") on node \"crc\" DevicePath \"\"" Dec 11 08:16:58 crc kubenswrapper[4860]: I1211 08:16:58.096251 4860 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/ccbceeb2-58ab-4409-82df-5c99189569b8-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 11 08:16:58 crc kubenswrapper[4860]: I1211 08:16:58.096264 4860 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/ccbceeb2-58ab-4409-82df-5c99189569b8-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 11 08:16:58 crc kubenswrapper[4860]: I1211 08:16:58.096279 4860 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/ccbceeb2-58ab-4409-82df-5c99189569b8-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 11 08:16:58 crc kubenswrapper[4860]: I1211 08:16:58.096294 4860 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ccbceeb2-58ab-4409-82df-5c99189569b8-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 08:16:58 crc kubenswrapper[4860]: I1211 08:16:58.197348 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/4a4ee064-015a-46b3-890e-9028f4205905-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-66dbbf5c6f-x2xtm\" (UID: \"4a4ee064-015a-46b3-890e-9028f4205905\") " pod="openshift-authentication/oauth-openshift-66dbbf5c6f-x2xtm" Dec 11 08:16:58 crc kubenswrapper[4860]: I1211 08:16:58.197436 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/4a4ee064-015a-46b3-890e-9028f4205905-v4-0-config-system-service-ca\") pod \"oauth-openshift-66dbbf5c6f-x2xtm\" (UID: \"4a4ee064-015a-46b3-890e-9028f4205905\") " pod="openshift-authentication/oauth-openshift-66dbbf5c6f-x2xtm" Dec 11 08:16:58 crc kubenswrapper[4860]: I1211 08:16:58.197470 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/4a4ee064-015a-46b3-890e-9028f4205905-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-66dbbf5c6f-x2xtm\" (UID: \"4a4ee064-015a-46b3-890e-9028f4205905\") " pod="openshift-authentication/oauth-openshift-66dbbf5c6f-x2xtm" Dec 11 08:16:58 crc kubenswrapper[4860]: I1211 08:16:58.197516 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4a4ee064-015a-46b3-890e-9028f4205905-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-66dbbf5c6f-x2xtm\" (UID: \"4a4ee064-015a-46b3-890e-9028f4205905\") " pod="openshift-authentication/oauth-openshift-66dbbf5c6f-x2xtm" Dec 11 08:16:58 crc kubenswrapper[4860]: I1211 08:16:58.197551 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/4a4ee064-015a-46b3-890e-9028f4205905-v4-0-config-system-cliconfig\") pod \"oauth-openshift-66dbbf5c6f-x2xtm\" (UID: \"4a4ee064-015a-46b3-890e-9028f4205905\") " pod="openshift-authentication/oauth-openshift-66dbbf5c6f-x2xtm" Dec 11 08:16:58 crc kubenswrapper[4860]: I1211 08:16:58.197587 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/4a4ee064-015a-46b3-890e-9028f4205905-audit-policies\") pod \"oauth-openshift-66dbbf5c6f-x2xtm\" (UID: \"4a4ee064-015a-46b3-890e-9028f4205905\") " pod="openshift-authentication/oauth-openshift-66dbbf5c6f-x2xtm" Dec 11 08:16:58 crc kubenswrapper[4860]: I1211 08:16:58.197680 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/4a4ee064-015a-46b3-890e-9028f4205905-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-66dbbf5c6f-x2xtm\" (UID: \"4a4ee064-015a-46b3-890e-9028f4205905\") " pod="openshift-authentication/oauth-openshift-66dbbf5c6f-x2xtm" Dec 11 08:16:58 crc kubenswrapper[4860]: I1211 08:16:58.197804 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/4a4ee064-015a-46b3-890e-9028f4205905-v4-0-config-user-template-login\") pod \"oauth-openshift-66dbbf5c6f-x2xtm\" (UID: \"4a4ee064-015a-46b3-890e-9028f4205905\") " pod="openshift-authentication/oauth-openshift-66dbbf5c6f-x2xtm" Dec 11 08:16:58 crc kubenswrapper[4860]: I1211 08:16:58.197862 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/4a4ee064-015a-46b3-890e-9028f4205905-v4-0-config-user-template-error\") pod \"oauth-openshift-66dbbf5c6f-x2xtm\" (UID: \"4a4ee064-015a-46b3-890e-9028f4205905\") " pod="openshift-authentication/oauth-openshift-66dbbf5c6f-x2xtm" Dec 11 08:16:58 crc kubenswrapper[4860]: I1211 08:16:58.197977 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9m7vs\" (UniqueName: \"kubernetes.io/projected/4a4ee064-015a-46b3-890e-9028f4205905-kube-api-access-9m7vs\") pod \"oauth-openshift-66dbbf5c6f-x2xtm\" (UID: \"4a4ee064-015a-46b3-890e-9028f4205905\") " pod="openshift-authentication/oauth-openshift-66dbbf5c6f-x2xtm" Dec 11 08:16:58 crc kubenswrapper[4860]: I1211 08:16:58.198079 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/4a4ee064-015a-46b3-890e-9028f4205905-v4-0-config-system-serving-cert\") pod \"oauth-openshift-66dbbf5c6f-x2xtm\" (UID: \"4a4ee064-015a-46b3-890e-9028f4205905\") " pod="openshift-authentication/oauth-openshift-66dbbf5c6f-x2xtm" Dec 11 08:16:58 crc kubenswrapper[4860]: I1211 08:16:58.198177 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/4a4ee064-015a-46b3-890e-9028f4205905-v4-0-config-system-router-certs\") pod \"oauth-openshift-66dbbf5c6f-x2xtm\" (UID: \"4a4ee064-015a-46b3-890e-9028f4205905\") " pod="openshift-authentication/oauth-openshift-66dbbf5c6f-x2xtm" Dec 11 08:16:58 crc kubenswrapper[4860]: I1211 08:16:58.198395 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/4a4ee064-015a-46b3-890e-9028f4205905-audit-dir\") pod \"oauth-openshift-66dbbf5c6f-x2xtm\" (UID: \"4a4ee064-015a-46b3-890e-9028f4205905\") " pod="openshift-authentication/oauth-openshift-66dbbf5c6f-x2xtm" Dec 11 08:16:58 crc kubenswrapper[4860]: I1211 08:16:58.198508 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/4a4ee064-015a-46b3-890e-9028f4205905-v4-0-config-system-session\") pod \"oauth-openshift-66dbbf5c6f-x2xtm\" (UID: \"4a4ee064-015a-46b3-890e-9028f4205905\") " pod="openshift-authentication/oauth-openshift-66dbbf5c6f-x2xtm" Dec 11 08:16:58 crc kubenswrapper[4860]: I1211 08:16:58.198933 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/4a4ee064-015a-46b3-890e-9028f4205905-v4-0-config-system-cliconfig\") pod \"oauth-openshift-66dbbf5c6f-x2xtm\" (UID: \"4a4ee064-015a-46b3-890e-9028f4205905\") " pod="openshift-authentication/oauth-openshift-66dbbf5c6f-x2xtm" Dec 11 08:16:58 crc kubenswrapper[4860]: I1211 08:16:58.199052 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/4a4ee064-015a-46b3-890e-9028f4205905-audit-dir\") pod \"oauth-openshift-66dbbf5c6f-x2xtm\" (UID: \"4a4ee064-015a-46b3-890e-9028f4205905\") " pod="openshift-authentication/oauth-openshift-66dbbf5c6f-x2xtm" Dec 11 08:16:58 crc kubenswrapper[4860]: I1211 08:16:58.199559 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/4a4ee064-015a-46b3-890e-9028f4205905-v4-0-config-system-service-ca\") pod \"oauth-openshift-66dbbf5c6f-x2xtm\" (UID: \"4a4ee064-015a-46b3-890e-9028f4205905\") " pod="openshift-authentication/oauth-openshift-66dbbf5c6f-x2xtm" Dec 11 08:16:58 crc kubenswrapper[4860]: I1211 08:16:58.200673 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4a4ee064-015a-46b3-890e-9028f4205905-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-66dbbf5c6f-x2xtm\" (UID: \"4a4ee064-015a-46b3-890e-9028f4205905\") " pod="openshift-authentication/oauth-openshift-66dbbf5c6f-x2xtm" Dec 11 08:16:58 crc kubenswrapper[4860]: I1211 08:16:58.200890 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/4a4ee064-015a-46b3-890e-9028f4205905-audit-policies\") pod \"oauth-openshift-66dbbf5c6f-x2xtm\" (UID: \"4a4ee064-015a-46b3-890e-9028f4205905\") " pod="openshift-authentication/oauth-openshift-66dbbf5c6f-x2xtm" Dec 11 08:16:58 crc kubenswrapper[4860]: I1211 08:16:58.201443 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/4a4ee064-015a-46b3-890e-9028f4205905-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-66dbbf5c6f-x2xtm\" (UID: \"4a4ee064-015a-46b3-890e-9028f4205905\") " pod="openshift-authentication/oauth-openshift-66dbbf5c6f-x2xtm" Dec 11 08:16:58 crc kubenswrapper[4860]: I1211 08:16:58.201663 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/4a4ee064-015a-46b3-890e-9028f4205905-v4-0-config-user-template-login\") pod \"oauth-openshift-66dbbf5c6f-x2xtm\" (UID: \"4a4ee064-015a-46b3-890e-9028f4205905\") " pod="openshift-authentication/oauth-openshift-66dbbf5c6f-x2xtm" Dec 11 08:16:58 crc kubenswrapper[4860]: I1211 08:16:58.201940 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/4a4ee064-015a-46b3-890e-9028f4205905-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-66dbbf5c6f-x2xtm\" (UID: \"4a4ee064-015a-46b3-890e-9028f4205905\") " pod="openshift-authentication/oauth-openshift-66dbbf5c6f-x2xtm" Dec 11 08:16:58 crc kubenswrapper[4860]: I1211 08:16:58.204916 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/4a4ee064-015a-46b3-890e-9028f4205905-v4-0-config-system-router-certs\") pod \"oauth-openshift-66dbbf5c6f-x2xtm\" (UID: \"4a4ee064-015a-46b3-890e-9028f4205905\") " pod="openshift-authentication/oauth-openshift-66dbbf5c6f-x2xtm" Dec 11 08:16:58 crc kubenswrapper[4860]: I1211 08:16:58.206993 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/4a4ee064-015a-46b3-890e-9028f4205905-v4-0-config-system-session\") pod \"oauth-openshift-66dbbf5c6f-x2xtm\" (UID: \"4a4ee064-015a-46b3-890e-9028f4205905\") " pod="openshift-authentication/oauth-openshift-66dbbf5c6f-x2xtm" Dec 11 08:16:58 crc kubenswrapper[4860]: I1211 08:16:58.207006 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/4a4ee064-015a-46b3-890e-9028f4205905-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-66dbbf5c6f-x2xtm\" (UID: \"4a4ee064-015a-46b3-890e-9028f4205905\") " pod="openshift-authentication/oauth-openshift-66dbbf5c6f-x2xtm" Dec 11 08:16:58 crc kubenswrapper[4860]: I1211 08:16:58.207565 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/4a4ee064-015a-46b3-890e-9028f4205905-v4-0-config-user-template-error\") pod \"oauth-openshift-66dbbf5c6f-x2xtm\" (UID: \"4a4ee064-015a-46b3-890e-9028f4205905\") " pod="openshift-authentication/oauth-openshift-66dbbf5c6f-x2xtm" Dec 11 08:16:58 crc kubenswrapper[4860]: I1211 08:16:58.208472 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/4a4ee064-015a-46b3-890e-9028f4205905-v4-0-config-system-serving-cert\") pod \"oauth-openshift-66dbbf5c6f-x2xtm\" (UID: \"4a4ee064-015a-46b3-890e-9028f4205905\") " pod="openshift-authentication/oauth-openshift-66dbbf5c6f-x2xtm" Dec 11 08:16:58 crc kubenswrapper[4860]: I1211 08:16:58.221232 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9m7vs\" (UniqueName: \"kubernetes.io/projected/4a4ee064-015a-46b3-890e-9028f4205905-kube-api-access-9m7vs\") pod \"oauth-openshift-66dbbf5c6f-x2xtm\" (UID: \"4a4ee064-015a-46b3-890e-9028f4205905\") " pod="openshift-authentication/oauth-openshift-66dbbf5c6f-x2xtm" Dec 11 08:16:58 crc kubenswrapper[4860]: I1211 08:16:58.366991 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-66dbbf5c6f-x2xtm" Dec 11 08:16:58 crc kubenswrapper[4860]: I1211 08:16:58.569755 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-dz7kr" event={"ID":"ccbceeb2-58ab-4409-82df-5c99189569b8","Type":"ContainerDied","Data":"905c393d94be1cc4f1936a9844ee4416c64fad01ddc896a25f003c6e8777386b"} Dec 11 08:16:58 crc kubenswrapper[4860]: I1211 08:16:58.569824 4860 scope.go:117] "RemoveContainer" containerID="da81c95fa79e88485cc7ff78c8483d365da07d5e70bf98b6b224c2b59dea5747" Dec 11 08:16:58 crc kubenswrapper[4860]: I1211 08:16:58.569902 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-dz7kr" Dec 11 08:16:58 crc kubenswrapper[4860]: I1211 08:16:58.605329 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-dz7kr"] Dec 11 08:16:58 crc kubenswrapper[4860]: I1211 08:16:58.608498 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-dz7kr"] Dec 11 08:16:58 crc kubenswrapper[4860]: I1211 08:16:58.823205 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-66dbbf5c6f-x2xtm"] Dec 11 08:16:59 crc kubenswrapper[4860]: I1211 08:16:59.576930 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-66dbbf5c6f-x2xtm" event={"ID":"4a4ee064-015a-46b3-890e-9028f4205905","Type":"ContainerStarted","Data":"1394bd04a58c943d1376ca49c9d4b4f114986103344900416400de28d662269b"} Dec 11 08:16:59 crc kubenswrapper[4860]: I1211 08:16:59.577613 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-66dbbf5c6f-x2xtm" Dec 11 08:16:59 crc kubenswrapper[4860]: I1211 08:16:59.577632 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-66dbbf5c6f-x2xtm" event={"ID":"4a4ee064-015a-46b3-890e-9028f4205905","Type":"ContainerStarted","Data":"89a6f4a28f4dcd9425008936ab2c591f9308c86aeb66dd8db018ee0f062c7745"} Dec 11 08:16:59 crc kubenswrapper[4860]: I1211 08:16:59.586554 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ccbceeb2-58ab-4409-82df-5c99189569b8" path="/var/lib/kubelet/pods/ccbceeb2-58ab-4409-82df-5c99189569b8/volumes" Dec 11 08:16:59 crc kubenswrapper[4860]: I1211 08:16:59.595354 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-66dbbf5c6f-x2xtm" podStartSLOduration=30.595332104 podStartE2EDuration="30.595332104s" podCreationTimestamp="2025-12-11 08:16:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:16:59.592726488 +0000 UTC m=+352.321245583" watchObservedRunningTime="2025-12-11 08:16:59.595332104 +0000 UTC m=+352.323851149" Dec 11 08:16:59 crc kubenswrapper[4860]: I1211 08:16:59.826388 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-66dbbf5c6f-x2xtm" Dec 11 08:17:03 crc kubenswrapper[4860]: I1211 08:17:03.422258 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-555fn" Dec 11 08:17:03 crc kubenswrapper[4860]: I1211 08:17:03.463990 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-zf572"] Dec 11 08:17:08 crc kubenswrapper[4860]: I1211 08:17:08.795340 4860 patch_prober.go:28] interesting pod/machine-config-daemon-99qgp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 08:17:08 crc kubenswrapper[4860]: I1211 08:17:08.795896 4860 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 08:17:28 crc kubenswrapper[4860]: I1211 08:17:28.520395 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-zf572" podUID="38005ee2-6635-498c-b156-ac57b4f2f4d9" containerName="registry" containerID="cri-o://027d8ed039343e3ce7d35642642a40ab245c222261386a8d009a083692ad1f2b" gracePeriod=30 Dec 11 08:17:28 crc kubenswrapper[4860]: I1211 08:17:28.768266 4860 generic.go:334] "Generic (PLEG): container finished" podID="38005ee2-6635-498c-b156-ac57b4f2f4d9" containerID="027d8ed039343e3ce7d35642642a40ab245c222261386a8d009a083692ad1f2b" exitCode=0 Dec 11 08:17:28 crc kubenswrapper[4860]: I1211 08:17:28.768559 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-zf572" event={"ID":"38005ee2-6635-498c-b156-ac57b4f2f4d9","Type":"ContainerDied","Data":"027d8ed039343e3ce7d35642642a40ab245c222261386a8d009a083692ad1f2b"} Dec 11 08:17:29 crc kubenswrapper[4860]: I1211 08:17:29.466471 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-zf572" Dec 11 08:17:29 crc kubenswrapper[4860]: I1211 08:17:29.484057 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/38005ee2-6635-498c-b156-ac57b4f2f4d9-trusted-ca\") pod \"38005ee2-6635-498c-b156-ac57b4f2f4d9\" (UID: \"38005ee2-6635-498c-b156-ac57b4f2f4d9\") " Dec 11 08:17:29 crc kubenswrapper[4860]: I1211 08:17:29.484094 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/38005ee2-6635-498c-b156-ac57b4f2f4d9-registry-certificates\") pod \"38005ee2-6635-498c-b156-ac57b4f2f4d9\" (UID: \"38005ee2-6635-498c-b156-ac57b4f2f4d9\") " Dec 11 08:17:29 crc kubenswrapper[4860]: I1211 08:17:29.485481 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/38005ee2-6635-498c-b156-ac57b4f2f4d9-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "38005ee2-6635-498c-b156-ac57b4f2f4d9" (UID: "38005ee2-6635-498c-b156-ac57b4f2f4d9"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:17:29 crc kubenswrapper[4860]: I1211 08:17:29.485609 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/38005ee2-6635-498c-b156-ac57b4f2f4d9-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "38005ee2-6635-498c-b156-ac57b4f2f4d9" (UID: "38005ee2-6635-498c-b156-ac57b4f2f4d9"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:17:29 crc kubenswrapper[4860]: I1211 08:17:29.584807 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/38005ee2-6635-498c-b156-ac57b4f2f4d9-bound-sa-token\") pod \"38005ee2-6635-498c-b156-ac57b4f2f4d9\" (UID: \"38005ee2-6635-498c-b156-ac57b4f2f4d9\") " Dec 11 08:17:29 crc kubenswrapper[4860]: I1211 08:17:29.585379 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"38005ee2-6635-498c-b156-ac57b4f2f4d9\" (UID: \"38005ee2-6635-498c-b156-ac57b4f2f4d9\") " Dec 11 08:17:29 crc kubenswrapper[4860]: I1211 08:17:29.585443 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/38005ee2-6635-498c-b156-ac57b4f2f4d9-ca-trust-extracted\") pod \"38005ee2-6635-498c-b156-ac57b4f2f4d9\" (UID: \"38005ee2-6635-498c-b156-ac57b4f2f4d9\") " Dec 11 08:17:29 crc kubenswrapper[4860]: I1211 08:17:29.585489 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/38005ee2-6635-498c-b156-ac57b4f2f4d9-installation-pull-secrets\") pod \"38005ee2-6635-498c-b156-ac57b4f2f4d9\" (UID: \"38005ee2-6635-498c-b156-ac57b4f2f4d9\") " Dec 11 08:17:29 crc kubenswrapper[4860]: I1211 08:17:29.585516 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/38005ee2-6635-498c-b156-ac57b4f2f4d9-registry-tls\") pod \"38005ee2-6635-498c-b156-ac57b4f2f4d9\" (UID: \"38005ee2-6635-498c-b156-ac57b4f2f4d9\") " Dec 11 08:17:29 crc kubenswrapper[4860]: I1211 08:17:29.585572 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-24srs\" (UniqueName: \"kubernetes.io/projected/38005ee2-6635-498c-b156-ac57b4f2f4d9-kube-api-access-24srs\") pod \"38005ee2-6635-498c-b156-ac57b4f2f4d9\" (UID: \"38005ee2-6635-498c-b156-ac57b4f2f4d9\") " Dec 11 08:17:29 crc kubenswrapper[4860]: I1211 08:17:29.585916 4860 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/38005ee2-6635-498c-b156-ac57b4f2f4d9-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 11 08:17:29 crc kubenswrapper[4860]: I1211 08:17:29.585935 4860 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/38005ee2-6635-498c-b156-ac57b4f2f4d9-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 11 08:17:29 crc kubenswrapper[4860]: I1211 08:17:29.592757 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/38005ee2-6635-498c-b156-ac57b4f2f4d9-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "38005ee2-6635-498c-b156-ac57b4f2f4d9" (UID: "38005ee2-6635-498c-b156-ac57b4f2f4d9"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:17:29 crc kubenswrapper[4860]: I1211 08:17:29.593491 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/38005ee2-6635-498c-b156-ac57b4f2f4d9-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "38005ee2-6635-498c-b156-ac57b4f2f4d9" (UID: "38005ee2-6635-498c-b156-ac57b4f2f4d9"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:17:29 crc kubenswrapper[4860]: I1211 08:17:29.593941 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/38005ee2-6635-498c-b156-ac57b4f2f4d9-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "38005ee2-6635-498c-b156-ac57b4f2f4d9" (UID: "38005ee2-6635-498c-b156-ac57b4f2f4d9"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:17:29 crc kubenswrapper[4860]: I1211 08:17:29.594426 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/38005ee2-6635-498c-b156-ac57b4f2f4d9-kube-api-access-24srs" (OuterVolumeSpecName: "kube-api-access-24srs") pod "38005ee2-6635-498c-b156-ac57b4f2f4d9" (UID: "38005ee2-6635-498c-b156-ac57b4f2f4d9"). InnerVolumeSpecName "kube-api-access-24srs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:17:29 crc kubenswrapper[4860]: I1211 08:17:29.606748 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "38005ee2-6635-498c-b156-ac57b4f2f4d9" (UID: "38005ee2-6635-498c-b156-ac57b4f2f4d9"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 11 08:17:29 crc kubenswrapper[4860]: I1211 08:17:29.619539 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/38005ee2-6635-498c-b156-ac57b4f2f4d9-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "38005ee2-6635-498c-b156-ac57b4f2f4d9" (UID: "38005ee2-6635-498c-b156-ac57b4f2f4d9"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:17:29 crc kubenswrapper[4860]: I1211 08:17:29.688214 4860 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/38005ee2-6635-498c-b156-ac57b4f2f4d9-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 11 08:17:29 crc kubenswrapper[4860]: I1211 08:17:29.688443 4860 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/38005ee2-6635-498c-b156-ac57b4f2f4d9-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 11 08:17:29 crc kubenswrapper[4860]: I1211 08:17:29.688502 4860 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/38005ee2-6635-498c-b156-ac57b4f2f4d9-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 11 08:17:29 crc kubenswrapper[4860]: I1211 08:17:29.688519 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-24srs\" (UniqueName: \"kubernetes.io/projected/38005ee2-6635-498c-b156-ac57b4f2f4d9-kube-api-access-24srs\") on node \"crc\" DevicePath \"\"" Dec 11 08:17:29 crc kubenswrapper[4860]: I1211 08:17:29.688532 4860 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/38005ee2-6635-498c-b156-ac57b4f2f4d9-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 11 08:17:29 crc kubenswrapper[4860]: I1211 08:17:29.775076 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-zf572" event={"ID":"38005ee2-6635-498c-b156-ac57b4f2f4d9","Type":"ContainerDied","Data":"292d1c897abc2bdf07e4b29e9ad59768345c805bda79fc0adb56dabb549b08ca"} Dec 11 08:17:29 crc kubenswrapper[4860]: I1211 08:17:29.775129 4860 scope.go:117] "RemoveContainer" containerID="027d8ed039343e3ce7d35642642a40ab245c222261386a8d009a083692ad1f2b" Dec 11 08:17:29 crc kubenswrapper[4860]: I1211 08:17:29.775230 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-zf572" Dec 11 08:17:29 crc kubenswrapper[4860]: I1211 08:17:29.816014 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-zf572"] Dec 11 08:17:29 crc kubenswrapper[4860]: I1211 08:17:29.820770 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-zf572"] Dec 11 08:17:31 crc kubenswrapper[4860]: I1211 08:17:31.590268 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="38005ee2-6635-498c-b156-ac57b4f2f4d9" path="/var/lib/kubelet/pods/38005ee2-6635-498c-b156-ac57b4f2f4d9/volumes" Dec 11 08:17:38 crc kubenswrapper[4860]: I1211 08:17:38.795069 4860 patch_prober.go:28] interesting pod/machine-config-daemon-99qgp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 08:17:38 crc kubenswrapper[4860]: I1211 08:17:38.795915 4860 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 08:18:08 crc kubenswrapper[4860]: I1211 08:18:08.795736 4860 patch_prober.go:28] interesting pod/machine-config-daemon-99qgp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 08:18:08 crc kubenswrapper[4860]: I1211 08:18:08.796401 4860 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 08:18:08 crc kubenswrapper[4860]: I1211 08:18:08.796454 4860 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" Dec 11 08:18:08 crc kubenswrapper[4860]: I1211 08:18:08.797119 4860 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"cddf60b7695e04c9618f575369ea700b35bf017d924547b72625c2bacba04e02"} pod="openshift-machine-config-operator/machine-config-daemon-99qgp" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 11 08:18:08 crc kubenswrapper[4860]: I1211 08:18:08.797192 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" containerName="machine-config-daemon" containerID="cri-o://cddf60b7695e04c9618f575369ea700b35bf017d924547b72625c2bacba04e02" gracePeriod=600 Dec 11 08:18:09 crc kubenswrapper[4860]: I1211 08:18:09.022113 4860 generic.go:334] "Generic (PLEG): container finished" podID="31c30642-6e60-41b4-a477-0d802424e0aa" containerID="cddf60b7695e04c9618f575369ea700b35bf017d924547b72625c2bacba04e02" exitCode=0 Dec 11 08:18:09 crc kubenswrapper[4860]: I1211 08:18:09.022145 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" event={"ID":"31c30642-6e60-41b4-a477-0d802424e0aa","Type":"ContainerDied","Data":"cddf60b7695e04c9618f575369ea700b35bf017d924547b72625c2bacba04e02"} Dec 11 08:18:09 crc kubenswrapper[4860]: I1211 08:18:09.022467 4860 scope.go:117] "RemoveContainer" containerID="b2d4d5de7106175087d91acb1f0c7552014f0c08bfcd65cbbf184d3f8d7826ff" Dec 11 08:18:10 crc kubenswrapper[4860]: I1211 08:18:10.032573 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" event={"ID":"31c30642-6e60-41b4-a477-0d802424e0aa","Type":"ContainerStarted","Data":"ff37f2c081c95677939e414d6be18406d4d67feea33292ff469d545943df00a5"} Dec 11 08:20:38 crc kubenswrapper[4860]: I1211 08:20:38.795573 4860 patch_prober.go:28] interesting pod/machine-config-daemon-99qgp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 08:20:38 crc kubenswrapper[4860]: I1211 08:20:38.796137 4860 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 08:21:08 crc kubenswrapper[4860]: I1211 08:21:08.795499 4860 patch_prober.go:28] interesting pod/machine-config-daemon-99qgp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 08:21:08 crc kubenswrapper[4860]: I1211 08:21:08.796086 4860 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 08:21:38 crc kubenswrapper[4860]: I1211 08:21:38.795763 4860 patch_prober.go:28] interesting pod/machine-config-daemon-99qgp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 08:21:38 crc kubenswrapper[4860]: I1211 08:21:38.796359 4860 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 08:21:38 crc kubenswrapper[4860]: I1211 08:21:38.796416 4860 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" Dec 11 08:21:38 crc kubenswrapper[4860]: I1211 08:21:38.797063 4860 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ff37f2c081c95677939e414d6be18406d4d67feea33292ff469d545943df00a5"} pod="openshift-machine-config-operator/machine-config-daemon-99qgp" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 11 08:21:38 crc kubenswrapper[4860]: I1211 08:21:38.797180 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" containerName="machine-config-daemon" containerID="cri-o://ff37f2c081c95677939e414d6be18406d4d67feea33292ff469d545943df00a5" gracePeriod=600 Dec 11 08:21:39 crc kubenswrapper[4860]: I1211 08:21:39.252831 4860 generic.go:334] "Generic (PLEG): container finished" podID="31c30642-6e60-41b4-a477-0d802424e0aa" containerID="ff37f2c081c95677939e414d6be18406d4d67feea33292ff469d545943df00a5" exitCode=0 Dec 11 08:21:39 crc kubenswrapper[4860]: I1211 08:21:39.252875 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" event={"ID":"31c30642-6e60-41b4-a477-0d802424e0aa","Type":"ContainerDied","Data":"ff37f2c081c95677939e414d6be18406d4d67feea33292ff469d545943df00a5"} Dec 11 08:21:39 crc kubenswrapper[4860]: I1211 08:21:39.252908 4860 scope.go:117] "RemoveContainer" containerID="cddf60b7695e04c9618f575369ea700b35bf017d924547b72625c2bacba04e02" Dec 11 08:21:40 crc kubenswrapper[4860]: I1211 08:21:40.260486 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" event={"ID":"31c30642-6e60-41b4-a477-0d802424e0aa","Type":"ContainerStarted","Data":"7abbc21a5acc8c61b19323dc53f4bb2e28d39cd9e4964ba8a2b38c1bd09ed9bf"} Dec 11 08:22:15 crc kubenswrapper[4860]: I1211 08:22:15.639890 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-qhfrs"] Dec 11 08:22:15 crc kubenswrapper[4860]: E1211 08:22:15.640700 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38005ee2-6635-498c-b156-ac57b4f2f4d9" containerName="registry" Dec 11 08:22:15 crc kubenswrapper[4860]: I1211 08:22:15.640717 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="38005ee2-6635-498c-b156-ac57b4f2f4d9" containerName="registry" Dec 11 08:22:15 crc kubenswrapper[4860]: I1211 08:22:15.640858 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="38005ee2-6635-498c-b156-ac57b4f2f4d9" containerName="registry" Dec 11 08:22:15 crc kubenswrapper[4860]: I1211 08:22:15.641305 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-qhfrs" Dec 11 08:22:15 crc kubenswrapper[4860]: I1211 08:22:15.643554 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Dec 11 08:22:15 crc kubenswrapper[4860]: I1211 08:22:15.643730 4860 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-trzbm" Dec 11 08:22:15 crc kubenswrapper[4860]: I1211 08:22:15.643796 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Dec 11 08:22:15 crc kubenswrapper[4860]: I1211 08:22:15.648569 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-7bszh"] Dec 11 08:22:15 crc kubenswrapper[4860]: I1211 08:22:15.649340 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-7bszh" Dec 11 08:22:15 crc kubenswrapper[4860]: I1211 08:22:15.651557 4860 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-n2g4j" Dec 11 08:22:15 crc kubenswrapper[4860]: I1211 08:22:15.652782 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-qhfrs"] Dec 11 08:22:15 crc kubenswrapper[4860]: I1211 08:22:15.661697 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-7bszh"] Dec 11 08:22:15 crc kubenswrapper[4860]: I1211 08:22:15.675878 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-rvtht"] Dec 11 08:22:15 crc kubenswrapper[4860]: I1211 08:22:15.676558 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-rvtht" Dec 11 08:22:15 crc kubenswrapper[4860]: I1211 08:22:15.684456 4860 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-c849x" Dec 11 08:22:15 crc kubenswrapper[4860]: I1211 08:22:15.697407 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-rvtht"] Dec 11 08:22:15 crc kubenswrapper[4860]: I1211 08:22:15.763007 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vscp8\" (UniqueName: \"kubernetes.io/projected/29ede248-cc17-41b1-a1c4-4feaa6010cf2-kube-api-access-vscp8\") pod \"cert-manager-cainjector-7f985d654d-qhfrs\" (UID: \"29ede248-cc17-41b1-a1c4-4feaa6010cf2\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-qhfrs" Dec 11 08:22:15 crc kubenswrapper[4860]: I1211 08:22:15.763368 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5dvtt\" (UniqueName: \"kubernetes.io/projected/5198d79a-f8ab-4a3a-9859-0925545321a4-kube-api-access-5dvtt\") pod \"cert-manager-5b446d88c5-7bszh\" (UID: \"5198d79a-f8ab-4a3a-9859-0925545321a4\") " pod="cert-manager/cert-manager-5b446d88c5-7bszh" Dec 11 08:22:15 crc kubenswrapper[4860]: I1211 08:22:15.864267 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5dvtt\" (UniqueName: \"kubernetes.io/projected/5198d79a-f8ab-4a3a-9859-0925545321a4-kube-api-access-5dvtt\") pod \"cert-manager-5b446d88c5-7bszh\" (UID: \"5198d79a-f8ab-4a3a-9859-0925545321a4\") " pod="cert-manager/cert-manager-5b446d88c5-7bszh" Dec 11 08:22:15 crc kubenswrapper[4860]: I1211 08:22:15.864339 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6dd2q\" (UniqueName: \"kubernetes.io/projected/7ea21c5c-320d-4bb0-9bea-186a528a61f4-kube-api-access-6dd2q\") pod \"cert-manager-webhook-5655c58dd6-rvtht\" (UID: \"7ea21c5c-320d-4bb0-9bea-186a528a61f4\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-rvtht" Dec 11 08:22:15 crc kubenswrapper[4860]: I1211 08:22:15.864377 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vscp8\" (UniqueName: \"kubernetes.io/projected/29ede248-cc17-41b1-a1c4-4feaa6010cf2-kube-api-access-vscp8\") pod \"cert-manager-cainjector-7f985d654d-qhfrs\" (UID: \"29ede248-cc17-41b1-a1c4-4feaa6010cf2\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-qhfrs" Dec 11 08:22:15 crc kubenswrapper[4860]: I1211 08:22:15.882082 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vscp8\" (UniqueName: \"kubernetes.io/projected/29ede248-cc17-41b1-a1c4-4feaa6010cf2-kube-api-access-vscp8\") pod \"cert-manager-cainjector-7f985d654d-qhfrs\" (UID: \"29ede248-cc17-41b1-a1c4-4feaa6010cf2\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-qhfrs" Dec 11 08:22:15 crc kubenswrapper[4860]: I1211 08:22:15.882082 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5dvtt\" (UniqueName: \"kubernetes.io/projected/5198d79a-f8ab-4a3a-9859-0925545321a4-kube-api-access-5dvtt\") pod \"cert-manager-5b446d88c5-7bszh\" (UID: \"5198d79a-f8ab-4a3a-9859-0925545321a4\") " pod="cert-manager/cert-manager-5b446d88c5-7bszh" Dec 11 08:22:15 crc kubenswrapper[4860]: I1211 08:22:15.958304 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-qhfrs" Dec 11 08:22:15 crc kubenswrapper[4860]: I1211 08:22:15.965823 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6dd2q\" (UniqueName: \"kubernetes.io/projected/7ea21c5c-320d-4bb0-9bea-186a528a61f4-kube-api-access-6dd2q\") pod \"cert-manager-webhook-5655c58dd6-rvtht\" (UID: \"7ea21c5c-320d-4bb0-9bea-186a528a61f4\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-rvtht" Dec 11 08:22:15 crc kubenswrapper[4860]: I1211 08:22:15.966260 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-7bszh" Dec 11 08:22:15 crc kubenswrapper[4860]: I1211 08:22:15.984616 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6dd2q\" (UniqueName: \"kubernetes.io/projected/7ea21c5c-320d-4bb0-9bea-186a528a61f4-kube-api-access-6dd2q\") pod \"cert-manager-webhook-5655c58dd6-rvtht\" (UID: \"7ea21c5c-320d-4bb0-9bea-186a528a61f4\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-rvtht" Dec 11 08:22:15 crc kubenswrapper[4860]: I1211 08:22:15.995612 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-rvtht" Dec 11 08:22:16 crc kubenswrapper[4860]: I1211 08:22:16.276538 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-qhfrs"] Dec 11 08:22:16 crc kubenswrapper[4860]: I1211 08:22:16.281601 4860 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 11 08:22:16 crc kubenswrapper[4860]: I1211 08:22:16.401434 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-7bszh"] Dec 11 08:22:16 crc kubenswrapper[4860]: W1211 08:22:16.404604 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5198d79a_f8ab_4a3a_9859_0925545321a4.slice/crio-64b6de3f6009e8b1e8df3234c5933a70e909e56d65bbf5a910fadce8e4c2eb60 WatchSource:0}: Error finding container 64b6de3f6009e8b1e8df3234c5933a70e909e56d65bbf5a910fadce8e4c2eb60: Status 404 returned error can't find the container with id 64b6de3f6009e8b1e8df3234c5933a70e909e56d65bbf5a910fadce8e4c2eb60 Dec 11 08:22:16 crc kubenswrapper[4860]: I1211 08:22:16.436495 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-7bszh" event={"ID":"5198d79a-f8ab-4a3a-9859-0925545321a4","Type":"ContainerStarted","Data":"64b6de3f6009e8b1e8df3234c5933a70e909e56d65bbf5a910fadce8e4c2eb60"} Dec 11 08:22:16 crc kubenswrapper[4860]: I1211 08:22:16.437576 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-qhfrs" event={"ID":"29ede248-cc17-41b1-a1c4-4feaa6010cf2","Type":"ContainerStarted","Data":"218475b0a82c10fae598113b2f43eacca55572f9bb6f41ba0228f8a1481e43a9"} Dec 11 08:22:16 crc kubenswrapper[4860]: I1211 08:22:16.438381 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-rvtht"] Dec 11 08:22:16 crc kubenswrapper[4860]: W1211 08:22:16.442833 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7ea21c5c_320d_4bb0_9bea_186a528a61f4.slice/crio-4422271bb5c1c1895072ee91568173ea8d3c45897ea26c8326ce0df5398e9bfa WatchSource:0}: Error finding container 4422271bb5c1c1895072ee91568173ea8d3c45897ea26c8326ce0df5398e9bfa: Status 404 returned error can't find the container with id 4422271bb5c1c1895072ee91568173ea8d3c45897ea26c8326ce0df5398e9bfa Dec 11 08:22:17 crc kubenswrapper[4860]: I1211 08:22:17.448387 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-rvtht" event={"ID":"7ea21c5c-320d-4bb0-9bea-186a528a61f4","Type":"ContainerStarted","Data":"4422271bb5c1c1895072ee91568173ea8d3c45897ea26c8326ce0df5398e9bfa"} Dec 11 08:22:19 crc kubenswrapper[4860]: I1211 08:22:19.481588 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-7bszh" event={"ID":"5198d79a-f8ab-4a3a-9859-0925545321a4","Type":"ContainerStarted","Data":"703fb87c3e2bf18a10a8a32c059fc86c8b51d96f1868cf1ac4bafc95e21b39b1"} Dec 11 08:22:19 crc kubenswrapper[4860]: I1211 08:22:19.485546 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-qhfrs" event={"ID":"29ede248-cc17-41b1-a1c4-4feaa6010cf2","Type":"ContainerStarted","Data":"bca7073b896cd5cc9fad999b235aa9c70f79f3740b73e716b2f2a84bc87813b0"} Dec 11 08:22:19 crc kubenswrapper[4860]: I1211 08:22:19.499515 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-7bszh" podStartSLOduration=1.741615771 podStartE2EDuration="4.49949365s" podCreationTimestamp="2025-12-11 08:22:15 +0000 UTC" firstStartedPulling="2025-12-11 08:22:16.407217971 +0000 UTC m=+669.135737026" lastFinishedPulling="2025-12-11 08:22:19.16509585 +0000 UTC m=+671.893614905" observedRunningTime="2025-12-11 08:22:19.496356852 +0000 UTC m=+672.224875907" watchObservedRunningTime="2025-12-11 08:22:19.49949365 +0000 UTC m=+672.228012705" Dec 11 08:22:20 crc kubenswrapper[4860]: I1211 08:22:20.491742 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-rvtht" event={"ID":"7ea21c5c-320d-4bb0-9bea-186a528a61f4","Type":"ContainerStarted","Data":"8f7cc07dae710cc50e5f9981575a33b57c95b0ba72b62a5c417efb2da32b093b"} Dec 11 08:22:20 crc kubenswrapper[4860]: I1211 08:22:20.507243 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-qhfrs" podStartSLOduration=2.686774192 podStartE2EDuration="5.507218629s" podCreationTimestamp="2025-12-11 08:22:15 +0000 UTC" firstStartedPulling="2025-12-11 08:22:16.281406876 +0000 UTC m=+669.009925931" lastFinishedPulling="2025-12-11 08:22:19.101851313 +0000 UTC m=+671.830370368" observedRunningTime="2025-12-11 08:22:19.519300049 +0000 UTC m=+672.247819104" watchObservedRunningTime="2025-12-11 08:22:20.507218629 +0000 UTC m=+673.235737704" Dec 11 08:22:20 crc kubenswrapper[4860]: I1211 08:22:20.512343 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-rvtht" podStartSLOduration=1.9484645409999999 podStartE2EDuration="5.512328279s" podCreationTimestamp="2025-12-11 08:22:15 +0000 UTC" firstStartedPulling="2025-12-11 08:22:16.4448834 +0000 UTC m=+669.173402455" lastFinishedPulling="2025-12-11 08:22:20.008747138 +0000 UTC m=+672.737266193" observedRunningTime="2025-12-11 08:22:20.506435654 +0000 UTC m=+673.234954719" watchObservedRunningTime="2025-12-11 08:22:20.512328279 +0000 UTC m=+673.240847334" Dec 11 08:22:20 crc kubenswrapper[4860]: I1211 08:22:20.996991 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-rvtht" Dec 11 08:22:25 crc kubenswrapper[4860]: I1211 08:22:25.998576 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-rvtht" Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.396306 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-xvnqp"] Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.397138 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" podUID="0589a204-c98c-417d-8256-bfe3bf747660" containerName="nbdb" containerID="cri-o://6dfcc4370edaae36a5a32b0a9c29290c50f8e4f427e4e147dc3c779beaa8b780" gracePeriod=30 Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.397246 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" podUID="0589a204-c98c-417d-8256-bfe3bf747660" containerName="sbdb" containerID="cri-o://bb80d9928e8349fc5963227d135355ad7ba820d4fd89877af601d005ca03334b" gracePeriod=30 Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.397126 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" podUID="0589a204-c98c-417d-8256-bfe3bf747660" containerName="ovn-controller" containerID="cri-o://c9984d5d242bb258405efe91fc9caf9e7ea592834445df2e78add9553a921c76" gracePeriod=30 Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.397358 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" podUID="0589a204-c98c-417d-8256-bfe3bf747660" containerName="kube-rbac-proxy-node" containerID="cri-o://3333b4346f59b04017e32d61f457ba7ff457166731128f339d7987ae6b56b06d" gracePeriod=30 Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.397369 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" podUID="0589a204-c98c-417d-8256-bfe3bf747660" containerName="northd" containerID="cri-o://d856ae95e9077524c518d82d10fb6b93b7fbcaaf2b720ffafe6f7fae58bd7caa" gracePeriod=30 Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.397384 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" podUID="0589a204-c98c-417d-8256-bfe3bf747660" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://52bb8ade314afb9494f4e026d0af2aea7b1970d64dc66102f55e257ae4678154" gracePeriod=30 Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.397434 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" podUID="0589a204-c98c-417d-8256-bfe3bf747660" containerName="ovn-acl-logging" containerID="cri-o://1426084ac580895c42354630910dd8c8e3a48ddd25f22abce1d28dc3de332a63" gracePeriod=30 Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.435154 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" podUID="0589a204-c98c-417d-8256-bfe3bf747660" containerName="ovnkube-controller" containerID="cri-o://4513cf6bc605dd49d55acb42fa24068e7e61a7b40104f7c5a17dc78cc5694388" gracePeriod=30 Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.525478 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-pznwl_5310c1fc-66c6-40aa-b1bf-5a59a2410f9c/kube-multus/2.log" Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.526055 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-pznwl_5310c1fc-66c6-40aa-b1bf-5a59a2410f9c/kube-multus/1.log" Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.526114 4860 generic.go:334] "Generic (PLEG): container finished" podID="5310c1fc-66c6-40aa-b1bf-5a59a2410f9c" containerID="c50698b594e2b9e096d97dd750433d216996139113b1349d3ba87a70d057511b" exitCode=2 Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.526194 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-pznwl" event={"ID":"5310c1fc-66c6-40aa-b1bf-5a59a2410f9c","Type":"ContainerDied","Data":"c50698b594e2b9e096d97dd750433d216996139113b1349d3ba87a70d057511b"} Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.526235 4860 scope.go:117] "RemoveContainer" containerID="34119a6367261d7e12174f914ee8cfaa55703a55210f83c58b0e4694c54a5b03" Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.526707 4860 scope.go:117] "RemoveContainer" containerID="c50698b594e2b9e096d97dd750433d216996139113b1349d3ba87a70d057511b" Dec 11 08:22:26 crc kubenswrapper[4860]: E1211 08:22:26.526905 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-pznwl_openshift-multus(5310c1fc-66c6-40aa-b1bf-5a59a2410f9c)\"" pod="openshift-multus/multus-pznwl" podUID="5310c1fc-66c6-40aa-b1bf-5a59a2410f9c" Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.538129 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-xvnqp_0589a204-c98c-417d-8256-bfe3bf747660/ovnkube-controller/3.log" Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.539535 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-xvnqp_0589a204-c98c-417d-8256-bfe3bf747660/ovn-acl-logging/0.log" Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.539932 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-xvnqp_0589a204-c98c-417d-8256-bfe3bf747660/ovn-controller/0.log" Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.540234 4860 generic.go:334] "Generic (PLEG): container finished" podID="0589a204-c98c-417d-8256-bfe3bf747660" containerID="d856ae95e9077524c518d82d10fb6b93b7fbcaaf2b720ffafe6f7fae58bd7caa" exitCode=0 Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.540258 4860 generic.go:334] "Generic (PLEG): container finished" podID="0589a204-c98c-417d-8256-bfe3bf747660" containerID="52bb8ade314afb9494f4e026d0af2aea7b1970d64dc66102f55e257ae4678154" exitCode=0 Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.540266 4860 generic.go:334] "Generic (PLEG): container finished" podID="0589a204-c98c-417d-8256-bfe3bf747660" containerID="3333b4346f59b04017e32d61f457ba7ff457166731128f339d7987ae6b56b06d" exitCode=0 Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.540274 4860 generic.go:334] "Generic (PLEG): container finished" podID="0589a204-c98c-417d-8256-bfe3bf747660" containerID="1426084ac580895c42354630910dd8c8e3a48ddd25f22abce1d28dc3de332a63" exitCode=143 Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.540281 4860 generic.go:334] "Generic (PLEG): container finished" podID="0589a204-c98c-417d-8256-bfe3bf747660" containerID="c9984d5d242bb258405efe91fc9caf9e7ea592834445df2e78add9553a921c76" exitCode=143 Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.540300 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" event={"ID":"0589a204-c98c-417d-8256-bfe3bf747660","Type":"ContainerDied","Data":"d856ae95e9077524c518d82d10fb6b93b7fbcaaf2b720ffafe6f7fae58bd7caa"} Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.540326 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" event={"ID":"0589a204-c98c-417d-8256-bfe3bf747660","Type":"ContainerDied","Data":"52bb8ade314afb9494f4e026d0af2aea7b1970d64dc66102f55e257ae4678154"} Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.540337 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" event={"ID":"0589a204-c98c-417d-8256-bfe3bf747660","Type":"ContainerDied","Data":"3333b4346f59b04017e32d61f457ba7ff457166731128f339d7987ae6b56b06d"} Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.540345 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" event={"ID":"0589a204-c98c-417d-8256-bfe3bf747660","Type":"ContainerDied","Data":"1426084ac580895c42354630910dd8c8e3a48ddd25f22abce1d28dc3de332a63"} Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.540354 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" event={"ID":"0589a204-c98c-417d-8256-bfe3bf747660","Type":"ContainerDied","Data":"c9984d5d242bb258405efe91fc9caf9e7ea592834445df2e78add9553a921c76"} Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.673068 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-xvnqp_0589a204-c98c-417d-8256-bfe3bf747660/ovnkube-controller/3.log" Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.677679 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-xvnqp_0589a204-c98c-417d-8256-bfe3bf747660/ovn-acl-logging/0.log" Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.678516 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-xvnqp_0589a204-c98c-417d-8256-bfe3bf747660/ovn-controller/0.log" Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.678970 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.724404 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-w2n7w"] Dec 11 08:22:26 crc kubenswrapper[4860]: E1211 08:22:26.724619 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0589a204-c98c-417d-8256-bfe3bf747660" containerName="ovn-controller" Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.724632 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="0589a204-c98c-417d-8256-bfe3bf747660" containerName="ovn-controller" Dec 11 08:22:26 crc kubenswrapper[4860]: E1211 08:22:26.724666 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0589a204-c98c-417d-8256-bfe3bf747660" containerName="ovnkube-controller" Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.724674 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="0589a204-c98c-417d-8256-bfe3bf747660" containerName="ovnkube-controller" Dec 11 08:22:26 crc kubenswrapper[4860]: E1211 08:22:26.724681 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0589a204-c98c-417d-8256-bfe3bf747660" containerName="ovnkube-controller" Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.724688 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="0589a204-c98c-417d-8256-bfe3bf747660" containerName="ovnkube-controller" Dec 11 08:22:26 crc kubenswrapper[4860]: E1211 08:22:26.724695 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0589a204-c98c-417d-8256-bfe3bf747660" containerName="kubecfg-setup" Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.724703 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="0589a204-c98c-417d-8256-bfe3bf747660" containerName="kubecfg-setup" Dec 11 08:22:26 crc kubenswrapper[4860]: E1211 08:22:26.724713 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0589a204-c98c-417d-8256-bfe3bf747660" containerName="kube-rbac-proxy-ovn-metrics" Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.724718 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="0589a204-c98c-417d-8256-bfe3bf747660" containerName="kube-rbac-proxy-ovn-metrics" Dec 11 08:22:26 crc kubenswrapper[4860]: E1211 08:22:26.724763 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0589a204-c98c-417d-8256-bfe3bf747660" containerName="sbdb" Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.724770 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="0589a204-c98c-417d-8256-bfe3bf747660" containerName="sbdb" Dec 11 08:22:26 crc kubenswrapper[4860]: E1211 08:22:26.724781 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0589a204-c98c-417d-8256-bfe3bf747660" containerName="ovnkube-controller" Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.724788 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="0589a204-c98c-417d-8256-bfe3bf747660" containerName="ovnkube-controller" Dec 11 08:22:26 crc kubenswrapper[4860]: E1211 08:22:26.724795 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0589a204-c98c-417d-8256-bfe3bf747660" containerName="ovn-acl-logging" Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.724801 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="0589a204-c98c-417d-8256-bfe3bf747660" containerName="ovn-acl-logging" Dec 11 08:22:26 crc kubenswrapper[4860]: E1211 08:22:26.724810 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0589a204-c98c-417d-8256-bfe3bf747660" containerName="kube-rbac-proxy-node" Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.724815 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="0589a204-c98c-417d-8256-bfe3bf747660" containerName="kube-rbac-proxy-node" Dec 11 08:22:26 crc kubenswrapper[4860]: E1211 08:22:26.724821 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0589a204-c98c-417d-8256-bfe3bf747660" containerName="northd" Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.724827 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="0589a204-c98c-417d-8256-bfe3bf747660" containerName="northd" Dec 11 08:22:26 crc kubenswrapper[4860]: E1211 08:22:26.724835 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0589a204-c98c-417d-8256-bfe3bf747660" containerName="nbdb" Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.724840 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="0589a204-c98c-417d-8256-bfe3bf747660" containerName="nbdb" Dec 11 08:22:26 crc kubenswrapper[4860]: E1211 08:22:26.724848 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0589a204-c98c-417d-8256-bfe3bf747660" containerName="ovnkube-controller" Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.724927 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="0589a204-c98c-417d-8256-bfe3bf747660" containerName="ovnkube-controller" Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.725023 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="0589a204-c98c-417d-8256-bfe3bf747660" containerName="ovnkube-controller" Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.725035 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="0589a204-c98c-417d-8256-bfe3bf747660" containerName="northd" Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.725045 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="0589a204-c98c-417d-8256-bfe3bf747660" containerName="ovn-acl-logging" Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.725056 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="0589a204-c98c-417d-8256-bfe3bf747660" containerName="kube-rbac-proxy-ovn-metrics" Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.725063 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="0589a204-c98c-417d-8256-bfe3bf747660" containerName="nbdb" Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.725070 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="0589a204-c98c-417d-8256-bfe3bf747660" containerName="ovn-controller" Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.725077 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="0589a204-c98c-417d-8256-bfe3bf747660" containerName="ovnkube-controller" Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.725084 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="0589a204-c98c-417d-8256-bfe3bf747660" containerName="ovnkube-controller" Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.725090 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="0589a204-c98c-417d-8256-bfe3bf747660" containerName="sbdb" Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.725097 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="0589a204-c98c-417d-8256-bfe3bf747660" containerName="ovnkube-controller" Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.725104 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="0589a204-c98c-417d-8256-bfe3bf747660" containerName="kube-rbac-proxy-node" Dec 11 08:22:26 crc kubenswrapper[4860]: E1211 08:22:26.725189 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0589a204-c98c-417d-8256-bfe3bf747660" containerName="ovnkube-controller" Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.725195 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="0589a204-c98c-417d-8256-bfe3bf747660" containerName="ovnkube-controller" Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.725301 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="0589a204-c98c-417d-8256-bfe3bf747660" containerName="ovnkube-controller" Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.727333 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-w2n7w" Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.809495 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/0589a204-c98c-417d-8256-bfe3bf747660-log-socket\") pod \"0589a204-c98c-417d-8256-bfe3bf747660\" (UID: \"0589a204-c98c-417d-8256-bfe3bf747660\") " Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.809539 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/0589a204-c98c-417d-8256-bfe3bf747660-run-ovn\") pod \"0589a204-c98c-417d-8256-bfe3bf747660\" (UID: \"0589a204-c98c-417d-8256-bfe3bf747660\") " Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.809555 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0589a204-c98c-417d-8256-bfe3bf747660-log-socket" (OuterVolumeSpecName: "log-socket") pod "0589a204-c98c-417d-8256-bfe3bf747660" (UID: "0589a204-c98c-417d-8256-bfe3bf747660"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.809574 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/0589a204-c98c-417d-8256-bfe3bf747660-host-kubelet\") pod \"0589a204-c98c-417d-8256-bfe3bf747660\" (UID: \"0589a204-c98c-417d-8256-bfe3bf747660\") " Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.809588 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0589a204-c98c-417d-8256-bfe3bf747660-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "0589a204-c98c-417d-8256-bfe3bf747660" (UID: "0589a204-c98c-417d-8256-bfe3bf747660"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.809592 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/0589a204-c98c-417d-8256-bfe3bf747660-node-log\") pod \"0589a204-c98c-417d-8256-bfe3bf747660\" (UID: \"0589a204-c98c-417d-8256-bfe3bf747660\") " Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.809607 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0589a204-c98c-417d-8256-bfe3bf747660-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "0589a204-c98c-417d-8256-bfe3bf747660" (UID: "0589a204-c98c-417d-8256-bfe3bf747660"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.809620 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/0589a204-c98c-417d-8256-bfe3bf747660-host-run-netns\") pod \"0589a204-c98c-417d-8256-bfe3bf747660\" (UID: \"0589a204-c98c-417d-8256-bfe3bf747660\") " Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.809660 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/0589a204-c98c-417d-8256-bfe3bf747660-run-systemd\") pod \"0589a204-c98c-417d-8256-bfe3bf747660\" (UID: \"0589a204-c98c-417d-8256-bfe3bf747660\") " Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.809671 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0589a204-c98c-417d-8256-bfe3bf747660-node-log" (OuterVolumeSpecName: "node-log") pod "0589a204-c98c-417d-8256-bfe3bf747660" (UID: "0589a204-c98c-417d-8256-bfe3bf747660"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.809699 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tftsj\" (UniqueName: \"kubernetes.io/projected/0589a204-c98c-417d-8256-bfe3bf747660-kube-api-access-tftsj\") pod \"0589a204-c98c-417d-8256-bfe3bf747660\" (UID: \"0589a204-c98c-417d-8256-bfe3bf747660\") " Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.809732 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/0589a204-c98c-417d-8256-bfe3bf747660-host-cni-bin\") pod \"0589a204-c98c-417d-8256-bfe3bf747660\" (UID: \"0589a204-c98c-417d-8256-bfe3bf747660\") " Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.809730 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0589a204-c98c-417d-8256-bfe3bf747660-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "0589a204-c98c-417d-8256-bfe3bf747660" (UID: "0589a204-c98c-417d-8256-bfe3bf747660"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.809783 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0589a204-c98c-417d-8256-bfe3bf747660-run-openvswitch\") pod \"0589a204-c98c-417d-8256-bfe3bf747660\" (UID: \"0589a204-c98c-417d-8256-bfe3bf747660\") " Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.809804 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/0589a204-c98c-417d-8256-bfe3bf747660-host-cni-netd\") pod \"0589a204-c98c-417d-8256-bfe3bf747660\" (UID: \"0589a204-c98c-417d-8256-bfe3bf747660\") " Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.809830 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0589a204-c98c-417d-8256-bfe3bf747660-var-lib-openvswitch\") pod \"0589a204-c98c-417d-8256-bfe3bf747660\" (UID: \"0589a204-c98c-417d-8256-bfe3bf747660\") " Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.809854 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0589a204-c98c-417d-8256-bfe3bf747660-host-run-ovn-kubernetes\") pod \"0589a204-c98c-417d-8256-bfe3bf747660\" (UID: \"0589a204-c98c-417d-8256-bfe3bf747660\") " Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.809803 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0589a204-c98c-417d-8256-bfe3bf747660-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "0589a204-c98c-417d-8256-bfe3bf747660" (UID: "0589a204-c98c-417d-8256-bfe3bf747660"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.809827 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0589a204-c98c-417d-8256-bfe3bf747660-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "0589a204-c98c-417d-8256-bfe3bf747660" (UID: "0589a204-c98c-417d-8256-bfe3bf747660"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.809876 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/0589a204-c98c-417d-8256-bfe3bf747660-host-slash\") pod \"0589a204-c98c-417d-8256-bfe3bf747660\" (UID: \"0589a204-c98c-417d-8256-bfe3bf747660\") " Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.809898 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/0589a204-c98c-417d-8256-bfe3bf747660-systemd-units\") pod \"0589a204-c98c-417d-8256-bfe3bf747660\" (UID: \"0589a204-c98c-417d-8256-bfe3bf747660\") " Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.809926 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/0589a204-c98c-417d-8256-bfe3bf747660-ovnkube-config\") pod \"0589a204-c98c-417d-8256-bfe3bf747660\" (UID: \"0589a204-c98c-417d-8256-bfe3bf747660\") " Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.809948 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0589a204-c98c-417d-8256-bfe3bf747660-host-var-lib-cni-networks-ovn-kubernetes\") pod \"0589a204-c98c-417d-8256-bfe3bf747660\" (UID: \"0589a204-c98c-417d-8256-bfe3bf747660\") " Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.809853 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0589a204-c98c-417d-8256-bfe3bf747660-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "0589a204-c98c-417d-8256-bfe3bf747660" (UID: "0589a204-c98c-417d-8256-bfe3bf747660"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.809898 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0589a204-c98c-417d-8256-bfe3bf747660-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "0589a204-c98c-417d-8256-bfe3bf747660" (UID: "0589a204-c98c-417d-8256-bfe3bf747660"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.809919 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0589a204-c98c-417d-8256-bfe3bf747660-host-slash" (OuterVolumeSpecName: "host-slash") pod "0589a204-c98c-417d-8256-bfe3bf747660" (UID: "0589a204-c98c-417d-8256-bfe3bf747660"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.809973 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0589a204-c98c-417d-8256-bfe3bf747660-etc-openvswitch\") pod \"0589a204-c98c-417d-8256-bfe3bf747660\" (UID: \"0589a204-c98c-417d-8256-bfe3bf747660\") " Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.810025 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/0589a204-c98c-417d-8256-bfe3bf747660-env-overrides\") pod \"0589a204-c98c-417d-8256-bfe3bf747660\" (UID: \"0589a204-c98c-417d-8256-bfe3bf747660\") " Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.810064 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/0589a204-c98c-417d-8256-bfe3bf747660-ovnkube-script-lib\") pod \"0589a204-c98c-417d-8256-bfe3bf747660\" (UID: \"0589a204-c98c-417d-8256-bfe3bf747660\") " Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.810087 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/0589a204-c98c-417d-8256-bfe3bf747660-ovn-node-metrics-cert\") pod \"0589a204-c98c-417d-8256-bfe3bf747660\" (UID: \"0589a204-c98c-417d-8256-bfe3bf747660\") " Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.810321 4860 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/0589a204-c98c-417d-8256-bfe3bf747660-log-socket\") on node \"crc\" DevicePath \"\"" Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.809935 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0589a204-c98c-417d-8256-bfe3bf747660-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "0589a204-c98c-417d-8256-bfe3bf747660" (UID: "0589a204-c98c-417d-8256-bfe3bf747660"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.809957 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0589a204-c98c-417d-8256-bfe3bf747660-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "0589a204-c98c-417d-8256-bfe3bf747660" (UID: "0589a204-c98c-417d-8256-bfe3bf747660"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.809993 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0589a204-c98c-417d-8256-bfe3bf747660-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "0589a204-c98c-417d-8256-bfe3bf747660" (UID: "0589a204-c98c-417d-8256-bfe3bf747660"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.810442 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0589a204-c98c-417d-8256-bfe3bf747660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "0589a204-c98c-417d-8256-bfe3bf747660" (UID: "0589a204-c98c-417d-8256-bfe3bf747660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.810458 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0589a204-c98c-417d-8256-bfe3bf747660-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "0589a204-c98c-417d-8256-bfe3bf747660" (UID: "0589a204-c98c-417d-8256-bfe3bf747660"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.810337 4860 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/0589a204-c98c-417d-8256-bfe3bf747660-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.810516 4860 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/0589a204-c98c-417d-8256-bfe3bf747660-host-kubelet\") on node \"crc\" DevicePath \"\"" Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.810527 4860 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/0589a204-c98c-417d-8256-bfe3bf747660-node-log\") on node \"crc\" DevicePath \"\"" Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.810537 4860 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/0589a204-c98c-417d-8256-bfe3bf747660-host-run-netns\") on node \"crc\" DevicePath \"\"" Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.810546 4860 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/0589a204-c98c-417d-8256-bfe3bf747660-host-cni-bin\") on node \"crc\" DevicePath \"\"" Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.810555 4860 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0589a204-c98c-417d-8256-bfe3bf747660-run-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.810563 4860 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/0589a204-c98c-417d-8256-bfe3bf747660-host-cni-netd\") on node \"crc\" DevicePath \"\"" Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.810573 4860 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0589a204-c98c-417d-8256-bfe3bf747660-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.810582 4860 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/0589a204-c98c-417d-8256-bfe3bf747660-host-slash\") on node \"crc\" DevicePath \"\"" Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.810874 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0589a204-c98c-417d-8256-bfe3bf747660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "0589a204-c98c-417d-8256-bfe3bf747660" (UID: "0589a204-c98c-417d-8256-bfe3bf747660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.811018 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0589a204-c98c-417d-8256-bfe3bf747660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "0589a204-c98c-417d-8256-bfe3bf747660" (UID: "0589a204-c98c-417d-8256-bfe3bf747660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.814755 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0589a204-c98c-417d-8256-bfe3bf747660-kube-api-access-tftsj" (OuterVolumeSpecName: "kube-api-access-tftsj") pod "0589a204-c98c-417d-8256-bfe3bf747660" (UID: "0589a204-c98c-417d-8256-bfe3bf747660"). InnerVolumeSpecName "kube-api-access-tftsj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.814958 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0589a204-c98c-417d-8256-bfe3bf747660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "0589a204-c98c-417d-8256-bfe3bf747660" (UID: "0589a204-c98c-417d-8256-bfe3bf747660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.822878 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0589a204-c98c-417d-8256-bfe3bf747660-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "0589a204-c98c-417d-8256-bfe3bf747660" (UID: "0589a204-c98c-417d-8256-bfe3bf747660"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.911398 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/430aa217-e9f9-445b-855a-69e8c797da38-node-log\") pod \"ovnkube-node-w2n7w\" (UID: \"430aa217-e9f9-445b-855a-69e8c797da38\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2n7w" Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.911464 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/430aa217-e9f9-445b-855a-69e8c797da38-host-cni-netd\") pod \"ovnkube-node-w2n7w\" (UID: \"430aa217-e9f9-445b-855a-69e8c797da38\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2n7w" Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.911484 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/430aa217-e9f9-445b-855a-69e8c797da38-host-run-netns\") pod \"ovnkube-node-w2n7w\" (UID: \"430aa217-e9f9-445b-855a-69e8c797da38\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2n7w" Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.911501 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/430aa217-e9f9-445b-855a-69e8c797da38-log-socket\") pod \"ovnkube-node-w2n7w\" (UID: \"430aa217-e9f9-445b-855a-69e8c797da38\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2n7w" Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.911516 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/430aa217-e9f9-445b-855a-69e8c797da38-run-systemd\") pod \"ovnkube-node-w2n7w\" (UID: \"430aa217-e9f9-445b-855a-69e8c797da38\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2n7w" Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.911535 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/430aa217-e9f9-445b-855a-69e8c797da38-env-overrides\") pod \"ovnkube-node-w2n7w\" (UID: \"430aa217-e9f9-445b-855a-69e8c797da38\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2n7w" Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.911555 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/430aa217-e9f9-445b-855a-69e8c797da38-var-lib-openvswitch\") pod \"ovnkube-node-w2n7w\" (UID: \"430aa217-e9f9-445b-855a-69e8c797da38\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2n7w" Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.911578 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/430aa217-e9f9-445b-855a-69e8c797da38-host-run-ovn-kubernetes\") pod \"ovnkube-node-w2n7w\" (UID: \"430aa217-e9f9-445b-855a-69e8c797da38\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2n7w" Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.911709 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/430aa217-e9f9-445b-855a-69e8c797da38-ovnkube-script-lib\") pod \"ovnkube-node-w2n7w\" (UID: \"430aa217-e9f9-445b-855a-69e8c797da38\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2n7w" Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.911729 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/430aa217-e9f9-445b-855a-69e8c797da38-run-ovn\") pod \"ovnkube-node-w2n7w\" (UID: \"430aa217-e9f9-445b-855a-69e8c797da38\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2n7w" Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.911746 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/430aa217-e9f9-445b-855a-69e8c797da38-ovn-node-metrics-cert\") pod \"ovnkube-node-w2n7w\" (UID: \"430aa217-e9f9-445b-855a-69e8c797da38\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2n7w" Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.911764 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/430aa217-e9f9-445b-855a-69e8c797da38-host-kubelet\") pod \"ovnkube-node-w2n7w\" (UID: \"430aa217-e9f9-445b-855a-69e8c797da38\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2n7w" Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.911780 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/430aa217-e9f9-445b-855a-69e8c797da38-ovnkube-config\") pod \"ovnkube-node-w2n7w\" (UID: \"430aa217-e9f9-445b-855a-69e8c797da38\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2n7w" Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.911803 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/430aa217-e9f9-445b-855a-69e8c797da38-etc-openvswitch\") pod \"ovnkube-node-w2n7w\" (UID: \"430aa217-e9f9-445b-855a-69e8c797da38\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2n7w" Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.911821 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/430aa217-e9f9-445b-855a-69e8c797da38-host-cni-bin\") pod \"ovnkube-node-w2n7w\" (UID: \"430aa217-e9f9-445b-855a-69e8c797da38\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2n7w" Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.911841 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/430aa217-e9f9-445b-855a-69e8c797da38-run-openvswitch\") pod \"ovnkube-node-w2n7w\" (UID: \"430aa217-e9f9-445b-855a-69e8c797da38\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2n7w" Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.911859 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/430aa217-e9f9-445b-855a-69e8c797da38-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-w2n7w\" (UID: \"430aa217-e9f9-445b-855a-69e8c797da38\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2n7w" Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.911878 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/430aa217-e9f9-445b-855a-69e8c797da38-host-slash\") pod \"ovnkube-node-w2n7w\" (UID: \"430aa217-e9f9-445b-855a-69e8c797da38\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2n7w" Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.911899 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-79jkc\" (UniqueName: \"kubernetes.io/projected/430aa217-e9f9-445b-855a-69e8c797da38-kube-api-access-79jkc\") pod \"ovnkube-node-w2n7w\" (UID: \"430aa217-e9f9-445b-855a-69e8c797da38\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2n7w" Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.911916 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/430aa217-e9f9-445b-855a-69e8c797da38-systemd-units\") pod \"ovnkube-node-w2n7w\" (UID: \"430aa217-e9f9-445b-855a-69e8c797da38\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2n7w" Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.911950 4860 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/0589a204-c98c-417d-8256-bfe3bf747660-run-systemd\") on node \"crc\" DevicePath \"\"" Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.911962 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tftsj\" (UniqueName: \"kubernetes.io/projected/0589a204-c98c-417d-8256-bfe3bf747660-kube-api-access-tftsj\") on node \"crc\" DevicePath \"\"" Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.911973 4860 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0589a204-c98c-417d-8256-bfe3bf747660-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.911982 4860 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/0589a204-c98c-417d-8256-bfe3bf747660-systemd-units\") on node \"crc\" DevicePath \"\"" Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.911990 4860 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/0589a204-c98c-417d-8256-bfe3bf747660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.911998 4860 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0589a204-c98c-417d-8256-bfe3bf747660-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.912009 4860 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0589a204-c98c-417d-8256-bfe3bf747660-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.912017 4860 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/0589a204-c98c-417d-8256-bfe3bf747660-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.912025 4860 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/0589a204-c98c-417d-8256-bfe3bf747660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 11 08:22:26 crc kubenswrapper[4860]: I1211 08:22:26.912033 4860 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/0589a204-c98c-417d-8256-bfe3bf747660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.013662 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/430aa217-e9f9-445b-855a-69e8c797da38-host-cni-netd\") pod \"ovnkube-node-w2n7w\" (UID: \"430aa217-e9f9-445b-855a-69e8c797da38\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2n7w" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.013735 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/430aa217-e9f9-445b-855a-69e8c797da38-host-run-netns\") pod \"ovnkube-node-w2n7w\" (UID: \"430aa217-e9f9-445b-855a-69e8c797da38\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2n7w" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.013759 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/430aa217-e9f9-445b-855a-69e8c797da38-log-socket\") pod \"ovnkube-node-w2n7w\" (UID: \"430aa217-e9f9-445b-855a-69e8c797da38\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2n7w" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.013785 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/430aa217-e9f9-445b-855a-69e8c797da38-run-systemd\") pod \"ovnkube-node-w2n7w\" (UID: \"430aa217-e9f9-445b-855a-69e8c797da38\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2n7w" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.013797 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/430aa217-e9f9-445b-855a-69e8c797da38-host-run-netns\") pod \"ovnkube-node-w2n7w\" (UID: \"430aa217-e9f9-445b-855a-69e8c797da38\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2n7w" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.013811 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/430aa217-e9f9-445b-855a-69e8c797da38-env-overrides\") pod \"ovnkube-node-w2n7w\" (UID: \"430aa217-e9f9-445b-855a-69e8c797da38\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2n7w" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.013883 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/430aa217-e9f9-445b-855a-69e8c797da38-run-systemd\") pod \"ovnkube-node-w2n7w\" (UID: \"430aa217-e9f9-445b-855a-69e8c797da38\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2n7w" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.013758 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/430aa217-e9f9-445b-855a-69e8c797da38-host-cni-netd\") pod \"ovnkube-node-w2n7w\" (UID: \"430aa217-e9f9-445b-855a-69e8c797da38\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2n7w" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.013892 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/430aa217-e9f9-445b-855a-69e8c797da38-var-lib-openvswitch\") pod \"ovnkube-node-w2n7w\" (UID: \"430aa217-e9f9-445b-855a-69e8c797da38\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2n7w" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.013928 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/430aa217-e9f9-445b-855a-69e8c797da38-var-lib-openvswitch\") pod \"ovnkube-node-w2n7w\" (UID: \"430aa217-e9f9-445b-855a-69e8c797da38\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2n7w" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.013941 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/430aa217-e9f9-445b-855a-69e8c797da38-host-run-ovn-kubernetes\") pod \"ovnkube-node-w2n7w\" (UID: \"430aa217-e9f9-445b-855a-69e8c797da38\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2n7w" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.013903 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/430aa217-e9f9-445b-855a-69e8c797da38-log-socket\") pod \"ovnkube-node-w2n7w\" (UID: \"430aa217-e9f9-445b-855a-69e8c797da38\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2n7w" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.014027 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/430aa217-e9f9-445b-855a-69e8c797da38-host-run-ovn-kubernetes\") pod \"ovnkube-node-w2n7w\" (UID: \"430aa217-e9f9-445b-855a-69e8c797da38\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2n7w" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.013963 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/430aa217-e9f9-445b-855a-69e8c797da38-ovnkube-script-lib\") pod \"ovnkube-node-w2n7w\" (UID: \"430aa217-e9f9-445b-855a-69e8c797da38\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2n7w" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.014076 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/430aa217-e9f9-445b-855a-69e8c797da38-run-ovn\") pod \"ovnkube-node-w2n7w\" (UID: \"430aa217-e9f9-445b-855a-69e8c797da38\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2n7w" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.014117 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/430aa217-e9f9-445b-855a-69e8c797da38-ovn-node-metrics-cert\") pod \"ovnkube-node-w2n7w\" (UID: \"430aa217-e9f9-445b-855a-69e8c797da38\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2n7w" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.014148 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/430aa217-e9f9-445b-855a-69e8c797da38-host-kubelet\") pod \"ovnkube-node-w2n7w\" (UID: \"430aa217-e9f9-445b-855a-69e8c797da38\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2n7w" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.014179 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/430aa217-e9f9-445b-855a-69e8c797da38-ovnkube-config\") pod \"ovnkube-node-w2n7w\" (UID: \"430aa217-e9f9-445b-855a-69e8c797da38\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2n7w" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.014543 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/430aa217-e9f9-445b-855a-69e8c797da38-etc-openvswitch\") pod \"ovnkube-node-w2n7w\" (UID: \"430aa217-e9f9-445b-855a-69e8c797da38\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2n7w" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.014586 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/430aa217-e9f9-445b-855a-69e8c797da38-host-cni-bin\") pod \"ovnkube-node-w2n7w\" (UID: \"430aa217-e9f9-445b-855a-69e8c797da38\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2n7w" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.014625 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/430aa217-e9f9-445b-855a-69e8c797da38-run-openvswitch\") pod \"ovnkube-node-w2n7w\" (UID: \"430aa217-e9f9-445b-855a-69e8c797da38\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2n7w" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.014681 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/430aa217-e9f9-445b-855a-69e8c797da38-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-w2n7w\" (UID: \"430aa217-e9f9-445b-855a-69e8c797da38\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2n7w" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.014718 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/430aa217-e9f9-445b-855a-69e8c797da38-host-slash\") pod \"ovnkube-node-w2n7w\" (UID: \"430aa217-e9f9-445b-855a-69e8c797da38\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2n7w" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.014176 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/430aa217-e9f9-445b-855a-69e8c797da38-run-ovn\") pod \"ovnkube-node-w2n7w\" (UID: \"430aa217-e9f9-445b-855a-69e8c797da38\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2n7w" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.014753 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-79jkc\" (UniqueName: \"kubernetes.io/projected/430aa217-e9f9-445b-855a-69e8c797da38-kube-api-access-79jkc\") pod \"ovnkube-node-w2n7w\" (UID: \"430aa217-e9f9-445b-855a-69e8c797da38\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2n7w" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.014206 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/430aa217-e9f9-445b-855a-69e8c797da38-host-kubelet\") pod \"ovnkube-node-w2n7w\" (UID: \"430aa217-e9f9-445b-855a-69e8c797da38\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2n7w" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.014790 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/430aa217-e9f9-445b-855a-69e8c797da38-systemd-units\") pod \"ovnkube-node-w2n7w\" (UID: \"430aa217-e9f9-445b-855a-69e8c797da38\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2n7w" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.014810 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/430aa217-e9f9-445b-855a-69e8c797da38-etc-openvswitch\") pod \"ovnkube-node-w2n7w\" (UID: \"430aa217-e9f9-445b-855a-69e8c797da38\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2n7w" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.014832 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/430aa217-e9f9-445b-855a-69e8c797da38-node-log\") pod \"ovnkube-node-w2n7w\" (UID: \"430aa217-e9f9-445b-855a-69e8c797da38\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2n7w" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.014855 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/430aa217-e9f9-445b-855a-69e8c797da38-run-openvswitch\") pod \"ovnkube-node-w2n7w\" (UID: \"430aa217-e9f9-445b-855a-69e8c797da38\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2n7w" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.014871 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/430aa217-e9f9-445b-855a-69e8c797da38-host-slash\") pod \"ovnkube-node-w2n7w\" (UID: \"430aa217-e9f9-445b-855a-69e8c797da38\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2n7w" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.014908 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/430aa217-e9f9-445b-855a-69e8c797da38-node-log\") pod \"ovnkube-node-w2n7w\" (UID: \"430aa217-e9f9-445b-855a-69e8c797da38\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2n7w" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.014914 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/430aa217-e9f9-445b-855a-69e8c797da38-host-cni-bin\") pod \"ovnkube-node-w2n7w\" (UID: \"430aa217-e9f9-445b-855a-69e8c797da38\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2n7w" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.014917 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/430aa217-e9f9-445b-855a-69e8c797da38-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-w2n7w\" (UID: \"430aa217-e9f9-445b-855a-69e8c797da38\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2n7w" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.014948 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/430aa217-e9f9-445b-855a-69e8c797da38-systemd-units\") pod \"ovnkube-node-w2n7w\" (UID: \"430aa217-e9f9-445b-855a-69e8c797da38\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2n7w" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.015176 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/430aa217-e9f9-445b-855a-69e8c797da38-env-overrides\") pod \"ovnkube-node-w2n7w\" (UID: \"430aa217-e9f9-445b-855a-69e8c797da38\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2n7w" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.015278 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/430aa217-e9f9-445b-855a-69e8c797da38-ovnkube-script-lib\") pod \"ovnkube-node-w2n7w\" (UID: \"430aa217-e9f9-445b-855a-69e8c797da38\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2n7w" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.015501 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/430aa217-e9f9-445b-855a-69e8c797da38-ovnkube-config\") pod \"ovnkube-node-w2n7w\" (UID: \"430aa217-e9f9-445b-855a-69e8c797da38\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2n7w" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.017034 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/430aa217-e9f9-445b-855a-69e8c797da38-ovn-node-metrics-cert\") pod \"ovnkube-node-w2n7w\" (UID: \"430aa217-e9f9-445b-855a-69e8c797da38\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2n7w" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.034013 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-79jkc\" (UniqueName: \"kubernetes.io/projected/430aa217-e9f9-445b-855a-69e8c797da38-kube-api-access-79jkc\") pod \"ovnkube-node-w2n7w\" (UID: \"430aa217-e9f9-445b-855a-69e8c797da38\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2n7w" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.042806 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-w2n7w" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.548475 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-xvnqp_0589a204-c98c-417d-8256-bfe3bf747660/ovnkube-controller/3.log" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.551214 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-xvnqp_0589a204-c98c-417d-8256-bfe3bf747660/ovn-acl-logging/0.log" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.551783 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-xvnqp_0589a204-c98c-417d-8256-bfe3bf747660/ovn-controller/0.log" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.552173 4860 generic.go:334] "Generic (PLEG): container finished" podID="0589a204-c98c-417d-8256-bfe3bf747660" containerID="4513cf6bc605dd49d55acb42fa24068e7e61a7b40104f7c5a17dc78cc5694388" exitCode=0 Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.552203 4860 generic.go:334] "Generic (PLEG): container finished" podID="0589a204-c98c-417d-8256-bfe3bf747660" containerID="bb80d9928e8349fc5963227d135355ad7ba820d4fd89877af601d005ca03334b" exitCode=0 Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.552214 4860 generic.go:334] "Generic (PLEG): container finished" podID="0589a204-c98c-417d-8256-bfe3bf747660" containerID="6dfcc4370edaae36a5a32b0a9c29290c50f8e4f427e4e147dc3c779beaa8b780" exitCode=0 Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.552258 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.552261 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" event={"ID":"0589a204-c98c-417d-8256-bfe3bf747660","Type":"ContainerDied","Data":"4513cf6bc605dd49d55acb42fa24068e7e61a7b40104f7c5a17dc78cc5694388"} Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.552321 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" event={"ID":"0589a204-c98c-417d-8256-bfe3bf747660","Type":"ContainerDied","Data":"bb80d9928e8349fc5963227d135355ad7ba820d4fd89877af601d005ca03334b"} Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.552334 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" event={"ID":"0589a204-c98c-417d-8256-bfe3bf747660","Type":"ContainerDied","Data":"6dfcc4370edaae36a5a32b0a9c29290c50f8e4f427e4e147dc3c779beaa8b780"} Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.552347 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xvnqp" event={"ID":"0589a204-c98c-417d-8256-bfe3bf747660","Type":"ContainerDied","Data":"7604d81409ead4ab386e79a3646e8237268fa238b5c89b25f14fdcfbc5e2b819"} Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.552389 4860 scope.go:117] "RemoveContainer" containerID="4513cf6bc605dd49d55acb42fa24068e7e61a7b40104f7c5a17dc78cc5694388" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.554741 4860 generic.go:334] "Generic (PLEG): container finished" podID="430aa217-e9f9-445b-855a-69e8c797da38" containerID="44af91198753749c2c5c0a8202f0fdc5cd2448bf732e7f1965048bb9a444449d" exitCode=0 Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.554813 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w2n7w" event={"ID":"430aa217-e9f9-445b-855a-69e8c797da38","Type":"ContainerDied","Data":"44af91198753749c2c5c0a8202f0fdc5cd2448bf732e7f1965048bb9a444449d"} Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.554844 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w2n7w" event={"ID":"430aa217-e9f9-445b-855a-69e8c797da38","Type":"ContainerStarted","Data":"7b19007b387ba58e0cd558c8808f967777fa6373ebc780be58a428310a41c553"} Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.568914 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-pznwl_5310c1fc-66c6-40aa-b1bf-5a59a2410f9c/kube-multus/2.log" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.582799 4860 scope.go:117] "RemoveContainer" containerID="d9e83cda3e40d325604114a37fe6e8b707ac1f91a034b0d979c70b5f4f24f2e7" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.619880 4860 scope.go:117] "RemoveContainer" containerID="bb80d9928e8349fc5963227d135355ad7ba820d4fd89877af601d005ca03334b" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.629620 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-xvnqp"] Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.633155 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-xvnqp"] Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.651905 4860 scope.go:117] "RemoveContainer" containerID="6dfcc4370edaae36a5a32b0a9c29290c50f8e4f427e4e147dc3c779beaa8b780" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.679382 4860 scope.go:117] "RemoveContainer" containerID="d856ae95e9077524c518d82d10fb6b93b7fbcaaf2b720ffafe6f7fae58bd7caa" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.695904 4860 scope.go:117] "RemoveContainer" containerID="52bb8ade314afb9494f4e026d0af2aea7b1970d64dc66102f55e257ae4678154" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.726199 4860 scope.go:117] "RemoveContainer" containerID="3333b4346f59b04017e32d61f457ba7ff457166731128f339d7987ae6b56b06d" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.747104 4860 scope.go:117] "RemoveContainer" containerID="1426084ac580895c42354630910dd8c8e3a48ddd25f22abce1d28dc3de332a63" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.761053 4860 scope.go:117] "RemoveContainer" containerID="c9984d5d242bb258405efe91fc9caf9e7ea592834445df2e78add9553a921c76" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.795535 4860 scope.go:117] "RemoveContainer" containerID="17cbcb49c06f3a9059cd5a0f05eedb8c785da30a3e502e31bae2f57b3486cb14" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.814876 4860 scope.go:117] "RemoveContainer" containerID="4513cf6bc605dd49d55acb42fa24068e7e61a7b40104f7c5a17dc78cc5694388" Dec 11 08:22:27 crc kubenswrapper[4860]: E1211 08:22:27.818028 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4513cf6bc605dd49d55acb42fa24068e7e61a7b40104f7c5a17dc78cc5694388\": container with ID starting with 4513cf6bc605dd49d55acb42fa24068e7e61a7b40104f7c5a17dc78cc5694388 not found: ID does not exist" containerID="4513cf6bc605dd49d55acb42fa24068e7e61a7b40104f7c5a17dc78cc5694388" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.818089 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4513cf6bc605dd49d55acb42fa24068e7e61a7b40104f7c5a17dc78cc5694388"} err="failed to get container status \"4513cf6bc605dd49d55acb42fa24068e7e61a7b40104f7c5a17dc78cc5694388\": rpc error: code = NotFound desc = could not find container \"4513cf6bc605dd49d55acb42fa24068e7e61a7b40104f7c5a17dc78cc5694388\": container with ID starting with 4513cf6bc605dd49d55acb42fa24068e7e61a7b40104f7c5a17dc78cc5694388 not found: ID does not exist" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.818130 4860 scope.go:117] "RemoveContainer" containerID="d9e83cda3e40d325604114a37fe6e8b707ac1f91a034b0d979c70b5f4f24f2e7" Dec 11 08:22:27 crc kubenswrapper[4860]: E1211 08:22:27.818782 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d9e83cda3e40d325604114a37fe6e8b707ac1f91a034b0d979c70b5f4f24f2e7\": container with ID starting with d9e83cda3e40d325604114a37fe6e8b707ac1f91a034b0d979c70b5f4f24f2e7 not found: ID does not exist" containerID="d9e83cda3e40d325604114a37fe6e8b707ac1f91a034b0d979c70b5f4f24f2e7" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.818835 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d9e83cda3e40d325604114a37fe6e8b707ac1f91a034b0d979c70b5f4f24f2e7"} err="failed to get container status \"d9e83cda3e40d325604114a37fe6e8b707ac1f91a034b0d979c70b5f4f24f2e7\": rpc error: code = NotFound desc = could not find container \"d9e83cda3e40d325604114a37fe6e8b707ac1f91a034b0d979c70b5f4f24f2e7\": container with ID starting with d9e83cda3e40d325604114a37fe6e8b707ac1f91a034b0d979c70b5f4f24f2e7 not found: ID does not exist" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.818865 4860 scope.go:117] "RemoveContainer" containerID="bb80d9928e8349fc5963227d135355ad7ba820d4fd89877af601d005ca03334b" Dec 11 08:22:27 crc kubenswrapper[4860]: E1211 08:22:27.819198 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bb80d9928e8349fc5963227d135355ad7ba820d4fd89877af601d005ca03334b\": container with ID starting with bb80d9928e8349fc5963227d135355ad7ba820d4fd89877af601d005ca03334b not found: ID does not exist" containerID="bb80d9928e8349fc5963227d135355ad7ba820d4fd89877af601d005ca03334b" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.819245 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bb80d9928e8349fc5963227d135355ad7ba820d4fd89877af601d005ca03334b"} err="failed to get container status \"bb80d9928e8349fc5963227d135355ad7ba820d4fd89877af601d005ca03334b\": rpc error: code = NotFound desc = could not find container \"bb80d9928e8349fc5963227d135355ad7ba820d4fd89877af601d005ca03334b\": container with ID starting with bb80d9928e8349fc5963227d135355ad7ba820d4fd89877af601d005ca03334b not found: ID does not exist" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.819262 4860 scope.go:117] "RemoveContainer" containerID="6dfcc4370edaae36a5a32b0a9c29290c50f8e4f427e4e147dc3c779beaa8b780" Dec 11 08:22:27 crc kubenswrapper[4860]: E1211 08:22:27.819553 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6dfcc4370edaae36a5a32b0a9c29290c50f8e4f427e4e147dc3c779beaa8b780\": container with ID starting with 6dfcc4370edaae36a5a32b0a9c29290c50f8e4f427e4e147dc3c779beaa8b780 not found: ID does not exist" containerID="6dfcc4370edaae36a5a32b0a9c29290c50f8e4f427e4e147dc3c779beaa8b780" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.819580 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6dfcc4370edaae36a5a32b0a9c29290c50f8e4f427e4e147dc3c779beaa8b780"} err="failed to get container status \"6dfcc4370edaae36a5a32b0a9c29290c50f8e4f427e4e147dc3c779beaa8b780\": rpc error: code = NotFound desc = could not find container \"6dfcc4370edaae36a5a32b0a9c29290c50f8e4f427e4e147dc3c779beaa8b780\": container with ID starting with 6dfcc4370edaae36a5a32b0a9c29290c50f8e4f427e4e147dc3c779beaa8b780 not found: ID does not exist" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.819595 4860 scope.go:117] "RemoveContainer" containerID="d856ae95e9077524c518d82d10fb6b93b7fbcaaf2b720ffafe6f7fae58bd7caa" Dec 11 08:22:27 crc kubenswrapper[4860]: E1211 08:22:27.819845 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d856ae95e9077524c518d82d10fb6b93b7fbcaaf2b720ffafe6f7fae58bd7caa\": container with ID starting with d856ae95e9077524c518d82d10fb6b93b7fbcaaf2b720ffafe6f7fae58bd7caa not found: ID does not exist" containerID="d856ae95e9077524c518d82d10fb6b93b7fbcaaf2b720ffafe6f7fae58bd7caa" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.819890 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d856ae95e9077524c518d82d10fb6b93b7fbcaaf2b720ffafe6f7fae58bd7caa"} err="failed to get container status \"d856ae95e9077524c518d82d10fb6b93b7fbcaaf2b720ffafe6f7fae58bd7caa\": rpc error: code = NotFound desc = could not find container \"d856ae95e9077524c518d82d10fb6b93b7fbcaaf2b720ffafe6f7fae58bd7caa\": container with ID starting with d856ae95e9077524c518d82d10fb6b93b7fbcaaf2b720ffafe6f7fae58bd7caa not found: ID does not exist" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.819905 4860 scope.go:117] "RemoveContainer" containerID="52bb8ade314afb9494f4e026d0af2aea7b1970d64dc66102f55e257ae4678154" Dec 11 08:22:27 crc kubenswrapper[4860]: E1211 08:22:27.820133 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"52bb8ade314afb9494f4e026d0af2aea7b1970d64dc66102f55e257ae4678154\": container with ID starting with 52bb8ade314afb9494f4e026d0af2aea7b1970d64dc66102f55e257ae4678154 not found: ID does not exist" containerID="52bb8ade314afb9494f4e026d0af2aea7b1970d64dc66102f55e257ae4678154" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.820164 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"52bb8ade314afb9494f4e026d0af2aea7b1970d64dc66102f55e257ae4678154"} err="failed to get container status \"52bb8ade314afb9494f4e026d0af2aea7b1970d64dc66102f55e257ae4678154\": rpc error: code = NotFound desc = could not find container \"52bb8ade314afb9494f4e026d0af2aea7b1970d64dc66102f55e257ae4678154\": container with ID starting with 52bb8ade314afb9494f4e026d0af2aea7b1970d64dc66102f55e257ae4678154 not found: ID does not exist" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.820181 4860 scope.go:117] "RemoveContainer" containerID="3333b4346f59b04017e32d61f457ba7ff457166731128f339d7987ae6b56b06d" Dec 11 08:22:27 crc kubenswrapper[4860]: E1211 08:22:27.820368 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3333b4346f59b04017e32d61f457ba7ff457166731128f339d7987ae6b56b06d\": container with ID starting with 3333b4346f59b04017e32d61f457ba7ff457166731128f339d7987ae6b56b06d not found: ID does not exist" containerID="3333b4346f59b04017e32d61f457ba7ff457166731128f339d7987ae6b56b06d" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.820393 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3333b4346f59b04017e32d61f457ba7ff457166731128f339d7987ae6b56b06d"} err="failed to get container status \"3333b4346f59b04017e32d61f457ba7ff457166731128f339d7987ae6b56b06d\": rpc error: code = NotFound desc = could not find container \"3333b4346f59b04017e32d61f457ba7ff457166731128f339d7987ae6b56b06d\": container with ID starting with 3333b4346f59b04017e32d61f457ba7ff457166731128f339d7987ae6b56b06d not found: ID does not exist" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.820418 4860 scope.go:117] "RemoveContainer" containerID="1426084ac580895c42354630910dd8c8e3a48ddd25f22abce1d28dc3de332a63" Dec 11 08:22:27 crc kubenswrapper[4860]: E1211 08:22:27.820842 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1426084ac580895c42354630910dd8c8e3a48ddd25f22abce1d28dc3de332a63\": container with ID starting with 1426084ac580895c42354630910dd8c8e3a48ddd25f22abce1d28dc3de332a63 not found: ID does not exist" containerID="1426084ac580895c42354630910dd8c8e3a48ddd25f22abce1d28dc3de332a63" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.820893 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1426084ac580895c42354630910dd8c8e3a48ddd25f22abce1d28dc3de332a63"} err="failed to get container status \"1426084ac580895c42354630910dd8c8e3a48ddd25f22abce1d28dc3de332a63\": rpc error: code = NotFound desc = could not find container \"1426084ac580895c42354630910dd8c8e3a48ddd25f22abce1d28dc3de332a63\": container with ID starting with 1426084ac580895c42354630910dd8c8e3a48ddd25f22abce1d28dc3de332a63 not found: ID does not exist" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.820908 4860 scope.go:117] "RemoveContainer" containerID="c9984d5d242bb258405efe91fc9caf9e7ea592834445df2e78add9553a921c76" Dec 11 08:22:27 crc kubenswrapper[4860]: E1211 08:22:27.821159 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c9984d5d242bb258405efe91fc9caf9e7ea592834445df2e78add9553a921c76\": container with ID starting with c9984d5d242bb258405efe91fc9caf9e7ea592834445df2e78add9553a921c76 not found: ID does not exist" containerID="c9984d5d242bb258405efe91fc9caf9e7ea592834445df2e78add9553a921c76" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.821183 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c9984d5d242bb258405efe91fc9caf9e7ea592834445df2e78add9553a921c76"} err="failed to get container status \"c9984d5d242bb258405efe91fc9caf9e7ea592834445df2e78add9553a921c76\": rpc error: code = NotFound desc = could not find container \"c9984d5d242bb258405efe91fc9caf9e7ea592834445df2e78add9553a921c76\": container with ID starting with c9984d5d242bb258405efe91fc9caf9e7ea592834445df2e78add9553a921c76 not found: ID does not exist" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.821203 4860 scope.go:117] "RemoveContainer" containerID="17cbcb49c06f3a9059cd5a0f05eedb8c785da30a3e502e31bae2f57b3486cb14" Dec 11 08:22:27 crc kubenswrapper[4860]: E1211 08:22:27.821391 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"17cbcb49c06f3a9059cd5a0f05eedb8c785da30a3e502e31bae2f57b3486cb14\": container with ID starting with 17cbcb49c06f3a9059cd5a0f05eedb8c785da30a3e502e31bae2f57b3486cb14 not found: ID does not exist" containerID="17cbcb49c06f3a9059cd5a0f05eedb8c785da30a3e502e31bae2f57b3486cb14" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.821415 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"17cbcb49c06f3a9059cd5a0f05eedb8c785da30a3e502e31bae2f57b3486cb14"} err="failed to get container status \"17cbcb49c06f3a9059cd5a0f05eedb8c785da30a3e502e31bae2f57b3486cb14\": rpc error: code = NotFound desc = could not find container \"17cbcb49c06f3a9059cd5a0f05eedb8c785da30a3e502e31bae2f57b3486cb14\": container with ID starting with 17cbcb49c06f3a9059cd5a0f05eedb8c785da30a3e502e31bae2f57b3486cb14 not found: ID does not exist" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.821431 4860 scope.go:117] "RemoveContainer" containerID="4513cf6bc605dd49d55acb42fa24068e7e61a7b40104f7c5a17dc78cc5694388" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.821807 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4513cf6bc605dd49d55acb42fa24068e7e61a7b40104f7c5a17dc78cc5694388"} err="failed to get container status \"4513cf6bc605dd49d55acb42fa24068e7e61a7b40104f7c5a17dc78cc5694388\": rpc error: code = NotFound desc = could not find container \"4513cf6bc605dd49d55acb42fa24068e7e61a7b40104f7c5a17dc78cc5694388\": container with ID starting with 4513cf6bc605dd49d55acb42fa24068e7e61a7b40104f7c5a17dc78cc5694388 not found: ID does not exist" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.821836 4860 scope.go:117] "RemoveContainer" containerID="d9e83cda3e40d325604114a37fe6e8b707ac1f91a034b0d979c70b5f4f24f2e7" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.822109 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d9e83cda3e40d325604114a37fe6e8b707ac1f91a034b0d979c70b5f4f24f2e7"} err="failed to get container status \"d9e83cda3e40d325604114a37fe6e8b707ac1f91a034b0d979c70b5f4f24f2e7\": rpc error: code = NotFound desc = could not find container \"d9e83cda3e40d325604114a37fe6e8b707ac1f91a034b0d979c70b5f4f24f2e7\": container with ID starting with d9e83cda3e40d325604114a37fe6e8b707ac1f91a034b0d979c70b5f4f24f2e7 not found: ID does not exist" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.822133 4860 scope.go:117] "RemoveContainer" containerID="bb80d9928e8349fc5963227d135355ad7ba820d4fd89877af601d005ca03334b" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.822446 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bb80d9928e8349fc5963227d135355ad7ba820d4fd89877af601d005ca03334b"} err="failed to get container status \"bb80d9928e8349fc5963227d135355ad7ba820d4fd89877af601d005ca03334b\": rpc error: code = NotFound desc = could not find container \"bb80d9928e8349fc5963227d135355ad7ba820d4fd89877af601d005ca03334b\": container with ID starting with bb80d9928e8349fc5963227d135355ad7ba820d4fd89877af601d005ca03334b not found: ID does not exist" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.822475 4860 scope.go:117] "RemoveContainer" containerID="6dfcc4370edaae36a5a32b0a9c29290c50f8e4f427e4e147dc3c779beaa8b780" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.822908 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6dfcc4370edaae36a5a32b0a9c29290c50f8e4f427e4e147dc3c779beaa8b780"} err="failed to get container status \"6dfcc4370edaae36a5a32b0a9c29290c50f8e4f427e4e147dc3c779beaa8b780\": rpc error: code = NotFound desc = could not find container \"6dfcc4370edaae36a5a32b0a9c29290c50f8e4f427e4e147dc3c779beaa8b780\": container with ID starting with 6dfcc4370edaae36a5a32b0a9c29290c50f8e4f427e4e147dc3c779beaa8b780 not found: ID does not exist" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.822987 4860 scope.go:117] "RemoveContainer" containerID="d856ae95e9077524c518d82d10fb6b93b7fbcaaf2b720ffafe6f7fae58bd7caa" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.823335 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d856ae95e9077524c518d82d10fb6b93b7fbcaaf2b720ffafe6f7fae58bd7caa"} err="failed to get container status \"d856ae95e9077524c518d82d10fb6b93b7fbcaaf2b720ffafe6f7fae58bd7caa\": rpc error: code = NotFound desc = could not find container \"d856ae95e9077524c518d82d10fb6b93b7fbcaaf2b720ffafe6f7fae58bd7caa\": container with ID starting with d856ae95e9077524c518d82d10fb6b93b7fbcaaf2b720ffafe6f7fae58bd7caa not found: ID does not exist" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.823362 4860 scope.go:117] "RemoveContainer" containerID="52bb8ade314afb9494f4e026d0af2aea7b1970d64dc66102f55e257ae4678154" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.823788 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"52bb8ade314afb9494f4e026d0af2aea7b1970d64dc66102f55e257ae4678154"} err="failed to get container status \"52bb8ade314afb9494f4e026d0af2aea7b1970d64dc66102f55e257ae4678154\": rpc error: code = NotFound desc = could not find container \"52bb8ade314afb9494f4e026d0af2aea7b1970d64dc66102f55e257ae4678154\": container with ID starting with 52bb8ade314afb9494f4e026d0af2aea7b1970d64dc66102f55e257ae4678154 not found: ID does not exist" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.823809 4860 scope.go:117] "RemoveContainer" containerID="3333b4346f59b04017e32d61f457ba7ff457166731128f339d7987ae6b56b06d" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.824126 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3333b4346f59b04017e32d61f457ba7ff457166731128f339d7987ae6b56b06d"} err="failed to get container status \"3333b4346f59b04017e32d61f457ba7ff457166731128f339d7987ae6b56b06d\": rpc error: code = NotFound desc = could not find container \"3333b4346f59b04017e32d61f457ba7ff457166731128f339d7987ae6b56b06d\": container with ID starting with 3333b4346f59b04017e32d61f457ba7ff457166731128f339d7987ae6b56b06d not found: ID does not exist" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.824153 4860 scope.go:117] "RemoveContainer" containerID="1426084ac580895c42354630910dd8c8e3a48ddd25f22abce1d28dc3de332a63" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.824453 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1426084ac580895c42354630910dd8c8e3a48ddd25f22abce1d28dc3de332a63"} err="failed to get container status \"1426084ac580895c42354630910dd8c8e3a48ddd25f22abce1d28dc3de332a63\": rpc error: code = NotFound desc = could not find container \"1426084ac580895c42354630910dd8c8e3a48ddd25f22abce1d28dc3de332a63\": container with ID starting with 1426084ac580895c42354630910dd8c8e3a48ddd25f22abce1d28dc3de332a63 not found: ID does not exist" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.824473 4860 scope.go:117] "RemoveContainer" containerID="c9984d5d242bb258405efe91fc9caf9e7ea592834445df2e78add9553a921c76" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.824844 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c9984d5d242bb258405efe91fc9caf9e7ea592834445df2e78add9553a921c76"} err="failed to get container status \"c9984d5d242bb258405efe91fc9caf9e7ea592834445df2e78add9553a921c76\": rpc error: code = NotFound desc = could not find container \"c9984d5d242bb258405efe91fc9caf9e7ea592834445df2e78add9553a921c76\": container with ID starting with c9984d5d242bb258405efe91fc9caf9e7ea592834445df2e78add9553a921c76 not found: ID does not exist" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.824870 4860 scope.go:117] "RemoveContainer" containerID="17cbcb49c06f3a9059cd5a0f05eedb8c785da30a3e502e31bae2f57b3486cb14" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.825130 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"17cbcb49c06f3a9059cd5a0f05eedb8c785da30a3e502e31bae2f57b3486cb14"} err="failed to get container status \"17cbcb49c06f3a9059cd5a0f05eedb8c785da30a3e502e31bae2f57b3486cb14\": rpc error: code = NotFound desc = could not find container \"17cbcb49c06f3a9059cd5a0f05eedb8c785da30a3e502e31bae2f57b3486cb14\": container with ID starting with 17cbcb49c06f3a9059cd5a0f05eedb8c785da30a3e502e31bae2f57b3486cb14 not found: ID does not exist" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.825154 4860 scope.go:117] "RemoveContainer" containerID="4513cf6bc605dd49d55acb42fa24068e7e61a7b40104f7c5a17dc78cc5694388" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.825354 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4513cf6bc605dd49d55acb42fa24068e7e61a7b40104f7c5a17dc78cc5694388"} err="failed to get container status \"4513cf6bc605dd49d55acb42fa24068e7e61a7b40104f7c5a17dc78cc5694388\": rpc error: code = NotFound desc = could not find container \"4513cf6bc605dd49d55acb42fa24068e7e61a7b40104f7c5a17dc78cc5694388\": container with ID starting with 4513cf6bc605dd49d55acb42fa24068e7e61a7b40104f7c5a17dc78cc5694388 not found: ID does not exist" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.825379 4860 scope.go:117] "RemoveContainer" containerID="d9e83cda3e40d325604114a37fe6e8b707ac1f91a034b0d979c70b5f4f24f2e7" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.825646 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d9e83cda3e40d325604114a37fe6e8b707ac1f91a034b0d979c70b5f4f24f2e7"} err="failed to get container status \"d9e83cda3e40d325604114a37fe6e8b707ac1f91a034b0d979c70b5f4f24f2e7\": rpc error: code = NotFound desc = could not find container \"d9e83cda3e40d325604114a37fe6e8b707ac1f91a034b0d979c70b5f4f24f2e7\": container with ID starting with d9e83cda3e40d325604114a37fe6e8b707ac1f91a034b0d979c70b5f4f24f2e7 not found: ID does not exist" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.825674 4860 scope.go:117] "RemoveContainer" containerID="bb80d9928e8349fc5963227d135355ad7ba820d4fd89877af601d005ca03334b" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.825938 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bb80d9928e8349fc5963227d135355ad7ba820d4fd89877af601d005ca03334b"} err="failed to get container status \"bb80d9928e8349fc5963227d135355ad7ba820d4fd89877af601d005ca03334b\": rpc error: code = NotFound desc = could not find container \"bb80d9928e8349fc5963227d135355ad7ba820d4fd89877af601d005ca03334b\": container with ID starting with bb80d9928e8349fc5963227d135355ad7ba820d4fd89877af601d005ca03334b not found: ID does not exist" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.825955 4860 scope.go:117] "RemoveContainer" containerID="6dfcc4370edaae36a5a32b0a9c29290c50f8e4f427e4e147dc3c779beaa8b780" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.826220 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6dfcc4370edaae36a5a32b0a9c29290c50f8e4f427e4e147dc3c779beaa8b780"} err="failed to get container status \"6dfcc4370edaae36a5a32b0a9c29290c50f8e4f427e4e147dc3c779beaa8b780\": rpc error: code = NotFound desc = could not find container \"6dfcc4370edaae36a5a32b0a9c29290c50f8e4f427e4e147dc3c779beaa8b780\": container with ID starting with 6dfcc4370edaae36a5a32b0a9c29290c50f8e4f427e4e147dc3c779beaa8b780 not found: ID does not exist" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.826251 4860 scope.go:117] "RemoveContainer" containerID="d856ae95e9077524c518d82d10fb6b93b7fbcaaf2b720ffafe6f7fae58bd7caa" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.826680 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d856ae95e9077524c518d82d10fb6b93b7fbcaaf2b720ffafe6f7fae58bd7caa"} err="failed to get container status \"d856ae95e9077524c518d82d10fb6b93b7fbcaaf2b720ffafe6f7fae58bd7caa\": rpc error: code = NotFound desc = could not find container \"d856ae95e9077524c518d82d10fb6b93b7fbcaaf2b720ffafe6f7fae58bd7caa\": container with ID starting with d856ae95e9077524c518d82d10fb6b93b7fbcaaf2b720ffafe6f7fae58bd7caa not found: ID does not exist" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.826709 4860 scope.go:117] "RemoveContainer" containerID="52bb8ade314afb9494f4e026d0af2aea7b1970d64dc66102f55e257ae4678154" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.827108 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"52bb8ade314afb9494f4e026d0af2aea7b1970d64dc66102f55e257ae4678154"} err="failed to get container status \"52bb8ade314afb9494f4e026d0af2aea7b1970d64dc66102f55e257ae4678154\": rpc error: code = NotFound desc = could not find container \"52bb8ade314afb9494f4e026d0af2aea7b1970d64dc66102f55e257ae4678154\": container with ID starting with 52bb8ade314afb9494f4e026d0af2aea7b1970d64dc66102f55e257ae4678154 not found: ID does not exist" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.827134 4860 scope.go:117] "RemoveContainer" containerID="3333b4346f59b04017e32d61f457ba7ff457166731128f339d7987ae6b56b06d" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.827367 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3333b4346f59b04017e32d61f457ba7ff457166731128f339d7987ae6b56b06d"} err="failed to get container status \"3333b4346f59b04017e32d61f457ba7ff457166731128f339d7987ae6b56b06d\": rpc error: code = NotFound desc = could not find container \"3333b4346f59b04017e32d61f457ba7ff457166731128f339d7987ae6b56b06d\": container with ID starting with 3333b4346f59b04017e32d61f457ba7ff457166731128f339d7987ae6b56b06d not found: ID does not exist" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.827403 4860 scope.go:117] "RemoveContainer" containerID="1426084ac580895c42354630910dd8c8e3a48ddd25f22abce1d28dc3de332a63" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.828026 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1426084ac580895c42354630910dd8c8e3a48ddd25f22abce1d28dc3de332a63"} err="failed to get container status \"1426084ac580895c42354630910dd8c8e3a48ddd25f22abce1d28dc3de332a63\": rpc error: code = NotFound desc = could not find container \"1426084ac580895c42354630910dd8c8e3a48ddd25f22abce1d28dc3de332a63\": container with ID starting with 1426084ac580895c42354630910dd8c8e3a48ddd25f22abce1d28dc3de332a63 not found: ID does not exist" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.828054 4860 scope.go:117] "RemoveContainer" containerID="c9984d5d242bb258405efe91fc9caf9e7ea592834445df2e78add9553a921c76" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.828299 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c9984d5d242bb258405efe91fc9caf9e7ea592834445df2e78add9553a921c76"} err="failed to get container status \"c9984d5d242bb258405efe91fc9caf9e7ea592834445df2e78add9553a921c76\": rpc error: code = NotFound desc = could not find container \"c9984d5d242bb258405efe91fc9caf9e7ea592834445df2e78add9553a921c76\": container with ID starting with c9984d5d242bb258405efe91fc9caf9e7ea592834445df2e78add9553a921c76 not found: ID does not exist" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.828318 4860 scope.go:117] "RemoveContainer" containerID="17cbcb49c06f3a9059cd5a0f05eedb8c785da30a3e502e31bae2f57b3486cb14" Dec 11 08:22:27 crc kubenswrapper[4860]: I1211 08:22:27.828512 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"17cbcb49c06f3a9059cd5a0f05eedb8c785da30a3e502e31bae2f57b3486cb14"} err="failed to get container status \"17cbcb49c06f3a9059cd5a0f05eedb8c785da30a3e502e31bae2f57b3486cb14\": rpc error: code = NotFound desc = could not find container \"17cbcb49c06f3a9059cd5a0f05eedb8c785da30a3e502e31bae2f57b3486cb14\": container with ID starting with 17cbcb49c06f3a9059cd5a0f05eedb8c785da30a3e502e31bae2f57b3486cb14 not found: ID does not exist" Dec 11 08:22:28 crc kubenswrapper[4860]: I1211 08:22:28.583763 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w2n7w" event={"ID":"430aa217-e9f9-445b-855a-69e8c797da38","Type":"ContainerStarted","Data":"5e07613a42e7899fa4aa16c4711460390623e1af858b7cf523955b7fc0fd42ad"} Dec 11 08:22:28 crc kubenswrapper[4860]: I1211 08:22:28.584075 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w2n7w" event={"ID":"430aa217-e9f9-445b-855a-69e8c797da38","Type":"ContainerStarted","Data":"6b6a556b96fef49dbb5f876c6a36acf6ddeb4195535ab1032e5601408427bd7a"} Dec 11 08:22:28 crc kubenswrapper[4860]: I1211 08:22:28.584086 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w2n7w" event={"ID":"430aa217-e9f9-445b-855a-69e8c797da38","Type":"ContainerStarted","Data":"4a41b561a5b3df9b7c29cd5e87e280c173aa3b23100b611f683ac2e0949ff71c"} Dec 11 08:22:28 crc kubenswrapper[4860]: I1211 08:22:28.584096 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w2n7w" event={"ID":"430aa217-e9f9-445b-855a-69e8c797da38","Type":"ContainerStarted","Data":"aefd774906907deb90d5cb77b0e63410ef81ec81dfef3ca93e4f6bbbf307f4c4"} Dec 11 08:22:28 crc kubenswrapper[4860]: I1211 08:22:28.584106 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w2n7w" event={"ID":"430aa217-e9f9-445b-855a-69e8c797da38","Type":"ContainerStarted","Data":"e306201ace9d76651a4c50d7da27ff287e516b3b2c563e8e62b8dfb31f545b59"} Dec 11 08:22:28 crc kubenswrapper[4860]: I1211 08:22:28.584116 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w2n7w" event={"ID":"430aa217-e9f9-445b-855a-69e8c797da38","Type":"ContainerStarted","Data":"fe2157d50c00dfdf826b4439b2db0636fa41e50c1c5b4983a7b5710119b11ec7"} Dec 11 08:22:29 crc kubenswrapper[4860]: I1211 08:22:29.586353 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0589a204-c98c-417d-8256-bfe3bf747660" path="/var/lib/kubelet/pods/0589a204-c98c-417d-8256-bfe3bf747660/volumes" Dec 11 08:22:30 crc kubenswrapper[4860]: I1211 08:22:30.596346 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w2n7w" event={"ID":"430aa217-e9f9-445b-855a-69e8c797da38","Type":"ContainerStarted","Data":"f4e3f67ce5253ee77cf89854d77b60c3eab029615fc9a5dc84700ff7ec2d8448"} Dec 11 08:22:33 crc kubenswrapper[4860]: I1211 08:22:33.618624 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w2n7w" event={"ID":"430aa217-e9f9-445b-855a-69e8c797da38","Type":"ContainerStarted","Data":"6955b8a1aca446e9521a2052740cad8d1a89fe3d00e61d96409f68eecd604616"} Dec 11 08:22:33 crc kubenswrapper[4860]: I1211 08:22:33.619218 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-w2n7w" Dec 11 08:22:33 crc kubenswrapper[4860]: I1211 08:22:33.619231 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-w2n7w" Dec 11 08:22:33 crc kubenswrapper[4860]: I1211 08:22:33.643520 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-w2n7w" Dec 11 08:22:33 crc kubenswrapper[4860]: I1211 08:22:33.659223 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-w2n7w" podStartSLOduration=7.659202706 podStartE2EDuration="7.659202706s" podCreationTimestamp="2025-12-11 08:22:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:22:33.650826004 +0000 UTC m=+686.379345059" watchObservedRunningTime="2025-12-11 08:22:33.659202706 +0000 UTC m=+686.387721761" Dec 11 08:22:34 crc kubenswrapper[4860]: I1211 08:22:34.624286 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-w2n7w" Dec 11 08:22:34 crc kubenswrapper[4860]: I1211 08:22:34.649267 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-w2n7w" Dec 11 08:22:38 crc kubenswrapper[4860]: I1211 08:22:38.578895 4860 scope.go:117] "RemoveContainer" containerID="c50698b594e2b9e096d97dd750433d216996139113b1349d3ba87a70d057511b" Dec 11 08:22:38 crc kubenswrapper[4860]: E1211 08:22:38.579756 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-pznwl_openshift-multus(5310c1fc-66c6-40aa-b1bf-5a59a2410f9c)\"" pod="openshift-multus/multus-pznwl" podUID="5310c1fc-66c6-40aa-b1bf-5a59a2410f9c" Dec 11 08:22:50 crc kubenswrapper[4860]: I1211 08:22:50.578906 4860 scope.go:117] "RemoveContainer" containerID="c50698b594e2b9e096d97dd750433d216996139113b1349d3ba87a70d057511b" Dec 11 08:22:50 crc kubenswrapper[4860]: I1211 08:22:50.737246 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-pznwl_5310c1fc-66c6-40aa-b1bf-5a59a2410f9c/kube-multus/2.log" Dec 11 08:22:50 crc kubenswrapper[4860]: I1211 08:22:50.737629 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-pznwl" event={"ID":"5310c1fc-66c6-40aa-b1bf-5a59a2410f9c","Type":"ContainerStarted","Data":"79edc20ef8f9e40c14bb665a301db5b69a5c26c85e6d3344521ecc6f1a10dd90"} Dec 11 08:22:57 crc kubenswrapper[4860]: I1211 08:22:57.069351 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-w2n7w" Dec 11 08:23:05 crc kubenswrapper[4860]: I1211 08:23:05.377588 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8nkkjk"] Dec 11 08:23:05 crc kubenswrapper[4860]: I1211 08:23:05.379374 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8nkkjk" Dec 11 08:23:05 crc kubenswrapper[4860]: I1211 08:23:05.384231 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 11 08:23:05 crc kubenswrapper[4860]: I1211 08:23:05.387532 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8nkkjk"] Dec 11 08:23:05 crc kubenswrapper[4860]: I1211 08:23:05.452924 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f80508d8-cf59-4656-af41-b0688b7b997c-bundle\") pod \"98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8nkkjk\" (UID: \"f80508d8-cf59-4656-af41-b0688b7b997c\") " pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8nkkjk" Dec 11 08:23:05 crc kubenswrapper[4860]: I1211 08:23:05.453204 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f80508d8-cf59-4656-af41-b0688b7b997c-util\") pod \"98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8nkkjk\" (UID: \"f80508d8-cf59-4656-af41-b0688b7b997c\") " pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8nkkjk" Dec 11 08:23:05 crc kubenswrapper[4860]: I1211 08:23:05.453301 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z575f\" (UniqueName: \"kubernetes.io/projected/f80508d8-cf59-4656-af41-b0688b7b997c-kube-api-access-z575f\") pod \"98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8nkkjk\" (UID: \"f80508d8-cf59-4656-af41-b0688b7b997c\") " pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8nkkjk" Dec 11 08:23:05 crc kubenswrapper[4860]: I1211 08:23:05.555374 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f80508d8-cf59-4656-af41-b0688b7b997c-bundle\") pod \"98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8nkkjk\" (UID: \"f80508d8-cf59-4656-af41-b0688b7b997c\") " pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8nkkjk" Dec 11 08:23:05 crc kubenswrapper[4860]: I1211 08:23:05.555914 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f80508d8-cf59-4656-af41-b0688b7b997c-util\") pod \"98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8nkkjk\" (UID: \"f80508d8-cf59-4656-af41-b0688b7b997c\") " pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8nkkjk" Dec 11 08:23:05 crc kubenswrapper[4860]: I1211 08:23:05.556498 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z575f\" (UniqueName: \"kubernetes.io/projected/f80508d8-cf59-4656-af41-b0688b7b997c-kube-api-access-z575f\") pod \"98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8nkkjk\" (UID: \"f80508d8-cf59-4656-af41-b0688b7b997c\") " pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8nkkjk" Dec 11 08:23:05 crc kubenswrapper[4860]: I1211 08:23:05.556544 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f80508d8-cf59-4656-af41-b0688b7b997c-util\") pod \"98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8nkkjk\" (UID: \"f80508d8-cf59-4656-af41-b0688b7b997c\") " pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8nkkjk" Dec 11 08:23:05 crc kubenswrapper[4860]: I1211 08:23:05.556567 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f80508d8-cf59-4656-af41-b0688b7b997c-bundle\") pod \"98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8nkkjk\" (UID: \"f80508d8-cf59-4656-af41-b0688b7b997c\") " pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8nkkjk" Dec 11 08:23:05 crc kubenswrapper[4860]: I1211 08:23:05.576853 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z575f\" (UniqueName: \"kubernetes.io/projected/f80508d8-cf59-4656-af41-b0688b7b997c-kube-api-access-z575f\") pod \"98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8nkkjk\" (UID: \"f80508d8-cf59-4656-af41-b0688b7b997c\") " pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8nkkjk" Dec 11 08:23:05 crc kubenswrapper[4860]: I1211 08:23:05.697755 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8nkkjk" Dec 11 08:23:06 crc kubenswrapper[4860]: I1211 08:23:06.083548 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8nkkjk"] Dec 11 08:23:06 crc kubenswrapper[4860]: I1211 08:23:06.832733 4860 generic.go:334] "Generic (PLEG): container finished" podID="f80508d8-cf59-4656-af41-b0688b7b997c" containerID="ed36de5b80d31900077759aecd74ed5200aa41fa481d5d823068db0d554a0082" exitCode=0 Dec 11 08:23:06 crc kubenswrapper[4860]: I1211 08:23:06.832848 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8nkkjk" event={"ID":"f80508d8-cf59-4656-af41-b0688b7b997c","Type":"ContainerDied","Data":"ed36de5b80d31900077759aecd74ed5200aa41fa481d5d823068db0d554a0082"} Dec 11 08:23:06 crc kubenswrapper[4860]: I1211 08:23:06.833138 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8nkkjk" event={"ID":"f80508d8-cf59-4656-af41-b0688b7b997c","Type":"ContainerStarted","Data":"a2c8d51840ad1d16019c4cb2e8e2acb9e732c8ba7ea3b9450f252261f906a99e"} Dec 11 08:23:08 crc kubenswrapper[4860]: I1211 08:23:08.844714 4860 generic.go:334] "Generic (PLEG): container finished" podID="f80508d8-cf59-4656-af41-b0688b7b997c" containerID="7decda45427567bab5f859400943997fef3c2a081b80ed2704fb239fb9aa93ef" exitCode=0 Dec 11 08:23:08 crc kubenswrapper[4860]: I1211 08:23:08.844769 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8nkkjk" event={"ID":"f80508d8-cf59-4656-af41-b0688b7b997c","Type":"ContainerDied","Data":"7decda45427567bab5f859400943997fef3c2a081b80ed2704fb239fb9aa93ef"} Dec 11 08:23:09 crc kubenswrapper[4860]: I1211 08:23:09.851284 4860 generic.go:334] "Generic (PLEG): container finished" podID="f80508d8-cf59-4656-af41-b0688b7b997c" containerID="af6f418968e3d3f50727bf956b1d703e5104d3199b5f75f857fea800d9cb3e75" exitCode=0 Dec 11 08:23:09 crc kubenswrapper[4860]: I1211 08:23:09.851370 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8nkkjk" event={"ID":"f80508d8-cf59-4656-af41-b0688b7b997c","Type":"ContainerDied","Data":"af6f418968e3d3f50727bf956b1d703e5104d3199b5f75f857fea800d9cb3e75"} Dec 11 08:23:11 crc kubenswrapper[4860]: I1211 08:23:11.082754 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8nkkjk" Dec 11 08:23:11 crc kubenswrapper[4860]: I1211 08:23:11.250039 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f80508d8-cf59-4656-af41-b0688b7b997c-bundle\") pod \"f80508d8-cf59-4656-af41-b0688b7b997c\" (UID: \"f80508d8-cf59-4656-af41-b0688b7b997c\") " Dec 11 08:23:11 crc kubenswrapper[4860]: I1211 08:23:11.250128 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z575f\" (UniqueName: \"kubernetes.io/projected/f80508d8-cf59-4656-af41-b0688b7b997c-kube-api-access-z575f\") pod \"f80508d8-cf59-4656-af41-b0688b7b997c\" (UID: \"f80508d8-cf59-4656-af41-b0688b7b997c\") " Dec 11 08:23:11 crc kubenswrapper[4860]: I1211 08:23:11.250205 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f80508d8-cf59-4656-af41-b0688b7b997c-util\") pod \"f80508d8-cf59-4656-af41-b0688b7b997c\" (UID: \"f80508d8-cf59-4656-af41-b0688b7b997c\") " Dec 11 08:23:11 crc kubenswrapper[4860]: I1211 08:23:11.251231 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f80508d8-cf59-4656-af41-b0688b7b997c-bundle" (OuterVolumeSpecName: "bundle") pod "f80508d8-cf59-4656-af41-b0688b7b997c" (UID: "f80508d8-cf59-4656-af41-b0688b7b997c"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:23:11 crc kubenswrapper[4860]: I1211 08:23:11.255832 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f80508d8-cf59-4656-af41-b0688b7b997c-kube-api-access-z575f" (OuterVolumeSpecName: "kube-api-access-z575f") pod "f80508d8-cf59-4656-af41-b0688b7b997c" (UID: "f80508d8-cf59-4656-af41-b0688b7b997c"). InnerVolumeSpecName "kube-api-access-z575f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:23:11 crc kubenswrapper[4860]: I1211 08:23:11.263746 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f80508d8-cf59-4656-af41-b0688b7b997c-util" (OuterVolumeSpecName: "util") pod "f80508d8-cf59-4656-af41-b0688b7b997c" (UID: "f80508d8-cf59-4656-af41-b0688b7b997c"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:23:11 crc kubenswrapper[4860]: I1211 08:23:11.351936 4860 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f80508d8-cf59-4656-af41-b0688b7b997c-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 08:23:11 crc kubenswrapper[4860]: I1211 08:23:11.351971 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z575f\" (UniqueName: \"kubernetes.io/projected/f80508d8-cf59-4656-af41-b0688b7b997c-kube-api-access-z575f\") on node \"crc\" DevicePath \"\"" Dec 11 08:23:11 crc kubenswrapper[4860]: I1211 08:23:11.351983 4860 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f80508d8-cf59-4656-af41-b0688b7b997c-util\") on node \"crc\" DevicePath \"\"" Dec 11 08:23:11 crc kubenswrapper[4860]: I1211 08:23:11.865166 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8nkkjk" event={"ID":"f80508d8-cf59-4656-af41-b0688b7b997c","Type":"ContainerDied","Data":"a2c8d51840ad1d16019c4cb2e8e2acb9e732c8ba7ea3b9450f252261f906a99e"} Dec 11 08:23:11 crc kubenswrapper[4860]: I1211 08:23:11.865462 4860 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a2c8d51840ad1d16019c4cb2e8e2acb9e732c8ba7ea3b9450f252261f906a99e" Dec 11 08:23:11 crc kubenswrapper[4860]: I1211 08:23:11.865230 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8nkkjk" Dec 11 08:23:16 crc kubenswrapper[4860]: I1211 08:23:16.753828 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-6769fb99d-49jqb"] Dec 11 08:23:16 crc kubenswrapper[4860]: E1211 08:23:16.757801 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f80508d8-cf59-4656-af41-b0688b7b997c" containerName="extract" Dec 11 08:23:16 crc kubenswrapper[4860]: I1211 08:23:16.757835 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="f80508d8-cf59-4656-af41-b0688b7b997c" containerName="extract" Dec 11 08:23:16 crc kubenswrapper[4860]: E1211 08:23:16.757855 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f80508d8-cf59-4656-af41-b0688b7b997c" containerName="util" Dec 11 08:23:16 crc kubenswrapper[4860]: I1211 08:23:16.757863 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="f80508d8-cf59-4656-af41-b0688b7b997c" containerName="util" Dec 11 08:23:16 crc kubenswrapper[4860]: E1211 08:23:16.757881 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f80508d8-cf59-4656-af41-b0688b7b997c" containerName="pull" Dec 11 08:23:16 crc kubenswrapper[4860]: I1211 08:23:16.757889 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="f80508d8-cf59-4656-af41-b0688b7b997c" containerName="pull" Dec 11 08:23:16 crc kubenswrapper[4860]: I1211 08:23:16.758109 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="f80508d8-cf59-4656-af41-b0688b7b997c" containerName="extract" Dec 11 08:23:16 crc kubenswrapper[4860]: I1211 08:23:16.758816 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-6769fb99d-49jqb" Dec 11 08:23:16 crc kubenswrapper[4860]: I1211 08:23:16.763407 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-hlzmx" Dec 11 08:23:16 crc kubenswrapper[4860]: I1211 08:23:16.763544 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Dec 11 08:23:16 crc kubenswrapper[4860]: I1211 08:23:16.763545 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Dec 11 08:23:16 crc kubenswrapper[4860]: I1211 08:23:16.764960 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-6769fb99d-49jqb"] Dec 11 08:23:16 crc kubenswrapper[4860]: I1211 08:23:16.843965 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4psw4\" (UniqueName: \"kubernetes.io/projected/4d7eadcc-525c-4503-a8c1-7c0b0c04a9e8-kube-api-access-4psw4\") pod \"nmstate-operator-6769fb99d-49jqb\" (UID: \"4d7eadcc-525c-4503-a8c1-7c0b0c04a9e8\") " pod="openshift-nmstate/nmstate-operator-6769fb99d-49jqb" Dec 11 08:23:16 crc kubenswrapper[4860]: I1211 08:23:16.945409 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4psw4\" (UniqueName: \"kubernetes.io/projected/4d7eadcc-525c-4503-a8c1-7c0b0c04a9e8-kube-api-access-4psw4\") pod \"nmstate-operator-6769fb99d-49jqb\" (UID: \"4d7eadcc-525c-4503-a8c1-7c0b0c04a9e8\") " pod="openshift-nmstate/nmstate-operator-6769fb99d-49jqb" Dec 11 08:23:16 crc kubenswrapper[4860]: I1211 08:23:16.964712 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4psw4\" (UniqueName: \"kubernetes.io/projected/4d7eadcc-525c-4503-a8c1-7c0b0c04a9e8-kube-api-access-4psw4\") pod \"nmstate-operator-6769fb99d-49jqb\" (UID: \"4d7eadcc-525c-4503-a8c1-7c0b0c04a9e8\") " pod="openshift-nmstate/nmstate-operator-6769fb99d-49jqb" Dec 11 08:23:17 crc kubenswrapper[4860]: I1211 08:23:17.111382 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-6769fb99d-49jqb" Dec 11 08:23:17 crc kubenswrapper[4860]: I1211 08:23:17.503631 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-6769fb99d-49jqb"] Dec 11 08:23:17 crc kubenswrapper[4860]: I1211 08:23:17.897665 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-6769fb99d-49jqb" event={"ID":"4d7eadcc-525c-4503-a8c1-7c0b0c04a9e8","Type":"ContainerStarted","Data":"162fdc7770f51533b4cb9d1e3ec099be870df1519ae98eed1d08051ecc97e91e"} Dec 11 08:23:19 crc kubenswrapper[4860]: I1211 08:23:19.913079 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-6769fb99d-49jqb" event={"ID":"4d7eadcc-525c-4503-a8c1-7c0b0c04a9e8","Type":"ContainerStarted","Data":"665fd800da53e4bb5c55467d792bec4b132412864cc0ed54bb36f8c95adf931e"} Dec 11 08:23:19 crc kubenswrapper[4860]: I1211 08:23:19.942979 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-6769fb99d-49jqb" podStartSLOduration=1.895870811 podStartE2EDuration="3.942950974s" podCreationTimestamp="2025-12-11 08:23:16 +0000 UTC" firstStartedPulling="2025-12-11 08:23:17.510976498 +0000 UTC m=+730.239495573" lastFinishedPulling="2025-12-11 08:23:19.558056681 +0000 UTC m=+732.286575736" observedRunningTime="2025-12-11 08:23:19.932107804 +0000 UTC m=+732.660626919" watchObservedRunningTime="2025-12-11 08:23:19.942950974 +0000 UTC m=+732.671470069" Dec 11 08:23:25 crc kubenswrapper[4860]: I1211 08:23:25.755365 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-7f7f7578db-qv9ws"] Dec 11 08:23:25 crc kubenswrapper[4860]: I1211 08:23:25.756699 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f7f7578db-qv9ws" Dec 11 08:23:25 crc kubenswrapper[4860]: I1211 08:23:25.758670 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-2lpjq" Dec 11 08:23:25 crc kubenswrapper[4860]: I1211 08:23:25.783559 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-f8fb84555-nnjsk"] Dec 11 08:23:25 crc kubenswrapper[4860]: I1211 08:23:25.784445 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-f8fb84555-nnjsk" Dec 11 08:23:25 crc kubenswrapper[4860]: I1211 08:23:25.788073 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Dec 11 08:23:25 crc kubenswrapper[4860]: I1211 08:23:25.796324 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-kv4zf"] Dec 11 08:23:25 crc kubenswrapper[4860]: I1211 08:23:25.799810 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-kv4zf" Dec 11 08:23:25 crc kubenswrapper[4860]: I1211 08:23:25.809756 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-f8fb84555-nnjsk"] Dec 11 08:23:25 crc kubenswrapper[4860]: I1211 08:23:25.866385 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f7f7578db-qv9ws"] Dec 11 08:23:25 crc kubenswrapper[4860]: I1211 08:23:25.872898 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9526l\" (UniqueName: \"kubernetes.io/projected/e18c8adc-cdba-4f7f-8b5d-f15a73397c2b-kube-api-access-9526l\") pod \"nmstate-metrics-7f7f7578db-qv9ws\" (UID: \"e18c8adc-cdba-4f7f-8b5d-f15a73397c2b\") " pod="openshift-nmstate/nmstate-metrics-7f7f7578db-qv9ws" Dec 11 08:23:25 crc kubenswrapper[4860]: I1211 08:23:25.957091 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6ff7998486-2q4xw"] Dec 11 08:23:25 crc kubenswrapper[4860]: I1211 08:23:25.958068 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-2q4xw" Dec 11 08:23:25 crc kubenswrapper[4860]: I1211 08:23:25.960252 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Dec 11 08:23:25 crc kubenswrapper[4860]: I1211 08:23:25.960268 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Dec 11 08:23:25 crc kubenswrapper[4860]: I1211 08:23:25.960614 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-bzgf2" Dec 11 08:23:25 crc kubenswrapper[4860]: I1211 08:23:25.977213 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zxhd2\" (UniqueName: \"kubernetes.io/projected/f17c9e82-c0f3-408d-a871-d24992f28860-kube-api-access-zxhd2\") pod \"nmstate-handler-kv4zf\" (UID: \"f17c9e82-c0f3-408d-a871-d24992f28860\") " pod="openshift-nmstate/nmstate-handler-kv4zf" Dec 11 08:23:25 crc kubenswrapper[4860]: I1211 08:23:25.977266 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n2mq9\" (UniqueName: \"kubernetes.io/projected/295f8d4b-cf0e-41e9-a9a6-8ff38c589ecc-kube-api-access-n2mq9\") pod \"nmstate-webhook-f8fb84555-nnjsk\" (UID: \"295f8d4b-cf0e-41e9-a9a6-8ff38c589ecc\") " pod="openshift-nmstate/nmstate-webhook-f8fb84555-nnjsk" Dec 11 08:23:25 crc kubenswrapper[4860]: I1211 08:23:25.977292 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/f17c9e82-c0f3-408d-a871-d24992f28860-dbus-socket\") pod \"nmstate-handler-kv4zf\" (UID: \"f17c9e82-c0f3-408d-a871-d24992f28860\") " pod="openshift-nmstate/nmstate-handler-kv4zf" Dec 11 08:23:25 crc kubenswrapper[4860]: I1211 08:23:25.977315 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5tk5z\" (UniqueName: \"kubernetes.io/projected/4cc66f42-2ab8-4a37-be52-5eb9be9fdd4b-kube-api-access-5tk5z\") pod \"nmstate-console-plugin-6ff7998486-2q4xw\" (UID: \"4cc66f42-2ab8-4a37-be52-5eb9be9fdd4b\") " pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-2q4xw" Dec 11 08:23:25 crc kubenswrapper[4860]: I1211 08:23:25.977336 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/f17c9e82-c0f3-408d-a871-d24992f28860-ovs-socket\") pod \"nmstate-handler-kv4zf\" (UID: \"f17c9e82-c0f3-408d-a871-d24992f28860\") " pod="openshift-nmstate/nmstate-handler-kv4zf" Dec 11 08:23:25 crc kubenswrapper[4860]: I1211 08:23:25.977391 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/f17c9e82-c0f3-408d-a871-d24992f28860-nmstate-lock\") pod \"nmstate-handler-kv4zf\" (UID: \"f17c9e82-c0f3-408d-a871-d24992f28860\") " pod="openshift-nmstate/nmstate-handler-kv4zf" Dec 11 08:23:25 crc kubenswrapper[4860]: I1211 08:23:25.977426 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9526l\" (UniqueName: \"kubernetes.io/projected/e18c8adc-cdba-4f7f-8b5d-f15a73397c2b-kube-api-access-9526l\") pod \"nmstate-metrics-7f7f7578db-qv9ws\" (UID: \"e18c8adc-cdba-4f7f-8b5d-f15a73397c2b\") " pod="openshift-nmstate/nmstate-metrics-7f7f7578db-qv9ws" Dec 11 08:23:25 crc kubenswrapper[4860]: I1211 08:23:25.977447 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/4cc66f42-2ab8-4a37-be52-5eb9be9fdd4b-plugin-serving-cert\") pod \"nmstate-console-plugin-6ff7998486-2q4xw\" (UID: \"4cc66f42-2ab8-4a37-be52-5eb9be9fdd4b\") " pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-2q4xw" Dec 11 08:23:25 crc kubenswrapper[4860]: I1211 08:23:25.977467 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/4cc66f42-2ab8-4a37-be52-5eb9be9fdd4b-nginx-conf\") pod \"nmstate-console-plugin-6ff7998486-2q4xw\" (UID: \"4cc66f42-2ab8-4a37-be52-5eb9be9fdd4b\") " pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-2q4xw" Dec 11 08:23:25 crc kubenswrapper[4860]: I1211 08:23:25.977539 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/295f8d4b-cf0e-41e9-a9a6-8ff38c589ecc-tls-key-pair\") pod \"nmstate-webhook-f8fb84555-nnjsk\" (UID: \"295f8d4b-cf0e-41e9-a9a6-8ff38c589ecc\") " pod="openshift-nmstate/nmstate-webhook-f8fb84555-nnjsk" Dec 11 08:23:26 crc kubenswrapper[4860]: I1211 08:23:26.007608 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9526l\" (UniqueName: \"kubernetes.io/projected/e18c8adc-cdba-4f7f-8b5d-f15a73397c2b-kube-api-access-9526l\") pod \"nmstate-metrics-7f7f7578db-qv9ws\" (UID: \"e18c8adc-cdba-4f7f-8b5d-f15a73397c2b\") " pod="openshift-nmstate/nmstate-metrics-7f7f7578db-qv9ws" Dec 11 08:23:26 crc kubenswrapper[4860]: I1211 08:23:26.015434 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6ff7998486-2q4xw"] Dec 11 08:23:26 crc kubenswrapper[4860]: I1211 08:23:26.075385 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f7f7578db-qv9ws" Dec 11 08:23:26 crc kubenswrapper[4860]: I1211 08:23:26.078905 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zxhd2\" (UniqueName: \"kubernetes.io/projected/f17c9e82-c0f3-408d-a871-d24992f28860-kube-api-access-zxhd2\") pod \"nmstate-handler-kv4zf\" (UID: \"f17c9e82-c0f3-408d-a871-d24992f28860\") " pod="openshift-nmstate/nmstate-handler-kv4zf" Dec 11 08:23:26 crc kubenswrapper[4860]: I1211 08:23:26.078944 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n2mq9\" (UniqueName: \"kubernetes.io/projected/295f8d4b-cf0e-41e9-a9a6-8ff38c589ecc-kube-api-access-n2mq9\") pod \"nmstate-webhook-f8fb84555-nnjsk\" (UID: \"295f8d4b-cf0e-41e9-a9a6-8ff38c589ecc\") " pod="openshift-nmstate/nmstate-webhook-f8fb84555-nnjsk" Dec 11 08:23:26 crc kubenswrapper[4860]: I1211 08:23:26.078968 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/f17c9e82-c0f3-408d-a871-d24992f28860-dbus-socket\") pod \"nmstate-handler-kv4zf\" (UID: \"f17c9e82-c0f3-408d-a871-d24992f28860\") " pod="openshift-nmstate/nmstate-handler-kv4zf" Dec 11 08:23:26 crc kubenswrapper[4860]: I1211 08:23:26.078987 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5tk5z\" (UniqueName: \"kubernetes.io/projected/4cc66f42-2ab8-4a37-be52-5eb9be9fdd4b-kube-api-access-5tk5z\") pod \"nmstate-console-plugin-6ff7998486-2q4xw\" (UID: \"4cc66f42-2ab8-4a37-be52-5eb9be9fdd4b\") " pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-2q4xw" Dec 11 08:23:26 crc kubenswrapper[4860]: I1211 08:23:26.079005 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/f17c9e82-c0f3-408d-a871-d24992f28860-ovs-socket\") pod \"nmstate-handler-kv4zf\" (UID: \"f17c9e82-c0f3-408d-a871-d24992f28860\") " pod="openshift-nmstate/nmstate-handler-kv4zf" Dec 11 08:23:26 crc kubenswrapper[4860]: I1211 08:23:26.079024 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/f17c9e82-c0f3-408d-a871-d24992f28860-nmstate-lock\") pod \"nmstate-handler-kv4zf\" (UID: \"f17c9e82-c0f3-408d-a871-d24992f28860\") " pod="openshift-nmstate/nmstate-handler-kv4zf" Dec 11 08:23:26 crc kubenswrapper[4860]: I1211 08:23:26.079055 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/4cc66f42-2ab8-4a37-be52-5eb9be9fdd4b-plugin-serving-cert\") pod \"nmstate-console-plugin-6ff7998486-2q4xw\" (UID: \"4cc66f42-2ab8-4a37-be52-5eb9be9fdd4b\") " pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-2q4xw" Dec 11 08:23:26 crc kubenswrapper[4860]: I1211 08:23:26.079074 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/4cc66f42-2ab8-4a37-be52-5eb9be9fdd4b-nginx-conf\") pod \"nmstate-console-plugin-6ff7998486-2q4xw\" (UID: \"4cc66f42-2ab8-4a37-be52-5eb9be9fdd4b\") " pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-2q4xw" Dec 11 08:23:26 crc kubenswrapper[4860]: I1211 08:23:26.079100 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/295f8d4b-cf0e-41e9-a9a6-8ff38c589ecc-tls-key-pair\") pod \"nmstate-webhook-f8fb84555-nnjsk\" (UID: \"295f8d4b-cf0e-41e9-a9a6-8ff38c589ecc\") " pod="openshift-nmstate/nmstate-webhook-f8fb84555-nnjsk" Dec 11 08:23:26 crc kubenswrapper[4860]: I1211 08:23:26.079978 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/f17c9e82-c0f3-408d-a871-d24992f28860-ovs-socket\") pod \"nmstate-handler-kv4zf\" (UID: \"f17c9e82-c0f3-408d-a871-d24992f28860\") " pod="openshift-nmstate/nmstate-handler-kv4zf" Dec 11 08:23:26 crc kubenswrapper[4860]: I1211 08:23:26.080021 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/f17c9e82-c0f3-408d-a871-d24992f28860-nmstate-lock\") pod \"nmstate-handler-kv4zf\" (UID: \"f17c9e82-c0f3-408d-a871-d24992f28860\") " pod="openshift-nmstate/nmstate-handler-kv4zf" Dec 11 08:23:26 crc kubenswrapper[4860]: I1211 08:23:26.081155 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/f17c9e82-c0f3-408d-a871-d24992f28860-dbus-socket\") pod \"nmstate-handler-kv4zf\" (UID: \"f17c9e82-c0f3-408d-a871-d24992f28860\") " pod="openshift-nmstate/nmstate-handler-kv4zf" Dec 11 08:23:26 crc kubenswrapper[4860]: I1211 08:23:26.082545 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/4cc66f42-2ab8-4a37-be52-5eb9be9fdd4b-nginx-conf\") pod \"nmstate-console-plugin-6ff7998486-2q4xw\" (UID: \"4cc66f42-2ab8-4a37-be52-5eb9be9fdd4b\") " pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-2q4xw" Dec 11 08:23:26 crc kubenswrapper[4860]: I1211 08:23:26.091391 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/295f8d4b-cf0e-41e9-a9a6-8ff38c589ecc-tls-key-pair\") pod \"nmstate-webhook-f8fb84555-nnjsk\" (UID: \"295f8d4b-cf0e-41e9-a9a6-8ff38c589ecc\") " pod="openshift-nmstate/nmstate-webhook-f8fb84555-nnjsk" Dec 11 08:23:26 crc kubenswrapper[4860]: I1211 08:23:26.091884 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/4cc66f42-2ab8-4a37-be52-5eb9be9fdd4b-plugin-serving-cert\") pod \"nmstate-console-plugin-6ff7998486-2q4xw\" (UID: \"4cc66f42-2ab8-4a37-be52-5eb9be9fdd4b\") " pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-2q4xw" Dec 11 08:23:26 crc kubenswrapper[4860]: I1211 08:23:26.108558 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zxhd2\" (UniqueName: \"kubernetes.io/projected/f17c9e82-c0f3-408d-a871-d24992f28860-kube-api-access-zxhd2\") pod \"nmstate-handler-kv4zf\" (UID: \"f17c9e82-c0f3-408d-a871-d24992f28860\") " pod="openshift-nmstate/nmstate-handler-kv4zf" Dec 11 08:23:26 crc kubenswrapper[4860]: I1211 08:23:26.110295 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5tk5z\" (UniqueName: \"kubernetes.io/projected/4cc66f42-2ab8-4a37-be52-5eb9be9fdd4b-kube-api-access-5tk5z\") pod \"nmstate-console-plugin-6ff7998486-2q4xw\" (UID: \"4cc66f42-2ab8-4a37-be52-5eb9be9fdd4b\") " pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-2q4xw" Dec 11 08:23:26 crc kubenswrapper[4860]: I1211 08:23:26.117708 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n2mq9\" (UniqueName: \"kubernetes.io/projected/295f8d4b-cf0e-41e9-a9a6-8ff38c589ecc-kube-api-access-n2mq9\") pod \"nmstate-webhook-f8fb84555-nnjsk\" (UID: \"295f8d4b-cf0e-41e9-a9a6-8ff38c589ecc\") " pod="openshift-nmstate/nmstate-webhook-f8fb84555-nnjsk" Dec 11 08:23:26 crc kubenswrapper[4860]: I1211 08:23:26.120346 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-kv4zf" Dec 11 08:23:26 crc kubenswrapper[4860]: I1211 08:23:26.178471 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-5b75dc8788-kd5v5"] Dec 11 08:23:26 crc kubenswrapper[4860]: I1211 08:23:26.180112 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5b75dc8788-kd5v5" Dec 11 08:23:26 crc kubenswrapper[4860]: I1211 08:23:26.196116 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-5b75dc8788-kd5v5"] Dec 11 08:23:26 crc kubenswrapper[4860]: I1211 08:23:26.273514 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-2q4xw" Dec 11 08:23:26 crc kubenswrapper[4860]: I1211 08:23:26.282654 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/5f00f919-3407-446e-8d6f-39bff7117a55-oauth-serving-cert\") pod \"console-5b75dc8788-kd5v5\" (UID: \"5f00f919-3407-446e-8d6f-39bff7117a55\") " pod="openshift-console/console-5b75dc8788-kd5v5" Dec 11 08:23:26 crc kubenswrapper[4860]: I1211 08:23:26.282719 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/5f00f919-3407-446e-8d6f-39bff7117a55-service-ca\") pod \"console-5b75dc8788-kd5v5\" (UID: \"5f00f919-3407-446e-8d6f-39bff7117a55\") " pod="openshift-console/console-5b75dc8788-kd5v5" Dec 11 08:23:26 crc kubenswrapper[4860]: I1211 08:23:26.282745 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/5f00f919-3407-446e-8d6f-39bff7117a55-console-serving-cert\") pod \"console-5b75dc8788-kd5v5\" (UID: \"5f00f919-3407-446e-8d6f-39bff7117a55\") " pod="openshift-console/console-5b75dc8788-kd5v5" Dec 11 08:23:26 crc kubenswrapper[4860]: I1211 08:23:26.282766 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/5f00f919-3407-446e-8d6f-39bff7117a55-console-oauth-config\") pod \"console-5b75dc8788-kd5v5\" (UID: \"5f00f919-3407-446e-8d6f-39bff7117a55\") " pod="openshift-console/console-5b75dc8788-kd5v5" Dec 11 08:23:26 crc kubenswrapper[4860]: I1211 08:23:26.282795 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/5f00f919-3407-446e-8d6f-39bff7117a55-console-config\") pod \"console-5b75dc8788-kd5v5\" (UID: \"5f00f919-3407-446e-8d6f-39bff7117a55\") " pod="openshift-console/console-5b75dc8788-kd5v5" Dec 11 08:23:26 crc kubenswrapper[4860]: I1211 08:23:26.282822 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5f00f919-3407-446e-8d6f-39bff7117a55-trusted-ca-bundle\") pod \"console-5b75dc8788-kd5v5\" (UID: \"5f00f919-3407-446e-8d6f-39bff7117a55\") " pod="openshift-console/console-5b75dc8788-kd5v5" Dec 11 08:23:26 crc kubenswrapper[4860]: I1211 08:23:26.282837 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6r7z4\" (UniqueName: \"kubernetes.io/projected/5f00f919-3407-446e-8d6f-39bff7117a55-kube-api-access-6r7z4\") pod \"console-5b75dc8788-kd5v5\" (UID: \"5f00f919-3407-446e-8d6f-39bff7117a55\") " pod="openshift-console/console-5b75dc8788-kd5v5" Dec 11 08:23:26 crc kubenswrapper[4860]: I1211 08:23:26.381830 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f7f7578db-qv9ws"] Dec 11 08:23:26 crc kubenswrapper[4860]: I1211 08:23:26.383940 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/5f00f919-3407-446e-8d6f-39bff7117a55-oauth-serving-cert\") pod \"console-5b75dc8788-kd5v5\" (UID: \"5f00f919-3407-446e-8d6f-39bff7117a55\") " pod="openshift-console/console-5b75dc8788-kd5v5" Dec 11 08:23:26 crc kubenswrapper[4860]: I1211 08:23:26.384023 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/5f00f919-3407-446e-8d6f-39bff7117a55-service-ca\") pod \"console-5b75dc8788-kd5v5\" (UID: \"5f00f919-3407-446e-8d6f-39bff7117a55\") " pod="openshift-console/console-5b75dc8788-kd5v5" Dec 11 08:23:26 crc kubenswrapper[4860]: I1211 08:23:26.384070 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/5f00f919-3407-446e-8d6f-39bff7117a55-console-serving-cert\") pod \"console-5b75dc8788-kd5v5\" (UID: \"5f00f919-3407-446e-8d6f-39bff7117a55\") " pod="openshift-console/console-5b75dc8788-kd5v5" Dec 11 08:23:26 crc kubenswrapper[4860]: I1211 08:23:26.384145 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/5f00f919-3407-446e-8d6f-39bff7117a55-console-oauth-config\") pod \"console-5b75dc8788-kd5v5\" (UID: \"5f00f919-3407-446e-8d6f-39bff7117a55\") " pod="openshift-console/console-5b75dc8788-kd5v5" Dec 11 08:23:26 crc kubenswrapper[4860]: I1211 08:23:26.384196 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/5f00f919-3407-446e-8d6f-39bff7117a55-console-config\") pod \"console-5b75dc8788-kd5v5\" (UID: \"5f00f919-3407-446e-8d6f-39bff7117a55\") " pod="openshift-console/console-5b75dc8788-kd5v5" Dec 11 08:23:26 crc kubenswrapper[4860]: I1211 08:23:26.384232 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5f00f919-3407-446e-8d6f-39bff7117a55-trusted-ca-bundle\") pod \"console-5b75dc8788-kd5v5\" (UID: \"5f00f919-3407-446e-8d6f-39bff7117a55\") " pod="openshift-console/console-5b75dc8788-kd5v5" Dec 11 08:23:26 crc kubenswrapper[4860]: I1211 08:23:26.384255 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6r7z4\" (UniqueName: \"kubernetes.io/projected/5f00f919-3407-446e-8d6f-39bff7117a55-kube-api-access-6r7z4\") pod \"console-5b75dc8788-kd5v5\" (UID: \"5f00f919-3407-446e-8d6f-39bff7117a55\") " pod="openshift-console/console-5b75dc8788-kd5v5" Dec 11 08:23:26 crc kubenswrapper[4860]: I1211 08:23:26.386382 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/5f00f919-3407-446e-8d6f-39bff7117a55-service-ca\") pod \"console-5b75dc8788-kd5v5\" (UID: \"5f00f919-3407-446e-8d6f-39bff7117a55\") " pod="openshift-console/console-5b75dc8788-kd5v5" Dec 11 08:23:26 crc kubenswrapper[4860]: I1211 08:23:26.386851 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/5f00f919-3407-446e-8d6f-39bff7117a55-oauth-serving-cert\") pod \"console-5b75dc8788-kd5v5\" (UID: \"5f00f919-3407-446e-8d6f-39bff7117a55\") " pod="openshift-console/console-5b75dc8788-kd5v5" Dec 11 08:23:26 crc kubenswrapper[4860]: I1211 08:23:26.386998 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/5f00f919-3407-446e-8d6f-39bff7117a55-console-config\") pod \"console-5b75dc8788-kd5v5\" (UID: \"5f00f919-3407-446e-8d6f-39bff7117a55\") " pod="openshift-console/console-5b75dc8788-kd5v5" Dec 11 08:23:26 crc kubenswrapper[4860]: I1211 08:23:26.388821 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5f00f919-3407-446e-8d6f-39bff7117a55-trusted-ca-bundle\") pod \"console-5b75dc8788-kd5v5\" (UID: \"5f00f919-3407-446e-8d6f-39bff7117a55\") " pod="openshift-console/console-5b75dc8788-kd5v5" Dec 11 08:23:26 crc kubenswrapper[4860]: I1211 08:23:26.396060 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/5f00f919-3407-446e-8d6f-39bff7117a55-console-serving-cert\") pod \"console-5b75dc8788-kd5v5\" (UID: \"5f00f919-3407-446e-8d6f-39bff7117a55\") " pod="openshift-console/console-5b75dc8788-kd5v5" Dec 11 08:23:26 crc kubenswrapper[4860]: I1211 08:23:26.396094 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/5f00f919-3407-446e-8d6f-39bff7117a55-console-oauth-config\") pod \"console-5b75dc8788-kd5v5\" (UID: \"5f00f919-3407-446e-8d6f-39bff7117a55\") " pod="openshift-console/console-5b75dc8788-kd5v5" Dec 11 08:23:26 crc kubenswrapper[4860]: I1211 08:23:26.404090 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-f8fb84555-nnjsk" Dec 11 08:23:26 crc kubenswrapper[4860]: I1211 08:23:26.405907 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6r7z4\" (UniqueName: \"kubernetes.io/projected/5f00f919-3407-446e-8d6f-39bff7117a55-kube-api-access-6r7z4\") pod \"console-5b75dc8788-kd5v5\" (UID: \"5f00f919-3407-446e-8d6f-39bff7117a55\") " pod="openshift-console/console-5b75dc8788-kd5v5" Dec 11 08:23:26 crc kubenswrapper[4860]: I1211 08:23:26.520429 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5b75dc8788-kd5v5" Dec 11 08:23:26 crc kubenswrapper[4860]: I1211 08:23:26.535410 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6ff7998486-2q4xw"] Dec 11 08:23:26 crc kubenswrapper[4860]: I1211 08:23:26.622415 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-f8fb84555-nnjsk"] Dec 11 08:23:26 crc kubenswrapper[4860]: I1211 08:23:26.756490 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-5b75dc8788-kd5v5"] Dec 11 08:23:26 crc kubenswrapper[4860]: W1211 08:23:26.765681 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5f00f919_3407_446e_8d6f_39bff7117a55.slice/crio-7e65ad86c8766f5a549177881d3cc9120870cbd037ec9e80c360e5376b9b983c WatchSource:0}: Error finding container 7e65ad86c8766f5a549177881d3cc9120870cbd037ec9e80c360e5376b9b983c: Status 404 returned error can't find the container with id 7e65ad86c8766f5a549177881d3cc9120870cbd037ec9e80c360e5376b9b983c Dec 11 08:23:26 crc kubenswrapper[4860]: I1211 08:23:26.951237 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f7f7578db-qv9ws" event={"ID":"e18c8adc-cdba-4f7f-8b5d-f15a73397c2b","Type":"ContainerStarted","Data":"66ea1037397bf961bb6c769c8b60ff297bbbdccb6243e08b4758e3f0a9e76a14"} Dec 11 08:23:26 crc kubenswrapper[4860]: I1211 08:23:26.952169 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-f8fb84555-nnjsk" event={"ID":"295f8d4b-cf0e-41e9-a9a6-8ff38c589ecc","Type":"ContainerStarted","Data":"7aa61ceac2e93d4824348cb838c9f9a609f03e246c3394420858ec359b60b1a1"} Dec 11 08:23:26 crc kubenswrapper[4860]: I1211 08:23:26.953309 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-2q4xw" event={"ID":"4cc66f42-2ab8-4a37-be52-5eb9be9fdd4b","Type":"ContainerStarted","Data":"5f2f9c1365ceeb614979145860c33b8f75dd4878cb1d8a4c69cb8b5f635daa1d"} Dec 11 08:23:26 crc kubenswrapper[4860]: I1211 08:23:26.955340 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5b75dc8788-kd5v5" event={"ID":"5f00f919-3407-446e-8d6f-39bff7117a55","Type":"ContainerStarted","Data":"426b0a4baf994b21a19eea35038cafcd661828a88b7f56fd8b80e49a031f274d"} Dec 11 08:23:26 crc kubenswrapper[4860]: I1211 08:23:26.955370 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5b75dc8788-kd5v5" event={"ID":"5f00f919-3407-446e-8d6f-39bff7117a55","Type":"ContainerStarted","Data":"7e65ad86c8766f5a549177881d3cc9120870cbd037ec9e80c360e5376b9b983c"} Dec 11 08:23:26 crc kubenswrapper[4860]: I1211 08:23:26.958275 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-kv4zf" event={"ID":"f17c9e82-c0f3-408d-a871-d24992f28860","Type":"ContainerStarted","Data":"768d5acd39ea9bf8ab1a798c0362d9e8202eceb8c1b56b4e186986c9136fe8c6"} Dec 11 08:23:26 crc kubenswrapper[4860]: I1211 08:23:26.982173 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-5b75dc8788-kd5v5" podStartSLOduration=0.982139723 podStartE2EDuration="982.139723ms" podCreationTimestamp="2025-12-11 08:23:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:23:26.975965569 +0000 UTC m=+739.704484624" watchObservedRunningTime="2025-12-11 08:23:26.982139723 +0000 UTC m=+739.710658768" Dec 11 08:23:29 crc kubenswrapper[4860]: I1211 08:23:29.994680 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-f8fb84555-nnjsk" event={"ID":"295f8d4b-cf0e-41e9-a9a6-8ff38c589ecc","Type":"ContainerStarted","Data":"3bd9dcf9df11fe635452d9b2d46df1a67b9557eca109b46fe4d8f5afe6fcfc85"} Dec 11 08:23:29 crc kubenswrapper[4860]: I1211 08:23:29.996510 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-f8fb84555-nnjsk" Dec 11 08:23:29 crc kubenswrapper[4860]: I1211 08:23:29.999710 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-2q4xw" event={"ID":"4cc66f42-2ab8-4a37-be52-5eb9be9fdd4b","Type":"ContainerStarted","Data":"84a3c2601b8845976f7d46a0f2f89cc26373b0c611d64b335dacf4ca311ee3bf"} Dec 11 08:23:30 crc kubenswrapper[4860]: I1211 08:23:30.003565 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-kv4zf" event={"ID":"f17c9e82-c0f3-408d-a871-d24992f28860","Type":"ContainerStarted","Data":"a9fb949de05a5476e26ae6017491165d5ec718e05b3d725fbe9f04c30c873f69"} Dec 11 08:23:30 crc kubenswrapper[4860]: I1211 08:23:30.003722 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-kv4zf" Dec 11 08:23:30 crc kubenswrapper[4860]: I1211 08:23:30.006576 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f7f7578db-qv9ws" event={"ID":"e18c8adc-cdba-4f7f-8b5d-f15a73397c2b","Type":"ContainerStarted","Data":"105d3d168c43315e9e791e2d5715524e755e7f154d0b18374f4101d6412e7a2b"} Dec 11 08:23:30 crc kubenswrapper[4860]: I1211 08:23:30.017527 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-f8fb84555-nnjsk" podStartSLOduration=2.438734526 podStartE2EDuration="5.016611759s" podCreationTimestamp="2025-12-11 08:23:25 +0000 UTC" firstStartedPulling="2025-12-11 08:23:26.638863145 +0000 UTC m=+739.367382200" lastFinishedPulling="2025-12-11 08:23:29.216740378 +0000 UTC m=+741.945259433" observedRunningTime="2025-12-11 08:23:30.013583474 +0000 UTC m=+742.742102539" watchObservedRunningTime="2025-12-11 08:23:30.016611759 +0000 UTC m=+742.745130814" Dec 11 08:23:30 crc kubenswrapper[4860]: I1211 08:23:30.035847 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-2q4xw" podStartSLOduration=2.389908713 podStartE2EDuration="5.035813691s" podCreationTimestamp="2025-12-11 08:23:25 +0000 UTC" firstStartedPulling="2025-12-11 08:23:26.548146889 +0000 UTC m=+739.276665944" lastFinishedPulling="2025-12-11 08:23:29.194051867 +0000 UTC m=+741.922570922" observedRunningTime="2025-12-11 08:23:30.031557338 +0000 UTC m=+742.760076393" watchObservedRunningTime="2025-12-11 08:23:30.035813691 +0000 UTC m=+742.764332746" Dec 11 08:23:30 crc kubenswrapper[4860]: I1211 08:23:30.071749 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-kv4zf" podStartSLOduration=2.037330764 podStartE2EDuration="5.071723007s" podCreationTimestamp="2025-12-11 08:23:25 +0000 UTC" firstStartedPulling="2025-12-11 08:23:26.182125138 +0000 UTC m=+738.910644193" lastFinishedPulling="2025-12-11 08:23:29.216517381 +0000 UTC m=+741.945036436" observedRunningTime="2025-12-11 08:23:30.063143859 +0000 UTC m=+742.791662914" watchObservedRunningTime="2025-12-11 08:23:30.071723007 +0000 UTC m=+742.800242052" Dec 11 08:23:32 crc kubenswrapper[4860]: I1211 08:23:32.022574 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f7f7578db-qv9ws" event={"ID":"e18c8adc-cdba-4f7f-8b5d-f15a73397c2b","Type":"ContainerStarted","Data":"c701257938ad254ce8b65443f18b45324b29dd443927acf55dd96604cc435f08"} Dec 11 08:23:32 crc kubenswrapper[4860]: I1211 08:23:32.047165 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-7f7f7578db-qv9ws" podStartSLOduration=1.998985342 podStartE2EDuration="7.047140693s" podCreationTimestamp="2025-12-11 08:23:25 +0000 UTC" firstStartedPulling="2025-12-11 08:23:26.412264647 +0000 UTC m=+739.140783702" lastFinishedPulling="2025-12-11 08:23:31.460420008 +0000 UTC m=+744.188939053" observedRunningTime="2025-12-11 08:23:32.044969015 +0000 UTC m=+744.773488090" watchObservedRunningTime="2025-12-11 08:23:32.047140693 +0000 UTC m=+744.775659748" Dec 11 08:23:36 crc kubenswrapper[4860]: I1211 08:23:36.148450 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-kv4zf" Dec 11 08:23:36 crc kubenswrapper[4860]: I1211 08:23:36.520694 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-5b75dc8788-kd5v5" Dec 11 08:23:36 crc kubenswrapper[4860]: I1211 08:23:36.520769 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-5b75dc8788-kd5v5" Dec 11 08:23:36 crc kubenswrapper[4860]: I1211 08:23:36.527122 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-5b75dc8788-kd5v5" Dec 11 08:23:37 crc kubenswrapper[4860]: I1211 08:23:37.064014 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-5b75dc8788-kd5v5" Dec 11 08:23:37 crc kubenswrapper[4860]: I1211 08:23:37.125938 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-rhl5v"] Dec 11 08:23:46 crc kubenswrapper[4860]: I1211 08:23:46.412622 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-f8fb84555-nnjsk" Dec 11 08:23:50 crc kubenswrapper[4860]: I1211 08:23:50.745442 4860 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 11 08:24:00 crc kubenswrapper[4860]: I1211 08:24:00.113871 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4vr6tn"] Dec 11 08:24:00 crc kubenswrapper[4860]: I1211 08:24:00.116327 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4vr6tn" Dec 11 08:24:00 crc kubenswrapper[4860]: I1211 08:24:00.118862 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 11 08:24:00 crc kubenswrapper[4860]: I1211 08:24:00.124567 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4vr6tn"] Dec 11 08:24:00 crc kubenswrapper[4860]: I1211 08:24:00.213515 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/da496881-c195-4cc2-9277-c8118bf651dc-util\") pod \"5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4vr6tn\" (UID: \"da496881-c195-4cc2-9277-c8118bf651dc\") " pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4vr6tn" Dec 11 08:24:00 crc kubenswrapper[4860]: I1211 08:24:00.213867 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qcfsr\" (UniqueName: \"kubernetes.io/projected/da496881-c195-4cc2-9277-c8118bf651dc-kube-api-access-qcfsr\") pod \"5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4vr6tn\" (UID: \"da496881-c195-4cc2-9277-c8118bf651dc\") " pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4vr6tn" Dec 11 08:24:00 crc kubenswrapper[4860]: I1211 08:24:00.213932 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/da496881-c195-4cc2-9277-c8118bf651dc-bundle\") pod \"5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4vr6tn\" (UID: \"da496881-c195-4cc2-9277-c8118bf651dc\") " pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4vr6tn" Dec 11 08:24:00 crc kubenswrapper[4860]: I1211 08:24:00.315381 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/da496881-c195-4cc2-9277-c8118bf651dc-bundle\") pod \"5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4vr6tn\" (UID: \"da496881-c195-4cc2-9277-c8118bf651dc\") " pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4vr6tn" Dec 11 08:24:00 crc kubenswrapper[4860]: I1211 08:24:00.316002 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/da496881-c195-4cc2-9277-c8118bf651dc-util\") pod \"5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4vr6tn\" (UID: \"da496881-c195-4cc2-9277-c8118bf651dc\") " pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4vr6tn" Dec 11 08:24:00 crc kubenswrapper[4860]: I1211 08:24:00.316085 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/da496881-c195-4cc2-9277-c8118bf651dc-bundle\") pod \"5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4vr6tn\" (UID: \"da496881-c195-4cc2-9277-c8118bf651dc\") " pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4vr6tn" Dec 11 08:24:00 crc kubenswrapper[4860]: I1211 08:24:00.316335 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qcfsr\" (UniqueName: \"kubernetes.io/projected/da496881-c195-4cc2-9277-c8118bf651dc-kube-api-access-qcfsr\") pod \"5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4vr6tn\" (UID: \"da496881-c195-4cc2-9277-c8118bf651dc\") " pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4vr6tn" Dec 11 08:24:00 crc kubenswrapper[4860]: I1211 08:24:00.316777 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/da496881-c195-4cc2-9277-c8118bf651dc-util\") pod \"5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4vr6tn\" (UID: \"da496881-c195-4cc2-9277-c8118bf651dc\") " pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4vr6tn" Dec 11 08:24:00 crc kubenswrapper[4860]: I1211 08:24:00.339502 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qcfsr\" (UniqueName: \"kubernetes.io/projected/da496881-c195-4cc2-9277-c8118bf651dc-kube-api-access-qcfsr\") pod \"5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4vr6tn\" (UID: \"da496881-c195-4cc2-9277-c8118bf651dc\") " pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4vr6tn" Dec 11 08:24:00 crc kubenswrapper[4860]: I1211 08:24:00.436300 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4vr6tn" Dec 11 08:24:00 crc kubenswrapper[4860]: I1211 08:24:00.829381 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4vr6tn"] Dec 11 08:24:01 crc kubenswrapper[4860]: I1211 08:24:01.208351 4860 generic.go:334] "Generic (PLEG): container finished" podID="da496881-c195-4cc2-9277-c8118bf651dc" containerID="9113e32b49d99e41be463b93c74237e56178d9971c786f53baf291a9ab7229cd" exitCode=0 Dec 11 08:24:01 crc kubenswrapper[4860]: I1211 08:24:01.208508 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4vr6tn" event={"ID":"da496881-c195-4cc2-9277-c8118bf651dc","Type":"ContainerDied","Data":"9113e32b49d99e41be463b93c74237e56178d9971c786f53baf291a9ab7229cd"} Dec 11 08:24:01 crc kubenswrapper[4860]: I1211 08:24:01.208852 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4vr6tn" event={"ID":"da496881-c195-4cc2-9277-c8118bf651dc","Type":"ContainerStarted","Data":"c7497b24fa1ed3880c1ec828ba86d83ae2a50e7d846d4c2dba165c0cedc5c518"} Dec 11 08:24:02 crc kubenswrapper[4860]: I1211 08:24:02.178323 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-rhl5v" podUID="d7d1536c-ed04-4199-b839-b94e9482a63c" containerName="console" containerID="cri-o://212f84085c22f84a02e17d9434646d1d104baaecfabef26acde4f527d50cfcdf" gracePeriod=15 Dec 11 08:24:02 crc kubenswrapper[4860]: I1211 08:24:02.614552 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-rhl5v_d7d1536c-ed04-4199-b839-b94e9482a63c/console/0.log" Dec 11 08:24:02 crc kubenswrapper[4860]: I1211 08:24:02.614788 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-rhl5v" Dec 11 08:24:02 crc kubenswrapper[4860]: I1211 08:24:02.754185 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d7d1536c-ed04-4199-b839-b94e9482a63c-service-ca\") pod \"d7d1536c-ed04-4199-b839-b94e9482a63c\" (UID: \"d7d1536c-ed04-4199-b839-b94e9482a63c\") " Dec 11 08:24:02 crc kubenswrapper[4860]: I1211 08:24:02.754308 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/d7d1536c-ed04-4199-b839-b94e9482a63c-oauth-serving-cert\") pod \"d7d1536c-ed04-4199-b839-b94e9482a63c\" (UID: \"d7d1536c-ed04-4199-b839-b94e9482a63c\") " Dec 11 08:24:02 crc kubenswrapper[4860]: I1211 08:24:02.754346 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c465q\" (UniqueName: \"kubernetes.io/projected/d7d1536c-ed04-4199-b839-b94e9482a63c-kube-api-access-c465q\") pod \"d7d1536c-ed04-4199-b839-b94e9482a63c\" (UID: \"d7d1536c-ed04-4199-b839-b94e9482a63c\") " Dec 11 08:24:02 crc kubenswrapper[4860]: I1211 08:24:02.754389 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d7d1536c-ed04-4199-b839-b94e9482a63c-trusted-ca-bundle\") pod \"d7d1536c-ed04-4199-b839-b94e9482a63c\" (UID: \"d7d1536c-ed04-4199-b839-b94e9482a63c\") " Dec 11 08:24:02 crc kubenswrapper[4860]: I1211 08:24:02.754451 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/d7d1536c-ed04-4199-b839-b94e9482a63c-console-serving-cert\") pod \"d7d1536c-ed04-4199-b839-b94e9482a63c\" (UID: \"d7d1536c-ed04-4199-b839-b94e9482a63c\") " Dec 11 08:24:02 crc kubenswrapper[4860]: I1211 08:24:02.754512 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/d7d1536c-ed04-4199-b839-b94e9482a63c-console-config\") pod \"d7d1536c-ed04-4199-b839-b94e9482a63c\" (UID: \"d7d1536c-ed04-4199-b839-b94e9482a63c\") " Dec 11 08:24:02 crc kubenswrapper[4860]: I1211 08:24:02.754558 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/d7d1536c-ed04-4199-b839-b94e9482a63c-console-oauth-config\") pod \"d7d1536c-ed04-4199-b839-b94e9482a63c\" (UID: \"d7d1536c-ed04-4199-b839-b94e9482a63c\") " Dec 11 08:24:02 crc kubenswrapper[4860]: I1211 08:24:02.755289 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7d1536c-ed04-4199-b839-b94e9482a63c-service-ca" (OuterVolumeSpecName: "service-ca") pod "d7d1536c-ed04-4199-b839-b94e9482a63c" (UID: "d7d1536c-ed04-4199-b839-b94e9482a63c"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:24:02 crc kubenswrapper[4860]: I1211 08:24:02.755276 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7d1536c-ed04-4199-b839-b94e9482a63c-console-config" (OuterVolumeSpecName: "console-config") pod "d7d1536c-ed04-4199-b839-b94e9482a63c" (UID: "d7d1536c-ed04-4199-b839-b94e9482a63c"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:24:02 crc kubenswrapper[4860]: I1211 08:24:02.755435 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7d1536c-ed04-4199-b839-b94e9482a63c-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "d7d1536c-ed04-4199-b839-b94e9482a63c" (UID: "d7d1536c-ed04-4199-b839-b94e9482a63c"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:24:02 crc kubenswrapper[4860]: I1211 08:24:02.755818 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7d1536c-ed04-4199-b839-b94e9482a63c-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "d7d1536c-ed04-4199-b839-b94e9482a63c" (UID: "d7d1536c-ed04-4199-b839-b94e9482a63c"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:24:02 crc kubenswrapper[4860]: I1211 08:24:02.763453 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7d1536c-ed04-4199-b839-b94e9482a63c-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "d7d1536c-ed04-4199-b839-b94e9482a63c" (UID: "d7d1536c-ed04-4199-b839-b94e9482a63c"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:24:02 crc kubenswrapper[4860]: I1211 08:24:02.764327 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7d1536c-ed04-4199-b839-b94e9482a63c-kube-api-access-c465q" (OuterVolumeSpecName: "kube-api-access-c465q") pod "d7d1536c-ed04-4199-b839-b94e9482a63c" (UID: "d7d1536c-ed04-4199-b839-b94e9482a63c"). InnerVolumeSpecName "kube-api-access-c465q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:24:02 crc kubenswrapper[4860]: I1211 08:24:02.764425 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7d1536c-ed04-4199-b839-b94e9482a63c-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "d7d1536c-ed04-4199-b839-b94e9482a63c" (UID: "d7d1536c-ed04-4199-b839-b94e9482a63c"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:24:02 crc kubenswrapper[4860]: I1211 08:24:02.856118 4860 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/d7d1536c-ed04-4199-b839-b94e9482a63c-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 08:24:02 crc kubenswrapper[4860]: I1211 08:24:02.856175 4860 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/d7d1536c-ed04-4199-b839-b94e9482a63c-console-config\") on node \"crc\" DevicePath \"\"" Dec 11 08:24:02 crc kubenswrapper[4860]: I1211 08:24:02.856192 4860 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/d7d1536c-ed04-4199-b839-b94e9482a63c-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 11 08:24:02 crc kubenswrapper[4860]: I1211 08:24:02.856205 4860 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d7d1536c-ed04-4199-b839-b94e9482a63c-service-ca\") on node \"crc\" DevicePath \"\"" Dec 11 08:24:02 crc kubenswrapper[4860]: I1211 08:24:02.856221 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c465q\" (UniqueName: \"kubernetes.io/projected/d7d1536c-ed04-4199-b839-b94e9482a63c-kube-api-access-c465q\") on node \"crc\" DevicePath \"\"" Dec 11 08:24:02 crc kubenswrapper[4860]: I1211 08:24:02.856236 4860 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/d7d1536c-ed04-4199-b839-b94e9482a63c-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 08:24:02 crc kubenswrapper[4860]: I1211 08:24:02.856249 4860 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d7d1536c-ed04-4199-b839-b94e9482a63c-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 08:24:03 crc kubenswrapper[4860]: I1211 08:24:03.225947 4860 generic.go:334] "Generic (PLEG): container finished" podID="da496881-c195-4cc2-9277-c8118bf651dc" containerID="93e7670fdd56bcffadf5e5223af1c5ae3037845b3b73f1b16b42edc73656c05c" exitCode=0 Dec 11 08:24:03 crc kubenswrapper[4860]: I1211 08:24:03.226069 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4vr6tn" event={"ID":"da496881-c195-4cc2-9277-c8118bf651dc","Type":"ContainerDied","Data":"93e7670fdd56bcffadf5e5223af1c5ae3037845b3b73f1b16b42edc73656c05c"} Dec 11 08:24:03 crc kubenswrapper[4860]: I1211 08:24:03.228368 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-rhl5v_d7d1536c-ed04-4199-b839-b94e9482a63c/console/0.log" Dec 11 08:24:03 crc kubenswrapper[4860]: I1211 08:24:03.228418 4860 generic.go:334] "Generic (PLEG): container finished" podID="d7d1536c-ed04-4199-b839-b94e9482a63c" containerID="212f84085c22f84a02e17d9434646d1d104baaecfabef26acde4f527d50cfcdf" exitCode=2 Dec 11 08:24:03 crc kubenswrapper[4860]: I1211 08:24:03.228458 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-rhl5v" event={"ID":"d7d1536c-ed04-4199-b839-b94e9482a63c","Type":"ContainerDied","Data":"212f84085c22f84a02e17d9434646d1d104baaecfabef26acde4f527d50cfcdf"} Dec 11 08:24:03 crc kubenswrapper[4860]: I1211 08:24:03.228511 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-rhl5v" event={"ID":"d7d1536c-ed04-4199-b839-b94e9482a63c","Type":"ContainerDied","Data":"f397ce5904d9e55432c6e8b65b71e54575de3faadf6cc259328986bc042780ce"} Dec 11 08:24:03 crc kubenswrapper[4860]: I1211 08:24:03.228516 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-rhl5v" Dec 11 08:24:03 crc kubenswrapper[4860]: I1211 08:24:03.228538 4860 scope.go:117] "RemoveContainer" containerID="212f84085c22f84a02e17d9434646d1d104baaecfabef26acde4f527d50cfcdf" Dec 11 08:24:03 crc kubenswrapper[4860]: I1211 08:24:03.257286 4860 scope.go:117] "RemoveContainer" containerID="212f84085c22f84a02e17d9434646d1d104baaecfabef26acde4f527d50cfcdf" Dec 11 08:24:03 crc kubenswrapper[4860]: E1211 08:24:03.258414 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"212f84085c22f84a02e17d9434646d1d104baaecfabef26acde4f527d50cfcdf\": container with ID starting with 212f84085c22f84a02e17d9434646d1d104baaecfabef26acde4f527d50cfcdf not found: ID does not exist" containerID="212f84085c22f84a02e17d9434646d1d104baaecfabef26acde4f527d50cfcdf" Dec 11 08:24:03 crc kubenswrapper[4860]: I1211 08:24:03.258456 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"212f84085c22f84a02e17d9434646d1d104baaecfabef26acde4f527d50cfcdf"} err="failed to get container status \"212f84085c22f84a02e17d9434646d1d104baaecfabef26acde4f527d50cfcdf\": rpc error: code = NotFound desc = could not find container \"212f84085c22f84a02e17d9434646d1d104baaecfabef26acde4f527d50cfcdf\": container with ID starting with 212f84085c22f84a02e17d9434646d1d104baaecfabef26acde4f527d50cfcdf not found: ID does not exist" Dec 11 08:24:03 crc kubenswrapper[4860]: I1211 08:24:03.267356 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-rhl5v"] Dec 11 08:24:03 crc kubenswrapper[4860]: I1211 08:24:03.279378 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-rhl5v"] Dec 11 08:24:03 crc kubenswrapper[4860]: I1211 08:24:03.471844 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-hbw2q"] Dec 11 08:24:03 crc kubenswrapper[4860]: E1211 08:24:03.472222 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7d1536c-ed04-4199-b839-b94e9482a63c" containerName="console" Dec 11 08:24:03 crc kubenswrapper[4860]: I1211 08:24:03.472247 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7d1536c-ed04-4199-b839-b94e9482a63c" containerName="console" Dec 11 08:24:03 crc kubenswrapper[4860]: I1211 08:24:03.472431 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7d1536c-ed04-4199-b839-b94e9482a63c" containerName="console" Dec 11 08:24:03 crc kubenswrapper[4860]: I1211 08:24:03.473434 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hbw2q" Dec 11 08:24:03 crc kubenswrapper[4860]: I1211 08:24:03.492890 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hbw2q"] Dec 11 08:24:03 crc kubenswrapper[4860]: I1211 08:24:03.566363 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dfaade06-9008-441e-ad04-990d22253a33-utilities\") pod \"redhat-operators-hbw2q\" (UID: \"dfaade06-9008-441e-ad04-990d22253a33\") " pod="openshift-marketplace/redhat-operators-hbw2q" Dec 11 08:24:03 crc kubenswrapper[4860]: I1211 08:24:03.566485 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dfaade06-9008-441e-ad04-990d22253a33-catalog-content\") pod \"redhat-operators-hbw2q\" (UID: \"dfaade06-9008-441e-ad04-990d22253a33\") " pod="openshift-marketplace/redhat-operators-hbw2q" Dec 11 08:24:03 crc kubenswrapper[4860]: I1211 08:24:03.566533 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dmhg2\" (UniqueName: \"kubernetes.io/projected/dfaade06-9008-441e-ad04-990d22253a33-kube-api-access-dmhg2\") pod \"redhat-operators-hbw2q\" (UID: \"dfaade06-9008-441e-ad04-990d22253a33\") " pod="openshift-marketplace/redhat-operators-hbw2q" Dec 11 08:24:03 crc kubenswrapper[4860]: I1211 08:24:03.587690 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d7d1536c-ed04-4199-b839-b94e9482a63c" path="/var/lib/kubelet/pods/d7d1536c-ed04-4199-b839-b94e9482a63c/volumes" Dec 11 08:24:03 crc kubenswrapper[4860]: I1211 08:24:03.668189 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dfaade06-9008-441e-ad04-990d22253a33-utilities\") pod \"redhat-operators-hbw2q\" (UID: \"dfaade06-9008-441e-ad04-990d22253a33\") " pod="openshift-marketplace/redhat-operators-hbw2q" Dec 11 08:24:03 crc kubenswrapper[4860]: I1211 08:24:03.668626 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dfaade06-9008-441e-ad04-990d22253a33-catalog-content\") pod \"redhat-operators-hbw2q\" (UID: \"dfaade06-9008-441e-ad04-990d22253a33\") " pod="openshift-marketplace/redhat-operators-hbw2q" Dec 11 08:24:03 crc kubenswrapper[4860]: I1211 08:24:03.668755 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dmhg2\" (UniqueName: \"kubernetes.io/projected/dfaade06-9008-441e-ad04-990d22253a33-kube-api-access-dmhg2\") pod \"redhat-operators-hbw2q\" (UID: \"dfaade06-9008-441e-ad04-990d22253a33\") " pod="openshift-marketplace/redhat-operators-hbw2q" Dec 11 08:24:03 crc kubenswrapper[4860]: I1211 08:24:03.669193 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dfaade06-9008-441e-ad04-990d22253a33-utilities\") pod \"redhat-operators-hbw2q\" (UID: \"dfaade06-9008-441e-ad04-990d22253a33\") " pod="openshift-marketplace/redhat-operators-hbw2q" Dec 11 08:24:03 crc kubenswrapper[4860]: I1211 08:24:03.669321 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dfaade06-9008-441e-ad04-990d22253a33-catalog-content\") pod \"redhat-operators-hbw2q\" (UID: \"dfaade06-9008-441e-ad04-990d22253a33\") " pod="openshift-marketplace/redhat-operators-hbw2q" Dec 11 08:24:03 crc kubenswrapper[4860]: I1211 08:24:03.699062 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dmhg2\" (UniqueName: \"kubernetes.io/projected/dfaade06-9008-441e-ad04-990d22253a33-kube-api-access-dmhg2\") pod \"redhat-operators-hbw2q\" (UID: \"dfaade06-9008-441e-ad04-990d22253a33\") " pod="openshift-marketplace/redhat-operators-hbw2q" Dec 11 08:24:03 crc kubenswrapper[4860]: I1211 08:24:03.792096 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hbw2q" Dec 11 08:24:04 crc kubenswrapper[4860]: I1211 08:24:04.006591 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hbw2q"] Dec 11 08:24:04 crc kubenswrapper[4860]: W1211 08:24:04.016868 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddfaade06_9008_441e_ad04_990d22253a33.slice/crio-6c58ce405276391d579a713f700fd069e359c0483cc35f57e2864a669b159826 WatchSource:0}: Error finding container 6c58ce405276391d579a713f700fd069e359c0483cc35f57e2864a669b159826: Status 404 returned error can't find the container with id 6c58ce405276391d579a713f700fd069e359c0483cc35f57e2864a669b159826 Dec 11 08:24:04 crc kubenswrapper[4860]: I1211 08:24:04.239997 4860 generic.go:334] "Generic (PLEG): container finished" podID="da496881-c195-4cc2-9277-c8118bf651dc" containerID="65dcff99f2f0fd5048ea9cbd1c859541be39c5b05128ffd51d7dadfceb67e7b8" exitCode=0 Dec 11 08:24:04 crc kubenswrapper[4860]: I1211 08:24:04.240432 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4vr6tn" event={"ID":"da496881-c195-4cc2-9277-c8118bf651dc","Type":"ContainerDied","Data":"65dcff99f2f0fd5048ea9cbd1c859541be39c5b05128ffd51d7dadfceb67e7b8"} Dec 11 08:24:04 crc kubenswrapper[4860]: I1211 08:24:04.244280 4860 generic.go:334] "Generic (PLEG): container finished" podID="dfaade06-9008-441e-ad04-990d22253a33" containerID="f2f937bf6a7eafbf269831b9791c4f63b0e752e769d5b2f63d1e262a14f4e2f3" exitCode=0 Dec 11 08:24:04 crc kubenswrapper[4860]: I1211 08:24:04.244349 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hbw2q" event={"ID":"dfaade06-9008-441e-ad04-990d22253a33","Type":"ContainerDied","Data":"f2f937bf6a7eafbf269831b9791c4f63b0e752e769d5b2f63d1e262a14f4e2f3"} Dec 11 08:24:04 crc kubenswrapper[4860]: I1211 08:24:04.244415 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hbw2q" event={"ID":"dfaade06-9008-441e-ad04-990d22253a33","Type":"ContainerStarted","Data":"6c58ce405276391d579a713f700fd069e359c0483cc35f57e2864a669b159826"} Dec 11 08:24:05 crc kubenswrapper[4860]: I1211 08:24:05.536779 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4vr6tn" Dec 11 08:24:05 crc kubenswrapper[4860]: I1211 08:24:05.693281 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/da496881-c195-4cc2-9277-c8118bf651dc-bundle\") pod \"da496881-c195-4cc2-9277-c8118bf651dc\" (UID: \"da496881-c195-4cc2-9277-c8118bf651dc\") " Dec 11 08:24:05 crc kubenswrapper[4860]: I1211 08:24:05.693442 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/da496881-c195-4cc2-9277-c8118bf651dc-util\") pod \"da496881-c195-4cc2-9277-c8118bf651dc\" (UID: \"da496881-c195-4cc2-9277-c8118bf651dc\") " Dec 11 08:24:05 crc kubenswrapper[4860]: I1211 08:24:05.693507 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qcfsr\" (UniqueName: \"kubernetes.io/projected/da496881-c195-4cc2-9277-c8118bf651dc-kube-api-access-qcfsr\") pod \"da496881-c195-4cc2-9277-c8118bf651dc\" (UID: \"da496881-c195-4cc2-9277-c8118bf651dc\") " Dec 11 08:24:05 crc kubenswrapper[4860]: I1211 08:24:05.696224 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/da496881-c195-4cc2-9277-c8118bf651dc-bundle" (OuterVolumeSpecName: "bundle") pod "da496881-c195-4cc2-9277-c8118bf651dc" (UID: "da496881-c195-4cc2-9277-c8118bf651dc"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:24:05 crc kubenswrapper[4860]: I1211 08:24:05.703724 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/da496881-c195-4cc2-9277-c8118bf651dc-kube-api-access-qcfsr" (OuterVolumeSpecName: "kube-api-access-qcfsr") pod "da496881-c195-4cc2-9277-c8118bf651dc" (UID: "da496881-c195-4cc2-9277-c8118bf651dc"). InnerVolumeSpecName "kube-api-access-qcfsr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:24:05 crc kubenswrapper[4860]: I1211 08:24:05.709274 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/da496881-c195-4cc2-9277-c8118bf651dc-util" (OuterVolumeSpecName: "util") pod "da496881-c195-4cc2-9277-c8118bf651dc" (UID: "da496881-c195-4cc2-9277-c8118bf651dc"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:24:05 crc kubenswrapper[4860]: I1211 08:24:05.795867 4860 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/da496881-c195-4cc2-9277-c8118bf651dc-util\") on node \"crc\" DevicePath \"\"" Dec 11 08:24:05 crc kubenswrapper[4860]: I1211 08:24:05.795915 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qcfsr\" (UniqueName: \"kubernetes.io/projected/da496881-c195-4cc2-9277-c8118bf651dc-kube-api-access-qcfsr\") on node \"crc\" DevicePath \"\"" Dec 11 08:24:05 crc kubenswrapper[4860]: I1211 08:24:05.795928 4860 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/da496881-c195-4cc2-9277-c8118bf651dc-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 08:24:06 crc kubenswrapper[4860]: I1211 08:24:06.259564 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4vr6tn" event={"ID":"da496881-c195-4cc2-9277-c8118bf651dc","Type":"ContainerDied","Data":"c7497b24fa1ed3880c1ec828ba86d83ae2a50e7d846d4c2dba165c0cedc5c518"} Dec 11 08:24:06 crc kubenswrapper[4860]: I1211 08:24:06.259633 4860 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c7497b24fa1ed3880c1ec828ba86d83ae2a50e7d846d4c2dba165c0cedc5c518" Dec 11 08:24:06 crc kubenswrapper[4860]: I1211 08:24:06.259773 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4vr6tn" Dec 11 08:24:06 crc kubenswrapper[4860]: I1211 08:24:06.262896 4860 generic.go:334] "Generic (PLEG): container finished" podID="dfaade06-9008-441e-ad04-990d22253a33" containerID="28dd3fe0c406822334dc10a7ae05f59f6ba067d50492b59f8912f4b1eed0f4c2" exitCode=0 Dec 11 08:24:06 crc kubenswrapper[4860]: I1211 08:24:06.263063 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hbw2q" event={"ID":"dfaade06-9008-441e-ad04-990d22253a33","Type":"ContainerDied","Data":"28dd3fe0c406822334dc10a7ae05f59f6ba067d50492b59f8912f4b1eed0f4c2"} Dec 11 08:24:07 crc kubenswrapper[4860]: I1211 08:24:07.270732 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hbw2q" event={"ID":"dfaade06-9008-441e-ad04-990d22253a33","Type":"ContainerStarted","Data":"f5a7b7bff8adc9a8d9d0f093a674130fe922df47d6d0ee6a1a35638217c3b560"} Dec 11 08:24:07 crc kubenswrapper[4860]: I1211 08:24:07.297829 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-hbw2q" podStartSLOduration=1.792989515 podStartE2EDuration="4.29780644s" podCreationTimestamp="2025-12-11 08:24:03 +0000 UTC" firstStartedPulling="2025-12-11 08:24:04.247232576 +0000 UTC m=+776.975751631" lastFinishedPulling="2025-12-11 08:24:06.752049501 +0000 UTC m=+779.480568556" observedRunningTime="2025-12-11 08:24:07.293510335 +0000 UTC m=+780.022029410" watchObservedRunningTime="2025-12-11 08:24:07.29780644 +0000 UTC m=+780.026325495" Dec 11 08:24:07 crc kubenswrapper[4860]: E1211 08:24:07.514432 4860 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/system.slice/NetworkManager-dispatcher.service\": RecentStats: unable to find data in memory cache]" Dec 11 08:24:08 crc kubenswrapper[4860]: I1211 08:24:08.795669 4860 patch_prober.go:28] interesting pod/machine-config-daemon-99qgp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 08:24:08 crc kubenswrapper[4860]: I1211 08:24:08.796337 4860 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 08:24:13 crc kubenswrapper[4860]: I1211 08:24:13.792819 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-hbw2q" Dec 11 08:24:13 crc kubenswrapper[4860]: I1211 08:24:13.794809 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-hbw2q" Dec 11 08:24:13 crc kubenswrapper[4860]: I1211 08:24:13.850767 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-hbw2q" Dec 11 08:24:14 crc kubenswrapper[4860]: I1211 08:24:14.358457 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-hbw2q" Dec 11 08:24:15 crc kubenswrapper[4860]: I1211 08:24:15.305428 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-d5785c6bc-n9n8q"] Dec 11 08:24:15 crc kubenswrapper[4860]: E1211 08:24:15.305747 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da496881-c195-4cc2-9277-c8118bf651dc" containerName="pull" Dec 11 08:24:15 crc kubenswrapper[4860]: I1211 08:24:15.305764 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="da496881-c195-4cc2-9277-c8118bf651dc" containerName="pull" Dec 11 08:24:15 crc kubenswrapper[4860]: E1211 08:24:15.305776 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da496881-c195-4cc2-9277-c8118bf651dc" containerName="extract" Dec 11 08:24:15 crc kubenswrapper[4860]: I1211 08:24:15.305782 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="da496881-c195-4cc2-9277-c8118bf651dc" containerName="extract" Dec 11 08:24:15 crc kubenswrapper[4860]: E1211 08:24:15.305801 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da496881-c195-4cc2-9277-c8118bf651dc" containerName="util" Dec 11 08:24:15 crc kubenswrapper[4860]: I1211 08:24:15.305807 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="da496881-c195-4cc2-9277-c8118bf651dc" containerName="util" Dec 11 08:24:15 crc kubenswrapper[4860]: I1211 08:24:15.305915 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="da496881-c195-4cc2-9277-c8118bf651dc" containerName="extract" Dec 11 08:24:15 crc kubenswrapper[4860]: I1211 08:24:15.306413 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-d5785c6bc-n9n8q" Dec 11 08:24:15 crc kubenswrapper[4860]: I1211 08:24:15.309944 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Dec 11 08:24:15 crc kubenswrapper[4860]: I1211 08:24:15.310252 4860 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Dec 11 08:24:15 crc kubenswrapper[4860]: I1211 08:24:15.310454 4860 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-7kljd" Dec 11 08:24:15 crc kubenswrapper[4860]: I1211 08:24:15.311477 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Dec 11 08:24:15 crc kubenswrapper[4860]: I1211 08:24:15.316168 4860 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Dec 11 08:24:15 crc kubenswrapper[4860]: I1211 08:24:15.336758 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-d5785c6bc-n9n8q"] Dec 11 08:24:15 crc kubenswrapper[4860]: I1211 08:24:15.441913 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e0e047cd-d665-4b39-8588-b0d396e3b36d-apiservice-cert\") pod \"metallb-operator-controller-manager-d5785c6bc-n9n8q\" (UID: \"e0e047cd-d665-4b39-8588-b0d396e3b36d\") " pod="metallb-system/metallb-operator-controller-manager-d5785c6bc-n9n8q" Dec 11 08:24:15 crc kubenswrapper[4860]: I1211 08:24:15.441990 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e0e047cd-d665-4b39-8588-b0d396e3b36d-webhook-cert\") pod \"metallb-operator-controller-manager-d5785c6bc-n9n8q\" (UID: \"e0e047cd-d665-4b39-8588-b0d396e3b36d\") " pod="metallb-system/metallb-operator-controller-manager-d5785c6bc-n9n8q" Dec 11 08:24:15 crc kubenswrapper[4860]: I1211 08:24:15.442024 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-72qcb\" (UniqueName: \"kubernetes.io/projected/e0e047cd-d665-4b39-8588-b0d396e3b36d-kube-api-access-72qcb\") pod \"metallb-operator-controller-manager-d5785c6bc-n9n8q\" (UID: \"e0e047cd-d665-4b39-8588-b0d396e3b36d\") " pod="metallb-system/metallb-operator-controller-manager-d5785c6bc-n9n8q" Dec 11 08:24:15 crc kubenswrapper[4860]: I1211 08:24:15.544174 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e0e047cd-d665-4b39-8588-b0d396e3b36d-webhook-cert\") pod \"metallb-operator-controller-manager-d5785c6bc-n9n8q\" (UID: \"e0e047cd-d665-4b39-8588-b0d396e3b36d\") " pod="metallb-system/metallb-operator-controller-manager-d5785c6bc-n9n8q" Dec 11 08:24:15 crc kubenswrapper[4860]: I1211 08:24:15.544262 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-72qcb\" (UniqueName: \"kubernetes.io/projected/e0e047cd-d665-4b39-8588-b0d396e3b36d-kube-api-access-72qcb\") pod \"metallb-operator-controller-manager-d5785c6bc-n9n8q\" (UID: \"e0e047cd-d665-4b39-8588-b0d396e3b36d\") " pod="metallb-system/metallb-operator-controller-manager-d5785c6bc-n9n8q" Dec 11 08:24:15 crc kubenswrapper[4860]: I1211 08:24:15.544346 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e0e047cd-d665-4b39-8588-b0d396e3b36d-apiservice-cert\") pod \"metallb-operator-controller-manager-d5785c6bc-n9n8q\" (UID: \"e0e047cd-d665-4b39-8588-b0d396e3b36d\") " pod="metallb-system/metallb-operator-controller-manager-d5785c6bc-n9n8q" Dec 11 08:24:15 crc kubenswrapper[4860]: I1211 08:24:15.554023 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e0e047cd-d665-4b39-8588-b0d396e3b36d-apiservice-cert\") pod \"metallb-operator-controller-manager-d5785c6bc-n9n8q\" (UID: \"e0e047cd-d665-4b39-8588-b0d396e3b36d\") " pod="metallb-system/metallb-operator-controller-manager-d5785c6bc-n9n8q" Dec 11 08:24:15 crc kubenswrapper[4860]: I1211 08:24:15.559447 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e0e047cd-d665-4b39-8588-b0d396e3b36d-webhook-cert\") pod \"metallb-operator-controller-manager-d5785c6bc-n9n8q\" (UID: \"e0e047cd-d665-4b39-8588-b0d396e3b36d\") " pod="metallb-system/metallb-operator-controller-manager-d5785c6bc-n9n8q" Dec 11 08:24:15 crc kubenswrapper[4860]: I1211 08:24:15.571210 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-72qcb\" (UniqueName: \"kubernetes.io/projected/e0e047cd-d665-4b39-8588-b0d396e3b36d-kube-api-access-72qcb\") pod \"metallb-operator-controller-manager-d5785c6bc-n9n8q\" (UID: \"e0e047cd-d665-4b39-8588-b0d396e3b36d\") " pod="metallb-system/metallb-operator-controller-manager-d5785c6bc-n9n8q" Dec 11 08:24:15 crc kubenswrapper[4860]: I1211 08:24:15.630864 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-d5785c6bc-n9n8q" Dec 11 08:24:15 crc kubenswrapper[4860]: I1211 08:24:15.685195 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-59475655bd-8sqgf"] Dec 11 08:24:15 crc kubenswrapper[4860]: I1211 08:24:15.686105 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-59475655bd-8sqgf" Dec 11 08:24:15 crc kubenswrapper[4860]: I1211 08:24:15.690305 4860 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Dec 11 08:24:15 crc kubenswrapper[4860]: I1211 08:24:15.690355 4860 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-2dfhg" Dec 11 08:24:15 crc kubenswrapper[4860]: I1211 08:24:15.690312 4860 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 11 08:24:15 crc kubenswrapper[4860]: I1211 08:24:15.722663 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-59475655bd-8sqgf"] Dec 11 08:24:15 crc kubenswrapper[4860]: I1211 08:24:15.849122 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8lmzc\" (UniqueName: \"kubernetes.io/projected/47efc842-7de4-461c-bafc-0fbaabb6c3ad-kube-api-access-8lmzc\") pod \"metallb-operator-webhook-server-59475655bd-8sqgf\" (UID: \"47efc842-7de4-461c-bafc-0fbaabb6c3ad\") " pod="metallb-system/metallb-operator-webhook-server-59475655bd-8sqgf" Dec 11 08:24:15 crc kubenswrapper[4860]: I1211 08:24:15.849223 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/47efc842-7de4-461c-bafc-0fbaabb6c3ad-apiservice-cert\") pod \"metallb-operator-webhook-server-59475655bd-8sqgf\" (UID: \"47efc842-7de4-461c-bafc-0fbaabb6c3ad\") " pod="metallb-system/metallb-operator-webhook-server-59475655bd-8sqgf" Dec 11 08:24:15 crc kubenswrapper[4860]: I1211 08:24:15.849277 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/47efc842-7de4-461c-bafc-0fbaabb6c3ad-webhook-cert\") pod \"metallb-operator-webhook-server-59475655bd-8sqgf\" (UID: \"47efc842-7de4-461c-bafc-0fbaabb6c3ad\") " pod="metallb-system/metallb-operator-webhook-server-59475655bd-8sqgf" Dec 11 08:24:15 crc kubenswrapper[4860]: I1211 08:24:15.950722 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8lmzc\" (UniqueName: \"kubernetes.io/projected/47efc842-7de4-461c-bafc-0fbaabb6c3ad-kube-api-access-8lmzc\") pod \"metallb-operator-webhook-server-59475655bd-8sqgf\" (UID: \"47efc842-7de4-461c-bafc-0fbaabb6c3ad\") " pod="metallb-system/metallb-operator-webhook-server-59475655bd-8sqgf" Dec 11 08:24:15 crc kubenswrapper[4860]: I1211 08:24:15.950824 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/47efc842-7de4-461c-bafc-0fbaabb6c3ad-apiservice-cert\") pod \"metallb-operator-webhook-server-59475655bd-8sqgf\" (UID: \"47efc842-7de4-461c-bafc-0fbaabb6c3ad\") " pod="metallb-system/metallb-operator-webhook-server-59475655bd-8sqgf" Dec 11 08:24:15 crc kubenswrapper[4860]: I1211 08:24:15.950890 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/47efc842-7de4-461c-bafc-0fbaabb6c3ad-webhook-cert\") pod \"metallb-operator-webhook-server-59475655bd-8sqgf\" (UID: \"47efc842-7de4-461c-bafc-0fbaabb6c3ad\") " pod="metallb-system/metallb-operator-webhook-server-59475655bd-8sqgf" Dec 11 08:24:15 crc kubenswrapper[4860]: I1211 08:24:15.959200 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/47efc842-7de4-461c-bafc-0fbaabb6c3ad-apiservice-cert\") pod \"metallb-operator-webhook-server-59475655bd-8sqgf\" (UID: \"47efc842-7de4-461c-bafc-0fbaabb6c3ad\") " pod="metallb-system/metallb-operator-webhook-server-59475655bd-8sqgf" Dec 11 08:24:15 crc kubenswrapper[4860]: I1211 08:24:15.991639 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/47efc842-7de4-461c-bafc-0fbaabb6c3ad-webhook-cert\") pod \"metallb-operator-webhook-server-59475655bd-8sqgf\" (UID: \"47efc842-7de4-461c-bafc-0fbaabb6c3ad\") " pod="metallb-system/metallb-operator-webhook-server-59475655bd-8sqgf" Dec 11 08:24:15 crc kubenswrapper[4860]: I1211 08:24:15.992146 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8lmzc\" (UniqueName: \"kubernetes.io/projected/47efc842-7de4-461c-bafc-0fbaabb6c3ad-kube-api-access-8lmzc\") pod \"metallb-operator-webhook-server-59475655bd-8sqgf\" (UID: \"47efc842-7de4-461c-bafc-0fbaabb6c3ad\") " pod="metallb-system/metallb-operator-webhook-server-59475655bd-8sqgf" Dec 11 08:24:16 crc kubenswrapper[4860]: I1211 08:24:16.024129 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-59475655bd-8sqgf" Dec 11 08:24:16 crc kubenswrapper[4860]: I1211 08:24:16.077569 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-d5785c6bc-n9n8q"] Dec 11 08:24:16 crc kubenswrapper[4860]: I1211 08:24:16.332603 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-d5785c6bc-n9n8q" event={"ID":"e0e047cd-d665-4b39-8588-b0d396e3b36d","Type":"ContainerStarted","Data":"e938f7b4be2697fb006cf459bf370e667a7666ba30cc0aff6ca996e2e72fd8fa"} Dec 11 08:24:16 crc kubenswrapper[4860]: I1211 08:24:16.622581 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-59475655bd-8sqgf"] Dec 11 08:24:16 crc kubenswrapper[4860]: W1211 08:24:16.629669 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod47efc842_7de4_461c_bafc_0fbaabb6c3ad.slice/crio-b1c1b93f24c4a6d739f2016dbeb8d600274c0035d5448182ebb2816832db886f WatchSource:0}: Error finding container b1c1b93f24c4a6d739f2016dbeb8d600274c0035d5448182ebb2816832db886f: Status 404 returned error can't find the container with id b1c1b93f24c4a6d739f2016dbeb8d600274c0035d5448182ebb2816832db886f Dec 11 08:24:17 crc kubenswrapper[4860]: I1211 08:24:17.341012 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-59475655bd-8sqgf" event={"ID":"47efc842-7de4-461c-bafc-0fbaabb6c3ad","Type":"ContainerStarted","Data":"b1c1b93f24c4a6d739f2016dbeb8d600274c0035d5448182ebb2816832db886f"} Dec 11 08:24:17 crc kubenswrapper[4860]: I1211 08:24:17.659378 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hbw2q"] Dec 11 08:24:17 crc kubenswrapper[4860]: I1211 08:24:17.659780 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-hbw2q" podUID="dfaade06-9008-441e-ad04-990d22253a33" containerName="registry-server" containerID="cri-o://f5a7b7bff8adc9a8d9d0f093a674130fe922df47d6d0ee6a1a35638217c3b560" gracePeriod=2 Dec 11 08:24:20 crc kubenswrapper[4860]: I1211 08:24:20.296539 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hbw2q" Dec 11 08:24:20 crc kubenswrapper[4860]: I1211 08:24:20.363592 4860 generic.go:334] "Generic (PLEG): container finished" podID="dfaade06-9008-441e-ad04-990d22253a33" containerID="f5a7b7bff8adc9a8d9d0f093a674130fe922df47d6d0ee6a1a35638217c3b560" exitCode=0 Dec 11 08:24:20 crc kubenswrapper[4860]: I1211 08:24:20.363675 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hbw2q" event={"ID":"dfaade06-9008-441e-ad04-990d22253a33","Type":"ContainerDied","Data":"f5a7b7bff8adc9a8d9d0f093a674130fe922df47d6d0ee6a1a35638217c3b560"} Dec 11 08:24:20 crc kubenswrapper[4860]: I1211 08:24:20.363718 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hbw2q" event={"ID":"dfaade06-9008-441e-ad04-990d22253a33","Type":"ContainerDied","Data":"6c58ce405276391d579a713f700fd069e359c0483cc35f57e2864a669b159826"} Dec 11 08:24:20 crc kubenswrapper[4860]: I1211 08:24:20.363744 4860 scope.go:117] "RemoveContainer" containerID="f5a7b7bff8adc9a8d9d0f093a674130fe922df47d6d0ee6a1a35638217c3b560" Dec 11 08:24:20 crc kubenswrapper[4860]: I1211 08:24:20.363916 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hbw2q" Dec 11 08:24:20 crc kubenswrapper[4860]: I1211 08:24:20.397932 4860 scope.go:117] "RemoveContainer" containerID="28dd3fe0c406822334dc10a7ae05f59f6ba067d50492b59f8912f4b1eed0f4c2" Dec 11 08:24:20 crc kubenswrapper[4860]: I1211 08:24:20.423021 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dfaade06-9008-441e-ad04-990d22253a33-utilities\") pod \"dfaade06-9008-441e-ad04-990d22253a33\" (UID: \"dfaade06-9008-441e-ad04-990d22253a33\") " Dec 11 08:24:20 crc kubenswrapper[4860]: I1211 08:24:20.423120 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dmhg2\" (UniqueName: \"kubernetes.io/projected/dfaade06-9008-441e-ad04-990d22253a33-kube-api-access-dmhg2\") pod \"dfaade06-9008-441e-ad04-990d22253a33\" (UID: \"dfaade06-9008-441e-ad04-990d22253a33\") " Dec 11 08:24:20 crc kubenswrapper[4860]: I1211 08:24:20.423224 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dfaade06-9008-441e-ad04-990d22253a33-catalog-content\") pod \"dfaade06-9008-441e-ad04-990d22253a33\" (UID: \"dfaade06-9008-441e-ad04-990d22253a33\") " Dec 11 08:24:20 crc kubenswrapper[4860]: I1211 08:24:20.424734 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dfaade06-9008-441e-ad04-990d22253a33-utilities" (OuterVolumeSpecName: "utilities") pod "dfaade06-9008-441e-ad04-990d22253a33" (UID: "dfaade06-9008-441e-ad04-990d22253a33"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:24:20 crc kubenswrapper[4860]: I1211 08:24:20.426171 4860 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dfaade06-9008-441e-ad04-990d22253a33-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 08:24:20 crc kubenswrapper[4860]: I1211 08:24:20.428284 4860 scope.go:117] "RemoveContainer" containerID="f2f937bf6a7eafbf269831b9791c4f63b0e752e769d5b2f63d1e262a14f4e2f3" Dec 11 08:24:20 crc kubenswrapper[4860]: I1211 08:24:20.432390 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dfaade06-9008-441e-ad04-990d22253a33-kube-api-access-dmhg2" (OuterVolumeSpecName: "kube-api-access-dmhg2") pod "dfaade06-9008-441e-ad04-990d22253a33" (UID: "dfaade06-9008-441e-ad04-990d22253a33"). InnerVolumeSpecName "kube-api-access-dmhg2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:24:20 crc kubenswrapper[4860]: I1211 08:24:20.471037 4860 scope.go:117] "RemoveContainer" containerID="f5a7b7bff8adc9a8d9d0f093a674130fe922df47d6d0ee6a1a35638217c3b560" Dec 11 08:24:20 crc kubenswrapper[4860]: E1211 08:24:20.471863 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f5a7b7bff8adc9a8d9d0f093a674130fe922df47d6d0ee6a1a35638217c3b560\": container with ID starting with f5a7b7bff8adc9a8d9d0f093a674130fe922df47d6d0ee6a1a35638217c3b560 not found: ID does not exist" containerID="f5a7b7bff8adc9a8d9d0f093a674130fe922df47d6d0ee6a1a35638217c3b560" Dec 11 08:24:20 crc kubenswrapper[4860]: I1211 08:24:20.471942 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f5a7b7bff8adc9a8d9d0f093a674130fe922df47d6d0ee6a1a35638217c3b560"} err="failed to get container status \"f5a7b7bff8adc9a8d9d0f093a674130fe922df47d6d0ee6a1a35638217c3b560\": rpc error: code = NotFound desc = could not find container \"f5a7b7bff8adc9a8d9d0f093a674130fe922df47d6d0ee6a1a35638217c3b560\": container with ID starting with f5a7b7bff8adc9a8d9d0f093a674130fe922df47d6d0ee6a1a35638217c3b560 not found: ID does not exist" Dec 11 08:24:20 crc kubenswrapper[4860]: I1211 08:24:20.471990 4860 scope.go:117] "RemoveContainer" containerID="28dd3fe0c406822334dc10a7ae05f59f6ba067d50492b59f8912f4b1eed0f4c2" Dec 11 08:24:20 crc kubenswrapper[4860]: E1211 08:24:20.472325 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"28dd3fe0c406822334dc10a7ae05f59f6ba067d50492b59f8912f4b1eed0f4c2\": container with ID starting with 28dd3fe0c406822334dc10a7ae05f59f6ba067d50492b59f8912f4b1eed0f4c2 not found: ID does not exist" containerID="28dd3fe0c406822334dc10a7ae05f59f6ba067d50492b59f8912f4b1eed0f4c2" Dec 11 08:24:20 crc kubenswrapper[4860]: I1211 08:24:20.472361 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28dd3fe0c406822334dc10a7ae05f59f6ba067d50492b59f8912f4b1eed0f4c2"} err="failed to get container status \"28dd3fe0c406822334dc10a7ae05f59f6ba067d50492b59f8912f4b1eed0f4c2\": rpc error: code = NotFound desc = could not find container \"28dd3fe0c406822334dc10a7ae05f59f6ba067d50492b59f8912f4b1eed0f4c2\": container with ID starting with 28dd3fe0c406822334dc10a7ae05f59f6ba067d50492b59f8912f4b1eed0f4c2 not found: ID does not exist" Dec 11 08:24:20 crc kubenswrapper[4860]: I1211 08:24:20.472383 4860 scope.go:117] "RemoveContainer" containerID="f2f937bf6a7eafbf269831b9791c4f63b0e752e769d5b2f63d1e262a14f4e2f3" Dec 11 08:24:20 crc kubenswrapper[4860]: E1211 08:24:20.472703 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f2f937bf6a7eafbf269831b9791c4f63b0e752e769d5b2f63d1e262a14f4e2f3\": container with ID starting with f2f937bf6a7eafbf269831b9791c4f63b0e752e769d5b2f63d1e262a14f4e2f3 not found: ID does not exist" containerID="f2f937bf6a7eafbf269831b9791c4f63b0e752e769d5b2f63d1e262a14f4e2f3" Dec 11 08:24:20 crc kubenswrapper[4860]: I1211 08:24:20.472736 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f2f937bf6a7eafbf269831b9791c4f63b0e752e769d5b2f63d1e262a14f4e2f3"} err="failed to get container status \"f2f937bf6a7eafbf269831b9791c4f63b0e752e769d5b2f63d1e262a14f4e2f3\": rpc error: code = NotFound desc = could not find container \"f2f937bf6a7eafbf269831b9791c4f63b0e752e769d5b2f63d1e262a14f4e2f3\": container with ID starting with f2f937bf6a7eafbf269831b9791c4f63b0e752e769d5b2f63d1e262a14f4e2f3 not found: ID does not exist" Dec 11 08:24:20 crc kubenswrapper[4860]: I1211 08:24:20.527427 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dmhg2\" (UniqueName: \"kubernetes.io/projected/dfaade06-9008-441e-ad04-990d22253a33-kube-api-access-dmhg2\") on node \"crc\" DevicePath \"\"" Dec 11 08:24:20 crc kubenswrapper[4860]: I1211 08:24:20.561573 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dfaade06-9008-441e-ad04-990d22253a33-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "dfaade06-9008-441e-ad04-990d22253a33" (UID: "dfaade06-9008-441e-ad04-990d22253a33"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:24:20 crc kubenswrapper[4860]: I1211 08:24:20.629260 4860 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dfaade06-9008-441e-ad04-990d22253a33-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 08:24:20 crc kubenswrapper[4860]: I1211 08:24:20.700831 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hbw2q"] Dec 11 08:24:20 crc kubenswrapper[4860]: I1211 08:24:20.735081 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-hbw2q"] Dec 11 08:24:21 crc kubenswrapper[4860]: I1211 08:24:21.606048 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dfaade06-9008-441e-ad04-990d22253a33" path="/var/lib/kubelet/pods/dfaade06-9008-441e-ad04-990d22253a33/volumes" Dec 11 08:24:24 crc kubenswrapper[4860]: I1211 08:24:24.400034 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-d5785c6bc-n9n8q" event={"ID":"e0e047cd-d665-4b39-8588-b0d396e3b36d","Type":"ContainerStarted","Data":"770a479100f3137ad9b6873625b9ea24eaa57095dc023ae8b9a39cb8f6343fce"} Dec 11 08:24:24 crc kubenswrapper[4860]: I1211 08:24:24.400574 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-d5785c6bc-n9n8q" Dec 11 08:24:24 crc kubenswrapper[4860]: I1211 08:24:24.402332 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-59475655bd-8sqgf" event={"ID":"47efc842-7de4-461c-bafc-0fbaabb6c3ad","Type":"ContainerStarted","Data":"452d2268f5c2b9c4df22cdada4f88dfe37a0635a20c7c9a8413e2eac015b64b3"} Dec 11 08:24:24 crc kubenswrapper[4860]: I1211 08:24:24.402497 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-59475655bd-8sqgf" Dec 11 08:24:24 crc kubenswrapper[4860]: I1211 08:24:24.425102 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-d5785c6bc-n9n8q" podStartSLOduration=2.113656207 podStartE2EDuration="9.425072542s" podCreationTimestamp="2025-12-11 08:24:15 +0000 UTC" firstStartedPulling="2025-12-11 08:24:16.128066373 +0000 UTC m=+788.856585428" lastFinishedPulling="2025-12-11 08:24:23.439482708 +0000 UTC m=+796.168001763" observedRunningTime="2025-12-11 08:24:24.423011976 +0000 UTC m=+797.151531031" watchObservedRunningTime="2025-12-11 08:24:24.425072542 +0000 UTC m=+797.153591597" Dec 11 08:24:24 crc kubenswrapper[4860]: I1211 08:24:24.446416 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-59475655bd-8sqgf" podStartSLOduration=2.620781887 podStartE2EDuration="9.446392745s" podCreationTimestamp="2025-12-11 08:24:15 +0000 UTC" firstStartedPulling="2025-12-11 08:24:16.632964972 +0000 UTC m=+789.361484027" lastFinishedPulling="2025-12-11 08:24:23.45857583 +0000 UTC m=+796.187094885" observedRunningTime="2025-12-11 08:24:24.445810696 +0000 UTC m=+797.174329761" watchObservedRunningTime="2025-12-11 08:24:24.446392745 +0000 UTC m=+797.174911800" Dec 11 08:24:36 crc kubenswrapper[4860]: I1211 08:24:36.040407 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-59475655bd-8sqgf" Dec 11 08:24:38 crc kubenswrapper[4860]: I1211 08:24:38.795134 4860 patch_prober.go:28] interesting pod/machine-config-daemon-99qgp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 08:24:38 crc kubenswrapper[4860]: I1211 08:24:38.795753 4860 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 08:24:55 crc kubenswrapper[4860]: I1211 08:24:55.634790 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-d5785c6bc-n9n8q" Dec 11 08:24:56 crc kubenswrapper[4860]: I1211 08:24:56.494622 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-c6crs"] Dec 11 08:24:56 crc kubenswrapper[4860]: E1211 08:24:56.495026 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dfaade06-9008-441e-ad04-990d22253a33" containerName="registry-server" Dec 11 08:24:56 crc kubenswrapper[4860]: I1211 08:24:56.495047 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="dfaade06-9008-441e-ad04-990d22253a33" containerName="registry-server" Dec 11 08:24:56 crc kubenswrapper[4860]: E1211 08:24:56.495063 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dfaade06-9008-441e-ad04-990d22253a33" containerName="extract-content" Dec 11 08:24:56 crc kubenswrapper[4860]: I1211 08:24:56.495072 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="dfaade06-9008-441e-ad04-990d22253a33" containerName="extract-content" Dec 11 08:24:56 crc kubenswrapper[4860]: E1211 08:24:56.495093 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dfaade06-9008-441e-ad04-990d22253a33" containerName="extract-utilities" Dec 11 08:24:56 crc kubenswrapper[4860]: I1211 08:24:56.495103 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="dfaade06-9008-441e-ad04-990d22253a33" containerName="extract-utilities" Dec 11 08:24:56 crc kubenswrapper[4860]: I1211 08:24:56.495302 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="dfaade06-9008-441e-ad04-990d22253a33" containerName="registry-server" Dec 11 08:24:56 crc kubenswrapper[4860]: I1211 08:24:56.497992 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-c6crs" Dec 11 08:24:56 crc kubenswrapper[4860]: I1211 08:24:56.500326 4860 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Dec 11 08:24:56 crc kubenswrapper[4860]: I1211 08:24:56.501421 4860 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-fmkmq" Dec 11 08:24:56 crc kubenswrapper[4860]: I1211 08:24:56.509786 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-7784b6fcf-2548h"] Dec 11 08:24:56 crc kubenswrapper[4860]: I1211 08:24:56.510720 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-2548h" Dec 11 08:24:56 crc kubenswrapper[4860]: I1211 08:24:56.511629 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Dec 11 08:24:56 crc kubenswrapper[4860]: I1211 08:24:56.513431 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/130f6554-c476-49a0-8270-593008c1ffc6-metrics\") pod \"frr-k8s-c6crs\" (UID: \"130f6554-c476-49a0-8270-593008c1ffc6\") " pod="metallb-system/frr-k8s-c6crs" Dec 11 08:24:56 crc kubenswrapper[4860]: I1211 08:24:56.513512 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l84rk\" (UniqueName: \"kubernetes.io/projected/98afb3e0-d363-472c-886e-2b2981cb13a4-kube-api-access-l84rk\") pod \"frr-k8s-webhook-server-7784b6fcf-2548h\" (UID: \"98afb3e0-d363-472c-886e-2b2981cb13a4\") " pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-2548h" Dec 11 08:24:56 crc kubenswrapper[4860]: I1211 08:24:56.513558 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-78mrl\" (UniqueName: \"kubernetes.io/projected/130f6554-c476-49a0-8270-593008c1ffc6-kube-api-access-78mrl\") pod \"frr-k8s-c6crs\" (UID: \"130f6554-c476-49a0-8270-593008c1ffc6\") " pod="metallb-system/frr-k8s-c6crs" Dec 11 08:24:56 crc kubenswrapper[4860]: I1211 08:24:56.513599 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/130f6554-c476-49a0-8270-593008c1ffc6-reloader\") pod \"frr-k8s-c6crs\" (UID: \"130f6554-c476-49a0-8270-593008c1ffc6\") " pod="metallb-system/frr-k8s-c6crs" Dec 11 08:24:56 crc kubenswrapper[4860]: I1211 08:24:56.513631 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/98afb3e0-d363-472c-886e-2b2981cb13a4-cert\") pod \"frr-k8s-webhook-server-7784b6fcf-2548h\" (UID: \"98afb3e0-d363-472c-886e-2b2981cb13a4\") " pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-2548h" Dec 11 08:24:56 crc kubenswrapper[4860]: I1211 08:24:56.513683 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/130f6554-c476-49a0-8270-593008c1ffc6-frr-sockets\") pod \"frr-k8s-c6crs\" (UID: \"130f6554-c476-49a0-8270-593008c1ffc6\") " pod="metallb-system/frr-k8s-c6crs" Dec 11 08:24:56 crc kubenswrapper[4860]: I1211 08:24:56.513712 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/130f6554-c476-49a0-8270-593008c1ffc6-frr-startup\") pod \"frr-k8s-c6crs\" (UID: \"130f6554-c476-49a0-8270-593008c1ffc6\") " pod="metallb-system/frr-k8s-c6crs" Dec 11 08:24:56 crc kubenswrapper[4860]: I1211 08:24:56.513751 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/130f6554-c476-49a0-8270-593008c1ffc6-frr-conf\") pod \"frr-k8s-c6crs\" (UID: \"130f6554-c476-49a0-8270-593008c1ffc6\") " pod="metallb-system/frr-k8s-c6crs" Dec 11 08:24:56 crc kubenswrapper[4860]: I1211 08:24:56.513779 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/130f6554-c476-49a0-8270-593008c1ffc6-metrics-certs\") pod \"frr-k8s-c6crs\" (UID: \"130f6554-c476-49a0-8270-593008c1ffc6\") " pod="metallb-system/frr-k8s-c6crs" Dec 11 08:24:56 crc kubenswrapper[4860]: I1211 08:24:56.514166 4860 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Dec 11 08:24:56 crc kubenswrapper[4860]: I1211 08:24:56.545826 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7784b6fcf-2548h"] Dec 11 08:24:56 crc kubenswrapper[4860]: I1211 08:24:56.617071 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l84rk\" (UniqueName: \"kubernetes.io/projected/98afb3e0-d363-472c-886e-2b2981cb13a4-kube-api-access-l84rk\") pod \"frr-k8s-webhook-server-7784b6fcf-2548h\" (UID: \"98afb3e0-d363-472c-886e-2b2981cb13a4\") " pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-2548h" Dec 11 08:24:56 crc kubenswrapper[4860]: I1211 08:24:56.617153 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-78mrl\" (UniqueName: \"kubernetes.io/projected/130f6554-c476-49a0-8270-593008c1ffc6-kube-api-access-78mrl\") pod \"frr-k8s-c6crs\" (UID: \"130f6554-c476-49a0-8270-593008c1ffc6\") " pod="metallb-system/frr-k8s-c6crs" Dec 11 08:24:56 crc kubenswrapper[4860]: I1211 08:24:56.617195 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/130f6554-c476-49a0-8270-593008c1ffc6-reloader\") pod \"frr-k8s-c6crs\" (UID: \"130f6554-c476-49a0-8270-593008c1ffc6\") " pod="metallb-system/frr-k8s-c6crs" Dec 11 08:24:56 crc kubenswrapper[4860]: I1211 08:24:56.617241 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/98afb3e0-d363-472c-886e-2b2981cb13a4-cert\") pod \"frr-k8s-webhook-server-7784b6fcf-2548h\" (UID: \"98afb3e0-d363-472c-886e-2b2981cb13a4\") " pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-2548h" Dec 11 08:24:56 crc kubenswrapper[4860]: I1211 08:24:56.617262 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/130f6554-c476-49a0-8270-593008c1ffc6-frr-sockets\") pod \"frr-k8s-c6crs\" (UID: \"130f6554-c476-49a0-8270-593008c1ffc6\") " pod="metallb-system/frr-k8s-c6crs" Dec 11 08:24:56 crc kubenswrapper[4860]: I1211 08:24:56.617285 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/130f6554-c476-49a0-8270-593008c1ffc6-frr-startup\") pod \"frr-k8s-c6crs\" (UID: \"130f6554-c476-49a0-8270-593008c1ffc6\") " pod="metallb-system/frr-k8s-c6crs" Dec 11 08:24:56 crc kubenswrapper[4860]: I1211 08:24:56.617313 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/130f6554-c476-49a0-8270-593008c1ffc6-frr-conf\") pod \"frr-k8s-c6crs\" (UID: \"130f6554-c476-49a0-8270-593008c1ffc6\") " pod="metallb-system/frr-k8s-c6crs" Dec 11 08:24:56 crc kubenswrapper[4860]: I1211 08:24:56.617335 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/130f6554-c476-49a0-8270-593008c1ffc6-metrics-certs\") pod \"frr-k8s-c6crs\" (UID: \"130f6554-c476-49a0-8270-593008c1ffc6\") " pod="metallb-system/frr-k8s-c6crs" Dec 11 08:24:56 crc kubenswrapper[4860]: I1211 08:24:56.617364 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/130f6554-c476-49a0-8270-593008c1ffc6-metrics\") pod \"frr-k8s-c6crs\" (UID: \"130f6554-c476-49a0-8270-593008c1ffc6\") " pod="metallb-system/frr-k8s-c6crs" Dec 11 08:24:56 crc kubenswrapper[4860]: I1211 08:24:56.617832 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/130f6554-c476-49a0-8270-593008c1ffc6-metrics\") pod \"frr-k8s-c6crs\" (UID: \"130f6554-c476-49a0-8270-593008c1ffc6\") " pod="metallb-system/frr-k8s-c6crs" Dec 11 08:24:56 crc kubenswrapper[4860]: I1211 08:24:56.618460 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/130f6554-c476-49a0-8270-593008c1ffc6-reloader\") pod \"frr-k8s-c6crs\" (UID: \"130f6554-c476-49a0-8270-593008c1ffc6\") " pod="metallb-system/frr-k8s-c6crs" Dec 11 08:24:56 crc kubenswrapper[4860]: I1211 08:24:56.618565 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/130f6554-c476-49a0-8270-593008c1ffc6-frr-sockets\") pod \"frr-k8s-c6crs\" (UID: \"130f6554-c476-49a0-8270-593008c1ffc6\") " pod="metallb-system/frr-k8s-c6crs" Dec 11 08:24:56 crc kubenswrapper[4860]: I1211 08:24:56.619084 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/130f6554-c476-49a0-8270-593008c1ffc6-frr-conf\") pod \"frr-k8s-c6crs\" (UID: \"130f6554-c476-49a0-8270-593008c1ffc6\") " pod="metallb-system/frr-k8s-c6crs" Dec 11 08:24:56 crc kubenswrapper[4860]: I1211 08:24:56.619444 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/130f6554-c476-49a0-8270-593008c1ffc6-frr-startup\") pod \"frr-k8s-c6crs\" (UID: \"130f6554-c476-49a0-8270-593008c1ffc6\") " pod="metallb-system/frr-k8s-c6crs" Dec 11 08:24:56 crc kubenswrapper[4860]: E1211 08:24:56.619543 4860 secret.go:188] Couldn't get secret metallb-system/frr-k8s-certs-secret: secret "frr-k8s-certs-secret" not found Dec 11 08:24:56 crc kubenswrapper[4860]: E1211 08:24:56.619598 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/130f6554-c476-49a0-8270-593008c1ffc6-metrics-certs podName:130f6554-c476-49a0-8270-593008c1ffc6 nodeName:}" failed. No retries permitted until 2025-12-11 08:24:57.119583823 +0000 UTC m=+829.848102878 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/130f6554-c476-49a0-8270-593008c1ffc6-metrics-certs") pod "frr-k8s-c6crs" (UID: "130f6554-c476-49a0-8270-593008c1ffc6") : secret "frr-k8s-certs-secret" not found Dec 11 08:24:56 crc kubenswrapper[4860]: E1211 08:24:56.629714 4860 secret.go:188] Couldn't get secret metallb-system/frr-k8s-webhook-server-cert: secret "frr-k8s-webhook-server-cert" not found Dec 11 08:24:56 crc kubenswrapper[4860]: E1211 08:24:56.629848 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/98afb3e0-d363-472c-886e-2b2981cb13a4-cert podName:98afb3e0-d363-472c-886e-2b2981cb13a4 nodeName:}" failed. No retries permitted until 2025-12-11 08:24:57.129824026 +0000 UTC m=+829.858343081 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/98afb3e0-d363-472c-886e-2b2981cb13a4-cert") pod "frr-k8s-webhook-server-7784b6fcf-2548h" (UID: "98afb3e0-d363-472c-886e-2b2981cb13a4") : secret "frr-k8s-webhook-server-cert" not found Dec 11 08:24:56 crc kubenswrapper[4860]: I1211 08:24:56.635382 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-qmw55"] Dec 11 08:24:56 crc kubenswrapper[4860]: I1211 08:24:56.636629 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-qmw55" Dec 11 08:24:56 crc kubenswrapper[4860]: I1211 08:24:56.642836 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Dec 11 08:24:56 crc kubenswrapper[4860]: I1211 08:24:56.643093 4860 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Dec 11 08:24:56 crc kubenswrapper[4860]: I1211 08:24:56.643217 4860 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-zhfst" Dec 11 08:24:56 crc kubenswrapper[4860]: I1211 08:24:56.643333 4860 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Dec 11 08:24:56 crc kubenswrapper[4860]: I1211 08:24:56.659384 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-5bddd4b946-96p7v"] Dec 11 08:24:56 crc kubenswrapper[4860]: I1211 08:24:56.660916 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-5bddd4b946-96p7v" Dec 11 08:24:56 crc kubenswrapper[4860]: I1211 08:24:56.663317 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-78mrl\" (UniqueName: \"kubernetes.io/projected/130f6554-c476-49a0-8270-593008c1ffc6-kube-api-access-78mrl\") pod \"frr-k8s-c6crs\" (UID: \"130f6554-c476-49a0-8270-593008c1ffc6\") " pod="metallb-system/frr-k8s-c6crs" Dec 11 08:24:56 crc kubenswrapper[4860]: I1211 08:24:56.663851 4860 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Dec 11 08:24:56 crc kubenswrapper[4860]: I1211 08:24:56.664861 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l84rk\" (UniqueName: \"kubernetes.io/projected/98afb3e0-d363-472c-886e-2b2981cb13a4-kube-api-access-l84rk\") pod \"frr-k8s-webhook-server-7784b6fcf-2548h\" (UID: \"98afb3e0-d363-472c-886e-2b2981cb13a4\") " pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-2548h" Dec 11 08:24:56 crc kubenswrapper[4860]: I1211 08:24:56.683984 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-5bddd4b946-96p7v"] Dec 11 08:24:56 crc kubenswrapper[4860]: I1211 08:24:56.718656 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/b5eafccf-57d6-4bcb-a43f-5ef3a2f04db0-metallb-excludel2\") pod \"speaker-qmw55\" (UID: \"b5eafccf-57d6-4bcb-a43f-5ef3a2f04db0\") " pod="metallb-system/speaker-qmw55" Dec 11 08:24:56 crc kubenswrapper[4860]: I1211 08:24:56.718718 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/b5eafccf-57d6-4bcb-a43f-5ef3a2f04db0-memberlist\") pod \"speaker-qmw55\" (UID: \"b5eafccf-57d6-4bcb-a43f-5ef3a2f04db0\") " pod="metallb-system/speaker-qmw55" Dec 11 08:24:56 crc kubenswrapper[4860]: I1211 08:24:56.718758 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b5eafccf-57d6-4bcb-a43f-5ef3a2f04db0-metrics-certs\") pod \"speaker-qmw55\" (UID: \"b5eafccf-57d6-4bcb-a43f-5ef3a2f04db0\") " pod="metallb-system/speaker-qmw55" Dec 11 08:24:56 crc kubenswrapper[4860]: I1211 08:24:56.718786 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d5t8r\" (UniqueName: \"kubernetes.io/projected/b5eafccf-57d6-4bcb-a43f-5ef3a2f04db0-kube-api-access-d5t8r\") pod \"speaker-qmw55\" (UID: \"b5eafccf-57d6-4bcb-a43f-5ef3a2f04db0\") " pod="metallb-system/speaker-qmw55" Dec 11 08:24:56 crc kubenswrapper[4860]: I1211 08:24:56.718806 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/31da3aab-3e08-4b87-a866-d49a4224cf29-metrics-certs\") pod \"controller-5bddd4b946-96p7v\" (UID: \"31da3aab-3e08-4b87-a866-d49a4224cf29\") " pod="metallb-system/controller-5bddd4b946-96p7v" Dec 11 08:24:56 crc kubenswrapper[4860]: I1211 08:24:56.718830 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m9t7c\" (UniqueName: \"kubernetes.io/projected/31da3aab-3e08-4b87-a866-d49a4224cf29-kube-api-access-m9t7c\") pod \"controller-5bddd4b946-96p7v\" (UID: \"31da3aab-3e08-4b87-a866-d49a4224cf29\") " pod="metallb-system/controller-5bddd4b946-96p7v" Dec 11 08:24:56 crc kubenswrapper[4860]: I1211 08:24:56.718865 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/31da3aab-3e08-4b87-a866-d49a4224cf29-cert\") pod \"controller-5bddd4b946-96p7v\" (UID: \"31da3aab-3e08-4b87-a866-d49a4224cf29\") " pod="metallb-system/controller-5bddd4b946-96p7v" Dec 11 08:24:56 crc kubenswrapper[4860]: I1211 08:24:56.819804 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/b5eafccf-57d6-4bcb-a43f-5ef3a2f04db0-metallb-excludel2\") pod \"speaker-qmw55\" (UID: \"b5eafccf-57d6-4bcb-a43f-5ef3a2f04db0\") " pod="metallb-system/speaker-qmw55" Dec 11 08:24:56 crc kubenswrapper[4860]: I1211 08:24:56.820242 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/b5eafccf-57d6-4bcb-a43f-5ef3a2f04db0-memberlist\") pod \"speaker-qmw55\" (UID: \"b5eafccf-57d6-4bcb-a43f-5ef3a2f04db0\") " pod="metallb-system/speaker-qmw55" Dec 11 08:24:56 crc kubenswrapper[4860]: I1211 08:24:56.820378 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b5eafccf-57d6-4bcb-a43f-5ef3a2f04db0-metrics-certs\") pod \"speaker-qmw55\" (UID: \"b5eafccf-57d6-4bcb-a43f-5ef3a2f04db0\") " pod="metallb-system/speaker-qmw55" Dec 11 08:24:56 crc kubenswrapper[4860]: I1211 08:24:56.820476 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d5t8r\" (UniqueName: \"kubernetes.io/projected/b5eafccf-57d6-4bcb-a43f-5ef3a2f04db0-kube-api-access-d5t8r\") pod \"speaker-qmw55\" (UID: \"b5eafccf-57d6-4bcb-a43f-5ef3a2f04db0\") " pod="metallb-system/speaker-qmw55" Dec 11 08:24:56 crc kubenswrapper[4860]: I1211 08:24:56.820569 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/31da3aab-3e08-4b87-a866-d49a4224cf29-metrics-certs\") pod \"controller-5bddd4b946-96p7v\" (UID: \"31da3aab-3e08-4b87-a866-d49a4224cf29\") " pod="metallb-system/controller-5bddd4b946-96p7v" Dec 11 08:24:56 crc kubenswrapper[4860]: I1211 08:24:56.820780 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m9t7c\" (UniqueName: \"kubernetes.io/projected/31da3aab-3e08-4b87-a866-d49a4224cf29-kube-api-access-m9t7c\") pod \"controller-5bddd4b946-96p7v\" (UID: \"31da3aab-3e08-4b87-a866-d49a4224cf29\") " pod="metallb-system/controller-5bddd4b946-96p7v" Dec 11 08:24:56 crc kubenswrapper[4860]: I1211 08:24:56.820918 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/31da3aab-3e08-4b87-a866-d49a4224cf29-cert\") pod \"controller-5bddd4b946-96p7v\" (UID: \"31da3aab-3e08-4b87-a866-d49a4224cf29\") " pod="metallb-system/controller-5bddd4b946-96p7v" Dec 11 08:24:56 crc kubenswrapper[4860]: E1211 08:24:56.820414 4860 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 11 08:24:56 crc kubenswrapper[4860]: E1211 08:24:56.821222 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b5eafccf-57d6-4bcb-a43f-5ef3a2f04db0-memberlist podName:b5eafccf-57d6-4bcb-a43f-5ef3a2f04db0 nodeName:}" failed. No retries permitted until 2025-12-11 08:24:57.321205908 +0000 UTC m=+830.049724963 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/b5eafccf-57d6-4bcb-a43f-5ef3a2f04db0-memberlist") pod "speaker-qmw55" (UID: "b5eafccf-57d6-4bcb-a43f-5ef3a2f04db0") : secret "metallb-memberlist" not found Dec 11 08:24:56 crc kubenswrapper[4860]: E1211 08:24:56.820454 4860 secret.go:188] Couldn't get secret metallb-system/speaker-certs-secret: secret "speaker-certs-secret" not found Dec 11 08:24:56 crc kubenswrapper[4860]: I1211 08:24:56.820806 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/b5eafccf-57d6-4bcb-a43f-5ef3a2f04db0-metallb-excludel2\") pod \"speaker-qmw55\" (UID: \"b5eafccf-57d6-4bcb-a43f-5ef3a2f04db0\") " pod="metallb-system/speaker-qmw55" Dec 11 08:24:56 crc kubenswrapper[4860]: E1211 08:24:56.821436 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b5eafccf-57d6-4bcb-a43f-5ef3a2f04db0-metrics-certs podName:b5eafccf-57d6-4bcb-a43f-5ef3a2f04db0 nodeName:}" failed. No retries permitted until 2025-12-11 08:24:57.321417765 +0000 UTC m=+830.049936820 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/b5eafccf-57d6-4bcb-a43f-5ef3a2f04db0-metrics-certs") pod "speaker-qmw55" (UID: "b5eafccf-57d6-4bcb-a43f-5ef3a2f04db0") : secret "speaker-certs-secret" not found Dec 11 08:24:56 crc kubenswrapper[4860]: E1211 08:24:56.820733 4860 secret.go:188] Couldn't get secret metallb-system/controller-certs-secret: secret "controller-certs-secret" not found Dec 11 08:24:56 crc kubenswrapper[4860]: E1211 08:24:56.821629 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/31da3aab-3e08-4b87-a866-d49a4224cf29-metrics-certs podName:31da3aab-3e08-4b87-a866-d49a4224cf29 nodeName:}" failed. No retries permitted until 2025-12-11 08:24:57.32159514 +0000 UTC m=+830.050114195 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/31da3aab-3e08-4b87-a866-d49a4224cf29-metrics-certs") pod "controller-5bddd4b946-96p7v" (UID: "31da3aab-3e08-4b87-a866-d49a4224cf29") : secret "controller-certs-secret" not found Dec 11 08:24:56 crc kubenswrapper[4860]: I1211 08:24:56.825107 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/31da3aab-3e08-4b87-a866-d49a4224cf29-cert\") pod \"controller-5bddd4b946-96p7v\" (UID: \"31da3aab-3e08-4b87-a866-d49a4224cf29\") " pod="metallb-system/controller-5bddd4b946-96p7v" Dec 11 08:24:56 crc kubenswrapper[4860]: I1211 08:24:56.855856 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d5t8r\" (UniqueName: \"kubernetes.io/projected/b5eafccf-57d6-4bcb-a43f-5ef3a2f04db0-kube-api-access-d5t8r\") pod \"speaker-qmw55\" (UID: \"b5eafccf-57d6-4bcb-a43f-5ef3a2f04db0\") " pod="metallb-system/speaker-qmw55" Dec 11 08:24:56 crc kubenswrapper[4860]: I1211 08:24:56.862355 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m9t7c\" (UniqueName: \"kubernetes.io/projected/31da3aab-3e08-4b87-a866-d49a4224cf29-kube-api-access-m9t7c\") pod \"controller-5bddd4b946-96p7v\" (UID: \"31da3aab-3e08-4b87-a866-d49a4224cf29\") " pod="metallb-system/controller-5bddd4b946-96p7v" Dec 11 08:24:57 crc kubenswrapper[4860]: I1211 08:24:57.125051 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/130f6554-c476-49a0-8270-593008c1ffc6-metrics-certs\") pod \"frr-k8s-c6crs\" (UID: \"130f6554-c476-49a0-8270-593008c1ffc6\") " pod="metallb-system/frr-k8s-c6crs" Dec 11 08:24:57 crc kubenswrapper[4860]: I1211 08:24:57.129849 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/130f6554-c476-49a0-8270-593008c1ffc6-metrics-certs\") pod \"frr-k8s-c6crs\" (UID: \"130f6554-c476-49a0-8270-593008c1ffc6\") " pod="metallb-system/frr-k8s-c6crs" Dec 11 08:24:57 crc kubenswrapper[4860]: I1211 08:24:57.132557 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-c6crs" Dec 11 08:24:57 crc kubenswrapper[4860]: I1211 08:24:57.227419 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/98afb3e0-d363-472c-886e-2b2981cb13a4-cert\") pod \"frr-k8s-webhook-server-7784b6fcf-2548h\" (UID: \"98afb3e0-d363-472c-886e-2b2981cb13a4\") " pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-2548h" Dec 11 08:24:57 crc kubenswrapper[4860]: I1211 08:24:57.232556 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/98afb3e0-d363-472c-886e-2b2981cb13a4-cert\") pod \"frr-k8s-webhook-server-7784b6fcf-2548h\" (UID: \"98afb3e0-d363-472c-886e-2b2981cb13a4\") " pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-2548h" Dec 11 08:24:57 crc kubenswrapper[4860]: I1211 08:24:57.329115 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/31da3aab-3e08-4b87-a866-d49a4224cf29-metrics-certs\") pod \"controller-5bddd4b946-96p7v\" (UID: \"31da3aab-3e08-4b87-a866-d49a4224cf29\") " pod="metallb-system/controller-5bddd4b946-96p7v" Dec 11 08:24:57 crc kubenswrapper[4860]: I1211 08:24:57.329262 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/b5eafccf-57d6-4bcb-a43f-5ef3a2f04db0-memberlist\") pod \"speaker-qmw55\" (UID: \"b5eafccf-57d6-4bcb-a43f-5ef3a2f04db0\") " pod="metallb-system/speaker-qmw55" Dec 11 08:24:57 crc kubenswrapper[4860]: I1211 08:24:57.329302 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b5eafccf-57d6-4bcb-a43f-5ef3a2f04db0-metrics-certs\") pod \"speaker-qmw55\" (UID: \"b5eafccf-57d6-4bcb-a43f-5ef3a2f04db0\") " pod="metallb-system/speaker-qmw55" Dec 11 08:24:57 crc kubenswrapper[4860]: E1211 08:24:57.329861 4860 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 11 08:24:57 crc kubenswrapper[4860]: E1211 08:24:57.329951 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b5eafccf-57d6-4bcb-a43f-5ef3a2f04db0-memberlist podName:b5eafccf-57d6-4bcb-a43f-5ef3a2f04db0 nodeName:}" failed. No retries permitted until 2025-12-11 08:24:58.329926127 +0000 UTC m=+831.058445182 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/b5eafccf-57d6-4bcb-a43f-5ef3a2f04db0-memberlist") pod "speaker-qmw55" (UID: "b5eafccf-57d6-4bcb-a43f-5ef3a2f04db0") : secret "metallb-memberlist" not found Dec 11 08:24:57 crc kubenswrapper[4860]: I1211 08:24:57.333066 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b5eafccf-57d6-4bcb-a43f-5ef3a2f04db0-metrics-certs\") pod \"speaker-qmw55\" (UID: \"b5eafccf-57d6-4bcb-a43f-5ef3a2f04db0\") " pod="metallb-system/speaker-qmw55" Dec 11 08:24:57 crc kubenswrapper[4860]: I1211 08:24:57.333145 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/31da3aab-3e08-4b87-a866-d49a4224cf29-metrics-certs\") pod \"controller-5bddd4b946-96p7v\" (UID: \"31da3aab-3e08-4b87-a866-d49a4224cf29\") " pod="metallb-system/controller-5bddd4b946-96p7v" Dec 11 08:24:57 crc kubenswrapper[4860]: I1211 08:24:57.459875 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-2548h" Dec 11 08:24:57 crc kubenswrapper[4860]: I1211 08:24:57.618619 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-5bddd4b946-96p7v" Dec 11 08:24:57 crc kubenswrapper[4860]: I1211 08:24:57.628278 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-c6crs" event={"ID":"130f6554-c476-49a0-8270-593008c1ffc6","Type":"ContainerStarted","Data":"4cd12167a026b0b52e1e449d9df8dacefc13c6cfbd611bf2ae7fd87181bda8ea"} Dec 11 08:24:57 crc kubenswrapper[4860]: I1211 08:24:57.834323 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-5bddd4b946-96p7v"] Dec 11 08:24:57 crc kubenswrapper[4860]: W1211 08:24:57.843626 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod31da3aab_3e08_4b87_a866_d49a4224cf29.slice/crio-baacccbedca164e0f63ae3c02e2ee033bf0070ff15c7a99b951c1c5554b4bf86 WatchSource:0}: Error finding container baacccbedca164e0f63ae3c02e2ee033bf0070ff15c7a99b951c1c5554b4bf86: Status 404 returned error can't find the container with id baacccbedca164e0f63ae3c02e2ee033bf0070ff15c7a99b951c1c5554b4bf86 Dec 11 08:24:57 crc kubenswrapper[4860]: I1211 08:24:57.927123 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7784b6fcf-2548h"] Dec 11 08:24:58 crc kubenswrapper[4860]: I1211 08:24:58.343811 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/b5eafccf-57d6-4bcb-a43f-5ef3a2f04db0-memberlist\") pod \"speaker-qmw55\" (UID: \"b5eafccf-57d6-4bcb-a43f-5ef3a2f04db0\") " pod="metallb-system/speaker-qmw55" Dec 11 08:24:58 crc kubenswrapper[4860]: I1211 08:24:58.353733 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/b5eafccf-57d6-4bcb-a43f-5ef3a2f04db0-memberlist\") pod \"speaker-qmw55\" (UID: \"b5eafccf-57d6-4bcb-a43f-5ef3a2f04db0\") " pod="metallb-system/speaker-qmw55" Dec 11 08:24:58 crc kubenswrapper[4860]: I1211 08:24:58.506968 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-qmw55" Dec 11 08:24:58 crc kubenswrapper[4860]: W1211 08:24:58.530622 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb5eafccf_57d6_4bcb_a43f_5ef3a2f04db0.slice/crio-3300206c19702538354a98ccabd71d11198f4f5a3efb2a1798ef1579e4ff379b WatchSource:0}: Error finding container 3300206c19702538354a98ccabd71d11198f4f5a3efb2a1798ef1579e4ff379b: Status 404 returned error can't find the container with id 3300206c19702538354a98ccabd71d11198f4f5a3efb2a1798ef1579e4ff379b Dec 11 08:24:58 crc kubenswrapper[4860]: I1211 08:24:58.635399 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-2548h" event={"ID":"98afb3e0-d363-472c-886e-2b2981cb13a4","Type":"ContainerStarted","Data":"f5f93a9e3392ba6fb2d86b4485e1309e8ece96e285e3630a1af1e5dbef379974"} Dec 11 08:24:58 crc kubenswrapper[4860]: I1211 08:24:58.637293 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-qmw55" event={"ID":"b5eafccf-57d6-4bcb-a43f-5ef3a2f04db0","Type":"ContainerStarted","Data":"3300206c19702538354a98ccabd71d11198f4f5a3efb2a1798ef1579e4ff379b"} Dec 11 08:24:58 crc kubenswrapper[4860]: I1211 08:24:58.639826 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-5bddd4b946-96p7v" event={"ID":"31da3aab-3e08-4b87-a866-d49a4224cf29","Type":"ContainerStarted","Data":"607a39ba7241a92a39965d90d036ec4267bd7379c323b4d468d0a3cdc352dea5"} Dec 11 08:24:58 crc kubenswrapper[4860]: I1211 08:24:58.639864 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-5bddd4b946-96p7v" event={"ID":"31da3aab-3e08-4b87-a866-d49a4224cf29","Type":"ContainerStarted","Data":"1585812cf5907d51c236bb6851f97dc34e3608d10e1d41fe8fbd75b01259ad73"} Dec 11 08:24:58 crc kubenswrapper[4860]: I1211 08:24:58.639880 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-5bddd4b946-96p7v" event={"ID":"31da3aab-3e08-4b87-a866-d49a4224cf29","Type":"ContainerStarted","Data":"baacccbedca164e0f63ae3c02e2ee033bf0070ff15c7a99b951c1c5554b4bf86"} Dec 11 08:24:58 crc kubenswrapper[4860]: I1211 08:24:58.641358 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-5bddd4b946-96p7v" Dec 11 08:24:58 crc kubenswrapper[4860]: I1211 08:24:58.666787 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-5bddd4b946-96p7v" podStartSLOduration=2.66676144 podStartE2EDuration="2.66676144s" podCreationTimestamp="2025-12-11 08:24:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:24:58.662471505 +0000 UTC m=+831.390990560" watchObservedRunningTime="2025-12-11 08:24:58.66676144 +0000 UTC m=+831.395280495" Dec 11 08:24:59 crc kubenswrapper[4860]: I1211 08:24:59.665660 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-qmw55" event={"ID":"b5eafccf-57d6-4bcb-a43f-5ef3a2f04db0","Type":"ContainerStarted","Data":"e9e5cbeaf4ec4c6740712398fb37fdd6214983221d79b339000e51a8a0e618af"} Dec 11 08:24:59 crc kubenswrapper[4860]: I1211 08:24:59.668802 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-qmw55" event={"ID":"b5eafccf-57d6-4bcb-a43f-5ef3a2f04db0","Type":"ContainerStarted","Data":"2222a35821df82fca47992c642b1b344bccc53bda6b7f5c4afaff5e7bb11e3cc"} Dec 11 08:24:59 crc kubenswrapper[4860]: I1211 08:24:59.704600 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-qmw55" podStartSLOduration=3.704569013 podStartE2EDuration="3.704569013s" podCreationTimestamp="2025-12-11 08:24:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:24:59.700018199 +0000 UTC m=+832.428537254" watchObservedRunningTime="2025-12-11 08:24:59.704569013 +0000 UTC m=+832.433088068" Dec 11 08:25:00 crc kubenswrapper[4860]: I1211 08:25:00.673943 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-qmw55" Dec 11 08:25:04 crc kubenswrapper[4860]: I1211 08:25:04.705189 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-2548h" event={"ID":"98afb3e0-d363-472c-886e-2b2981cb13a4","Type":"ContainerStarted","Data":"e1a0622c560bcd0649c89b3d218d320baa898bf63a36b494ebb985b8e54454a5"} Dec 11 08:25:04 crc kubenswrapper[4860]: I1211 08:25:04.706860 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-2548h" Dec 11 08:25:04 crc kubenswrapper[4860]: I1211 08:25:04.711550 4860 generic.go:334] "Generic (PLEG): container finished" podID="130f6554-c476-49a0-8270-593008c1ffc6" containerID="f9de791e03c179832aa73b73421bdf0487fab22042b5c12006f4d6389089be06" exitCode=0 Dec 11 08:25:04 crc kubenswrapper[4860]: I1211 08:25:04.711670 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-c6crs" event={"ID":"130f6554-c476-49a0-8270-593008c1ffc6","Type":"ContainerDied","Data":"f9de791e03c179832aa73b73421bdf0487fab22042b5c12006f4d6389089be06"} Dec 11 08:25:04 crc kubenswrapper[4860]: I1211 08:25:04.757070 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-2548h" podStartSLOduration=2.310615456 podStartE2EDuration="8.757043355s" podCreationTimestamp="2025-12-11 08:24:56 +0000 UTC" firstStartedPulling="2025-12-11 08:24:57.990010365 +0000 UTC m=+830.718529420" lastFinishedPulling="2025-12-11 08:25:04.436438264 +0000 UTC m=+837.164957319" observedRunningTime="2025-12-11 08:25:04.72519497 +0000 UTC m=+837.453714025" watchObservedRunningTime="2025-12-11 08:25:04.757043355 +0000 UTC m=+837.485562410" Dec 11 08:25:05 crc kubenswrapper[4860]: I1211 08:25:05.721938 4860 generic.go:334] "Generic (PLEG): container finished" podID="130f6554-c476-49a0-8270-593008c1ffc6" containerID="8708260680bd97f976f2a409a1a1a31cdbaa344711fb1095b1dad8fee0382bcf" exitCode=0 Dec 11 08:25:05 crc kubenswrapper[4860]: I1211 08:25:05.722083 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-c6crs" event={"ID":"130f6554-c476-49a0-8270-593008c1ffc6","Type":"ContainerDied","Data":"8708260680bd97f976f2a409a1a1a31cdbaa344711fb1095b1dad8fee0382bcf"} Dec 11 08:25:06 crc kubenswrapper[4860]: I1211 08:25:06.733409 4860 generic.go:334] "Generic (PLEG): container finished" podID="130f6554-c476-49a0-8270-593008c1ffc6" containerID="d65017f1bdedaa7ca5760a1f1e0d11d35c5a80ee1b94693f48d9800f77364f09" exitCode=0 Dec 11 08:25:06 crc kubenswrapper[4860]: I1211 08:25:06.733688 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-c6crs" event={"ID":"130f6554-c476-49a0-8270-593008c1ffc6","Type":"ContainerDied","Data":"d65017f1bdedaa7ca5760a1f1e0d11d35c5a80ee1b94693f48d9800f77364f09"} Dec 11 08:25:07 crc kubenswrapper[4860]: I1211 08:25:07.746186 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-c6crs" event={"ID":"130f6554-c476-49a0-8270-593008c1ffc6","Type":"ContainerStarted","Data":"44c3f0b21105a79f743be6077b851029958583a03b16513252884a92376de1f1"} Dec 11 08:25:07 crc kubenswrapper[4860]: I1211 08:25:07.747409 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-c6crs" event={"ID":"130f6554-c476-49a0-8270-593008c1ffc6","Type":"ContainerStarted","Data":"15deac9cd425353f887b99ebe222a95ba4eac055d55b26f2ac45d13edf5bb245"} Dec 11 08:25:07 crc kubenswrapper[4860]: I1211 08:25:07.747497 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-c6crs" event={"ID":"130f6554-c476-49a0-8270-593008c1ffc6","Type":"ContainerStarted","Data":"3d0c76da5da3ba7aa333bfe34f2fe3e88b4789d4b21e8157c9415b350796f9da"} Dec 11 08:25:07 crc kubenswrapper[4860]: I1211 08:25:07.747586 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-c6crs" event={"ID":"130f6554-c476-49a0-8270-593008c1ffc6","Type":"ContainerStarted","Data":"3128d1e4a9bbcd79aac707064346aa26b28399194497d8ae4554d33ae9db1a51"} Dec 11 08:25:08 crc kubenswrapper[4860]: I1211 08:25:08.512548 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-qmw55" Dec 11 08:25:08 crc kubenswrapper[4860]: I1211 08:25:08.760082 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-c6crs" event={"ID":"130f6554-c476-49a0-8270-593008c1ffc6","Type":"ContainerStarted","Data":"811454d281f43e72ff0f45855481b16f5a2836692ac5c7212a631e65ab0e685c"} Dec 11 08:25:08 crc kubenswrapper[4860]: I1211 08:25:08.760147 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-c6crs" event={"ID":"130f6554-c476-49a0-8270-593008c1ffc6","Type":"ContainerStarted","Data":"3a16318f34f120ca5e456f2b30aee33a7697cae1b6f187fc0e747760760b517c"} Dec 11 08:25:08 crc kubenswrapper[4860]: I1211 08:25:08.761463 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-c6crs" Dec 11 08:25:08 crc kubenswrapper[4860]: I1211 08:25:08.791718 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-c6crs" podStartSLOduration=5.942910684 podStartE2EDuration="12.791689675s" podCreationTimestamp="2025-12-11 08:24:56 +0000 UTC" firstStartedPulling="2025-12-11 08:24:57.570231093 +0000 UTC m=+830.298750148" lastFinishedPulling="2025-12-11 08:25:04.419010084 +0000 UTC m=+837.147529139" observedRunningTime="2025-12-11 08:25:08.78837015 +0000 UTC m=+841.516889215" watchObservedRunningTime="2025-12-11 08:25:08.791689675 +0000 UTC m=+841.520208750" Dec 11 08:25:08 crc kubenswrapper[4860]: I1211 08:25:08.795548 4860 patch_prober.go:28] interesting pod/machine-config-daemon-99qgp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 08:25:08 crc kubenswrapper[4860]: I1211 08:25:08.795631 4860 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 08:25:08 crc kubenswrapper[4860]: I1211 08:25:08.795706 4860 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" Dec 11 08:25:08 crc kubenswrapper[4860]: I1211 08:25:08.796568 4860 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7abbc21a5acc8c61b19323dc53f4bb2e28d39cd9e4964ba8a2b38c1bd09ed9bf"} pod="openshift-machine-config-operator/machine-config-daemon-99qgp" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 11 08:25:08 crc kubenswrapper[4860]: I1211 08:25:08.796632 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" containerName="machine-config-daemon" containerID="cri-o://7abbc21a5acc8c61b19323dc53f4bb2e28d39cd9e4964ba8a2b38c1bd09ed9bf" gracePeriod=600 Dec 11 08:25:09 crc kubenswrapper[4860]: I1211 08:25:09.770005 4860 generic.go:334] "Generic (PLEG): container finished" podID="31c30642-6e60-41b4-a477-0d802424e0aa" containerID="7abbc21a5acc8c61b19323dc53f4bb2e28d39cd9e4964ba8a2b38c1bd09ed9bf" exitCode=0 Dec 11 08:25:09 crc kubenswrapper[4860]: I1211 08:25:09.770090 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" event={"ID":"31c30642-6e60-41b4-a477-0d802424e0aa","Type":"ContainerDied","Data":"7abbc21a5acc8c61b19323dc53f4bb2e28d39cd9e4964ba8a2b38c1bd09ed9bf"} Dec 11 08:25:09 crc kubenswrapper[4860]: I1211 08:25:09.771055 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" event={"ID":"31c30642-6e60-41b4-a477-0d802424e0aa","Type":"ContainerStarted","Data":"565fbddca3ff6a011767df936f9e699e4560197af6e486d467234b4599b2d2f6"} Dec 11 08:25:09 crc kubenswrapper[4860]: I1211 08:25:09.771084 4860 scope.go:117] "RemoveContainer" containerID="ff37f2c081c95677939e414d6be18406d4d67feea33292ff469d545943df00a5" Dec 11 08:25:11 crc kubenswrapper[4860]: I1211 08:25:11.323057 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-gblg8"] Dec 11 08:25:11 crc kubenswrapper[4860]: I1211 08:25:11.324958 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-gblg8" Dec 11 08:25:11 crc kubenswrapper[4860]: I1211 08:25:11.332003 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-2hzb2" Dec 11 08:25:11 crc kubenswrapper[4860]: I1211 08:25:11.333170 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Dec 11 08:25:11 crc kubenswrapper[4860]: I1211 08:25:11.333836 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Dec 11 08:25:11 crc kubenswrapper[4860]: I1211 08:25:11.336960 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-gblg8"] Dec 11 08:25:11 crc kubenswrapper[4860]: I1211 08:25:11.462314 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4wlfr\" (UniqueName: \"kubernetes.io/projected/bae66ca8-9554-4c05-adb3-e73819e23c1b-kube-api-access-4wlfr\") pod \"openstack-operator-index-gblg8\" (UID: \"bae66ca8-9554-4c05-adb3-e73819e23c1b\") " pod="openstack-operators/openstack-operator-index-gblg8" Dec 11 08:25:11 crc kubenswrapper[4860]: I1211 08:25:11.564333 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4wlfr\" (UniqueName: \"kubernetes.io/projected/bae66ca8-9554-4c05-adb3-e73819e23c1b-kube-api-access-4wlfr\") pod \"openstack-operator-index-gblg8\" (UID: \"bae66ca8-9554-4c05-adb3-e73819e23c1b\") " pod="openstack-operators/openstack-operator-index-gblg8" Dec 11 08:25:11 crc kubenswrapper[4860]: I1211 08:25:11.590117 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4wlfr\" (UniqueName: \"kubernetes.io/projected/bae66ca8-9554-4c05-adb3-e73819e23c1b-kube-api-access-4wlfr\") pod \"openstack-operator-index-gblg8\" (UID: \"bae66ca8-9554-4c05-adb3-e73819e23c1b\") " pod="openstack-operators/openstack-operator-index-gblg8" Dec 11 08:25:11 crc kubenswrapper[4860]: I1211 08:25:11.644431 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-gblg8" Dec 11 08:25:12 crc kubenswrapper[4860]: I1211 08:25:12.109259 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-gblg8"] Dec 11 08:25:12 crc kubenswrapper[4860]: W1211 08:25:12.118666 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbae66ca8_9554_4c05_adb3_e73819e23c1b.slice/crio-f5ce7d6fe252053d60410fdde592e1d640f1622fc3b2e75b6b98d3aa7998cc35 WatchSource:0}: Error finding container f5ce7d6fe252053d60410fdde592e1d640f1622fc3b2e75b6b98d3aa7998cc35: Status 404 returned error can't find the container with id f5ce7d6fe252053d60410fdde592e1d640f1622fc3b2e75b6b98d3aa7998cc35 Dec 11 08:25:12 crc kubenswrapper[4860]: I1211 08:25:12.133446 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-c6crs" Dec 11 08:25:12 crc kubenswrapper[4860]: I1211 08:25:12.185452 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-c6crs" Dec 11 08:25:12 crc kubenswrapper[4860]: I1211 08:25:12.808773 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-gblg8" event={"ID":"bae66ca8-9554-4c05-adb3-e73819e23c1b","Type":"ContainerStarted","Data":"f5ce7d6fe252053d60410fdde592e1d640f1622fc3b2e75b6b98d3aa7998cc35"} Dec 11 08:25:14 crc kubenswrapper[4860]: I1211 08:25:14.501982 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-gblg8"] Dec 11 08:25:15 crc kubenswrapper[4860]: I1211 08:25:15.117036 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-hm8hr"] Dec 11 08:25:15 crc kubenswrapper[4860]: I1211 08:25:15.118300 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-hm8hr" Dec 11 08:25:15 crc kubenswrapper[4860]: I1211 08:25:15.124577 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-hm8hr"] Dec 11 08:25:15 crc kubenswrapper[4860]: I1211 08:25:15.226034 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zbl6r\" (UniqueName: \"kubernetes.io/projected/aec175c6-0b2d-40d8-855a-6132fb5e66a0-kube-api-access-zbl6r\") pod \"openstack-operator-index-hm8hr\" (UID: \"aec175c6-0b2d-40d8-855a-6132fb5e66a0\") " pod="openstack-operators/openstack-operator-index-hm8hr" Dec 11 08:25:15 crc kubenswrapper[4860]: I1211 08:25:15.328009 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zbl6r\" (UniqueName: \"kubernetes.io/projected/aec175c6-0b2d-40d8-855a-6132fb5e66a0-kube-api-access-zbl6r\") pod \"openstack-operator-index-hm8hr\" (UID: \"aec175c6-0b2d-40d8-855a-6132fb5e66a0\") " pod="openstack-operators/openstack-operator-index-hm8hr" Dec 11 08:25:15 crc kubenswrapper[4860]: I1211 08:25:15.351004 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zbl6r\" (UniqueName: \"kubernetes.io/projected/aec175c6-0b2d-40d8-855a-6132fb5e66a0-kube-api-access-zbl6r\") pod \"openstack-operator-index-hm8hr\" (UID: \"aec175c6-0b2d-40d8-855a-6132fb5e66a0\") " pod="openstack-operators/openstack-operator-index-hm8hr" Dec 11 08:25:15 crc kubenswrapper[4860]: I1211 08:25:15.477804 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-hm8hr" Dec 11 08:25:15 crc kubenswrapper[4860]: I1211 08:25:15.831256 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-gblg8" event={"ID":"bae66ca8-9554-4c05-adb3-e73819e23c1b","Type":"ContainerStarted","Data":"d99d69a38ff6989763758da2d2bd3b736813980c1dc832da5c8b39ddbd7145c1"} Dec 11 08:25:15 crc kubenswrapper[4860]: I1211 08:25:15.831498 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-gblg8" podUID="bae66ca8-9554-4c05-adb3-e73819e23c1b" containerName="registry-server" containerID="cri-o://d99d69a38ff6989763758da2d2bd3b736813980c1dc832da5c8b39ddbd7145c1" gracePeriod=2 Dec 11 08:25:15 crc kubenswrapper[4860]: I1211 08:25:15.847861 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-gblg8" podStartSLOduration=1.432740395 podStartE2EDuration="4.847829653s" podCreationTimestamp="2025-12-11 08:25:11 +0000 UTC" firstStartedPulling="2025-12-11 08:25:12.122262598 +0000 UTC m=+844.850781663" lastFinishedPulling="2025-12-11 08:25:15.537351866 +0000 UTC m=+848.265870921" observedRunningTime="2025-12-11 08:25:15.847248304 +0000 UTC m=+848.575767359" watchObservedRunningTime="2025-12-11 08:25:15.847829653 +0000 UTC m=+848.576348708" Dec 11 08:25:15 crc kubenswrapper[4860]: I1211 08:25:15.937775 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-hm8hr"] Dec 11 08:25:15 crc kubenswrapper[4860]: W1211 08:25:15.954659 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaec175c6_0b2d_40d8_855a_6132fb5e66a0.slice/crio-a00ac3d1d66af99a414f5ec847d7c9229b50b7357918860bf03698794825b142 WatchSource:0}: Error finding container a00ac3d1d66af99a414f5ec847d7c9229b50b7357918860bf03698794825b142: Status 404 returned error can't find the container with id a00ac3d1d66af99a414f5ec847d7c9229b50b7357918860bf03698794825b142 Dec 11 08:25:16 crc kubenswrapper[4860]: I1211 08:25:16.193510 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-gblg8" Dec 11 08:25:16 crc kubenswrapper[4860]: I1211 08:25:16.386183 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4wlfr\" (UniqueName: \"kubernetes.io/projected/bae66ca8-9554-4c05-adb3-e73819e23c1b-kube-api-access-4wlfr\") pod \"bae66ca8-9554-4c05-adb3-e73819e23c1b\" (UID: \"bae66ca8-9554-4c05-adb3-e73819e23c1b\") " Dec 11 08:25:16 crc kubenswrapper[4860]: I1211 08:25:16.395748 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bae66ca8-9554-4c05-adb3-e73819e23c1b-kube-api-access-4wlfr" (OuterVolumeSpecName: "kube-api-access-4wlfr") pod "bae66ca8-9554-4c05-adb3-e73819e23c1b" (UID: "bae66ca8-9554-4c05-adb3-e73819e23c1b"). InnerVolumeSpecName "kube-api-access-4wlfr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:25:16 crc kubenswrapper[4860]: I1211 08:25:16.491597 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4wlfr\" (UniqueName: \"kubernetes.io/projected/bae66ca8-9554-4c05-adb3-e73819e23c1b-kube-api-access-4wlfr\") on node \"crc\" DevicePath \"\"" Dec 11 08:25:16 crc kubenswrapper[4860]: I1211 08:25:16.840670 4860 generic.go:334] "Generic (PLEG): container finished" podID="bae66ca8-9554-4c05-adb3-e73819e23c1b" containerID="d99d69a38ff6989763758da2d2bd3b736813980c1dc832da5c8b39ddbd7145c1" exitCode=0 Dec 11 08:25:16 crc kubenswrapper[4860]: I1211 08:25:16.840757 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-gblg8" Dec 11 08:25:16 crc kubenswrapper[4860]: I1211 08:25:16.840763 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-gblg8" event={"ID":"bae66ca8-9554-4c05-adb3-e73819e23c1b","Type":"ContainerDied","Data":"d99d69a38ff6989763758da2d2bd3b736813980c1dc832da5c8b39ddbd7145c1"} Dec 11 08:25:16 crc kubenswrapper[4860]: I1211 08:25:16.840867 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-gblg8" event={"ID":"bae66ca8-9554-4c05-adb3-e73819e23c1b","Type":"ContainerDied","Data":"f5ce7d6fe252053d60410fdde592e1d640f1622fc3b2e75b6b98d3aa7998cc35"} Dec 11 08:25:16 crc kubenswrapper[4860]: I1211 08:25:16.840894 4860 scope.go:117] "RemoveContainer" containerID="d99d69a38ff6989763758da2d2bd3b736813980c1dc832da5c8b39ddbd7145c1" Dec 11 08:25:16 crc kubenswrapper[4860]: I1211 08:25:16.845605 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-hm8hr" event={"ID":"aec175c6-0b2d-40d8-855a-6132fb5e66a0","Type":"ContainerStarted","Data":"236aeab635005e41c03b531f90f97abbcb8cd8ff8303a7136d11c55806582033"} Dec 11 08:25:16 crc kubenswrapper[4860]: I1211 08:25:16.845724 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-hm8hr" event={"ID":"aec175c6-0b2d-40d8-855a-6132fb5e66a0","Type":"ContainerStarted","Data":"a00ac3d1d66af99a414f5ec847d7c9229b50b7357918860bf03698794825b142"} Dec 11 08:25:16 crc kubenswrapper[4860]: I1211 08:25:16.865164 4860 scope.go:117] "RemoveContainer" containerID="d99d69a38ff6989763758da2d2bd3b736813980c1dc832da5c8b39ddbd7145c1" Dec 11 08:25:16 crc kubenswrapper[4860]: I1211 08:25:16.864951 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-hm8hr" podStartSLOduration=1.8162255040000002 podStartE2EDuration="1.864922597s" podCreationTimestamp="2025-12-11 08:25:15 +0000 UTC" firstStartedPulling="2025-12-11 08:25:15.962966707 +0000 UTC m=+848.691485762" lastFinishedPulling="2025-12-11 08:25:16.0116638 +0000 UTC m=+848.740182855" observedRunningTime="2025-12-11 08:25:16.861590552 +0000 UTC m=+849.590109607" watchObservedRunningTime="2025-12-11 08:25:16.864922597 +0000 UTC m=+849.593441652" Dec 11 08:25:16 crc kubenswrapper[4860]: E1211 08:25:16.865822 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d99d69a38ff6989763758da2d2bd3b736813980c1dc832da5c8b39ddbd7145c1\": container with ID starting with d99d69a38ff6989763758da2d2bd3b736813980c1dc832da5c8b39ddbd7145c1 not found: ID does not exist" containerID="d99d69a38ff6989763758da2d2bd3b736813980c1dc832da5c8b39ddbd7145c1" Dec 11 08:25:16 crc kubenswrapper[4860]: I1211 08:25:16.865858 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d99d69a38ff6989763758da2d2bd3b736813980c1dc832da5c8b39ddbd7145c1"} err="failed to get container status \"d99d69a38ff6989763758da2d2bd3b736813980c1dc832da5c8b39ddbd7145c1\": rpc error: code = NotFound desc = could not find container \"d99d69a38ff6989763758da2d2bd3b736813980c1dc832da5c8b39ddbd7145c1\": container with ID starting with d99d69a38ff6989763758da2d2bd3b736813980c1dc832da5c8b39ddbd7145c1 not found: ID does not exist" Dec 11 08:25:16 crc kubenswrapper[4860]: I1211 08:25:16.878511 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-gblg8"] Dec 11 08:25:16 crc kubenswrapper[4860]: I1211 08:25:16.883493 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-gblg8"] Dec 11 08:25:17 crc kubenswrapper[4860]: I1211 08:25:17.135784 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-c6crs" Dec 11 08:25:17 crc kubenswrapper[4860]: I1211 08:25:17.467075 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-2548h" Dec 11 08:25:17 crc kubenswrapper[4860]: I1211 08:25:17.597714 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bae66ca8-9554-4c05-adb3-e73819e23c1b" path="/var/lib/kubelet/pods/bae66ca8-9554-4c05-adb3-e73819e23c1b/volumes" Dec 11 08:25:17 crc kubenswrapper[4860]: I1211 08:25:17.625841 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-5bddd4b946-96p7v" Dec 11 08:25:24 crc kubenswrapper[4860]: I1211 08:25:24.313048 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-ggdl7"] Dec 11 08:25:24 crc kubenswrapper[4860]: E1211 08:25:24.314506 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bae66ca8-9554-4c05-adb3-e73819e23c1b" containerName="registry-server" Dec 11 08:25:24 crc kubenswrapper[4860]: I1211 08:25:24.314526 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="bae66ca8-9554-4c05-adb3-e73819e23c1b" containerName="registry-server" Dec 11 08:25:24 crc kubenswrapper[4860]: I1211 08:25:24.314757 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="bae66ca8-9554-4c05-adb3-e73819e23c1b" containerName="registry-server" Dec 11 08:25:24 crc kubenswrapper[4860]: I1211 08:25:24.316515 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ggdl7" Dec 11 08:25:24 crc kubenswrapper[4860]: I1211 08:25:24.318598 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/929e897f-23a2-48ef-ac90-79e420fdf77f-utilities\") pod \"certified-operators-ggdl7\" (UID: \"929e897f-23a2-48ef-ac90-79e420fdf77f\") " pod="openshift-marketplace/certified-operators-ggdl7" Dec 11 08:25:24 crc kubenswrapper[4860]: I1211 08:25:24.318745 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8mdgv\" (UniqueName: \"kubernetes.io/projected/929e897f-23a2-48ef-ac90-79e420fdf77f-kube-api-access-8mdgv\") pod \"certified-operators-ggdl7\" (UID: \"929e897f-23a2-48ef-ac90-79e420fdf77f\") " pod="openshift-marketplace/certified-operators-ggdl7" Dec 11 08:25:24 crc kubenswrapper[4860]: I1211 08:25:24.318808 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/929e897f-23a2-48ef-ac90-79e420fdf77f-catalog-content\") pod \"certified-operators-ggdl7\" (UID: \"929e897f-23a2-48ef-ac90-79e420fdf77f\") " pod="openshift-marketplace/certified-operators-ggdl7" Dec 11 08:25:24 crc kubenswrapper[4860]: I1211 08:25:24.331019 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-ggdl7"] Dec 11 08:25:24 crc kubenswrapper[4860]: I1211 08:25:24.420332 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8mdgv\" (UniqueName: \"kubernetes.io/projected/929e897f-23a2-48ef-ac90-79e420fdf77f-kube-api-access-8mdgv\") pod \"certified-operators-ggdl7\" (UID: \"929e897f-23a2-48ef-ac90-79e420fdf77f\") " pod="openshift-marketplace/certified-operators-ggdl7" Dec 11 08:25:24 crc kubenswrapper[4860]: I1211 08:25:24.420628 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/929e897f-23a2-48ef-ac90-79e420fdf77f-catalog-content\") pod \"certified-operators-ggdl7\" (UID: \"929e897f-23a2-48ef-ac90-79e420fdf77f\") " pod="openshift-marketplace/certified-operators-ggdl7" Dec 11 08:25:24 crc kubenswrapper[4860]: I1211 08:25:24.420715 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/929e897f-23a2-48ef-ac90-79e420fdf77f-utilities\") pod \"certified-operators-ggdl7\" (UID: \"929e897f-23a2-48ef-ac90-79e420fdf77f\") " pod="openshift-marketplace/certified-operators-ggdl7" Dec 11 08:25:24 crc kubenswrapper[4860]: I1211 08:25:24.421326 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/929e897f-23a2-48ef-ac90-79e420fdf77f-utilities\") pod \"certified-operators-ggdl7\" (UID: \"929e897f-23a2-48ef-ac90-79e420fdf77f\") " pod="openshift-marketplace/certified-operators-ggdl7" Dec 11 08:25:24 crc kubenswrapper[4860]: I1211 08:25:24.421470 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/929e897f-23a2-48ef-ac90-79e420fdf77f-catalog-content\") pod \"certified-operators-ggdl7\" (UID: \"929e897f-23a2-48ef-ac90-79e420fdf77f\") " pod="openshift-marketplace/certified-operators-ggdl7" Dec 11 08:25:24 crc kubenswrapper[4860]: I1211 08:25:24.449572 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8mdgv\" (UniqueName: \"kubernetes.io/projected/929e897f-23a2-48ef-ac90-79e420fdf77f-kube-api-access-8mdgv\") pod \"certified-operators-ggdl7\" (UID: \"929e897f-23a2-48ef-ac90-79e420fdf77f\") " pod="openshift-marketplace/certified-operators-ggdl7" Dec 11 08:25:24 crc kubenswrapper[4860]: I1211 08:25:24.649533 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ggdl7" Dec 11 08:25:24 crc kubenswrapper[4860]: I1211 08:25:24.928136 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-ggdl7"] Dec 11 08:25:25 crc kubenswrapper[4860]: I1211 08:25:25.478376 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-hm8hr" Dec 11 08:25:25 crc kubenswrapper[4860]: I1211 08:25:25.478447 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-hm8hr" Dec 11 08:25:25 crc kubenswrapper[4860]: I1211 08:25:25.506755 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-hm8hr" Dec 11 08:25:25 crc kubenswrapper[4860]: I1211 08:25:25.916946 4860 generic.go:334] "Generic (PLEG): container finished" podID="929e897f-23a2-48ef-ac90-79e420fdf77f" containerID="2049598cc557986568b4f1539b295cfa17c17ae4bf8d0ab99458d59cebe928c1" exitCode=0 Dec 11 08:25:25 crc kubenswrapper[4860]: I1211 08:25:25.917061 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ggdl7" event={"ID":"929e897f-23a2-48ef-ac90-79e420fdf77f","Type":"ContainerDied","Data":"2049598cc557986568b4f1539b295cfa17c17ae4bf8d0ab99458d59cebe928c1"} Dec 11 08:25:25 crc kubenswrapper[4860]: I1211 08:25:25.917113 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ggdl7" event={"ID":"929e897f-23a2-48ef-ac90-79e420fdf77f","Type":"ContainerStarted","Data":"cfe93820a7634f7c4f852a3b454629fa4d33ba20dda01a93a62c9a54633fa52e"} Dec 11 08:25:25 crc kubenswrapper[4860]: I1211 08:25:25.942723 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-hm8hr" Dec 11 08:25:26 crc kubenswrapper[4860]: I1211 08:25:26.928912 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ggdl7" event={"ID":"929e897f-23a2-48ef-ac90-79e420fdf77f","Type":"ContainerStarted","Data":"7147f9b1be91971b5ae8ea2b0a4371e07e6071a177847857fbde1fb070142170"} Dec 11 08:25:27 crc kubenswrapper[4860]: I1211 08:25:27.937127 4860 generic.go:334] "Generic (PLEG): container finished" podID="929e897f-23a2-48ef-ac90-79e420fdf77f" containerID="7147f9b1be91971b5ae8ea2b0a4371e07e6071a177847857fbde1fb070142170" exitCode=0 Dec 11 08:25:27 crc kubenswrapper[4860]: I1211 08:25:27.937246 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ggdl7" event={"ID":"929e897f-23a2-48ef-ac90-79e420fdf77f","Type":"ContainerDied","Data":"7147f9b1be91971b5ae8ea2b0a4371e07e6071a177847857fbde1fb070142170"} Dec 11 08:25:27 crc kubenswrapper[4860]: I1211 08:25:27.981573 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/941ec9b01b7a72127ba64306553d6766ab71ba59f32f9ca01c76932a62n7v8f"] Dec 11 08:25:27 crc kubenswrapper[4860]: I1211 08:25:27.983122 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/941ec9b01b7a72127ba64306553d6766ab71ba59f32f9ca01c76932a62n7v8f" Dec 11 08:25:27 crc kubenswrapper[4860]: I1211 08:25:27.988391 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-c95bg" Dec 11 08:25:27 crc kubenswrapper[4860]: I1211 08:25:27.997942 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/941ec9b01b7a72127ba64306553d6766ab71ba59f32f9ca01c76932a62n7v8f"] Dec 11 08:25:28 crc kubenswrapper[4860]: I1211 08:25:28.180771 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tjfv5\" (UniqueName: \"kubernetes.io/projected/bf285af3-cbc9-4e8f-bb3a-bc1d16a3ea54-kube-api-access-tjfv5\") pod \"941ec9b01b7a72127ba64306553d6766ab71ba59f32f9ca01c76932a62n7v8f\" (UID: \"bf285af3-cbc9-4e8f-bb3a-bc1d16a3ea54\") " pod="openstack-operators/941ec9b01b7a72127ba64306553d6766ab71ba59f32f9ca01c76932a62n7v8f" Dec 11 08:25:28 crc kubenswrapper[4860]: I1211 08:25:28.180853 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bf285af3-cbc9-4e8f-bb3a-bc1d16a3ea54-util\") pod \"941ec9b01b7a72127ba64306553d6766ab71ba59f32f9ca01c76932a62n7v8f\" (UID: \"bf285af3-cbc9-4e8f-bb3a-bc1d16a3ea54\") " pod="openstack-operators/941ec9b01b7a72127ba64306553d6766ab71ba59f32f9ca01c76932a62n7v8f" Dec 11 08:25:28 crc kubenswrapper[4860]: I1211 08:25:28.180971 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bf285af3-cbc9-4e8f-bb3a-bc1d16a3ea54-bundle\") pod \"941ec9b01b7a72127ba64306553d6766ab71ba59f32f9ca01c76932a62n7v8f\" (UID: \"bf285af3-cbc9-4e8f-bb3a-bc1d16a3ea54\") " pod="openstack-operators/941ec9b01b7a72127ba64306553d6766ab71ba59f32f9ca01c76932a62n7v8f" Dec 11 08:25:28 crc kubenswrapper[4860]: I1211 08:25:28.282937 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bf285af3-cbc9-4e8f-bb3a-bc1d16a3ea54-util\") pod \"941ec9b01b7a72127ba64306553d6766ab71ba59f32f9ca01c76932a62n7v8f\" (UID: \"bf285af3-cbc9-4e8f-bb3a-bc1d16a3ea54\") " pod="openstack-operators/941ec9b01b7a72127ba64306553d6766ab71ba59f32f9ca01c76932a62n7v8f" Dec 11 08:25:28 crc kubenswrapper[4860]: I1211 08:25:28.283071 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bf285af3-cbc9-4e8f-bb3a-bc1d16a3ea54-bundle\") pod \"941ec9b01b7a72127ba64306553d6766ab71ba59f32f9ca01c76932a62n7v8f\" (UID: \"bf285af3-cbc9-4e8f-bb3a-bc1d16a3ea54\") " pod="openstack-operators/941ec9b01b7a72127ba64306553d6766ab71ba59f32f9ca01c76932a62n7v8f" Dec 11 08:25:28 crc kubenswrapper[4860]: I1211 08:25:28.283110 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tjfv5\" (UniqueName: \"kubernetes.io/projected/bf285af3-cbc9-4e8f-bb3a-bc1d16a3ea54-kube-api-access-tjfv5\") pod \"941ec9b01b7a72127ba64306553d6766ab71ba59f32f9ca01c76932a62n7v8f\" (UID: \"bf285af3-cbc9-4e8f-bb3a-bc1d16a3ea54\") " pod="openstack-operators/941ec9b01b7a72127ba64306553d6766ab71ba59f32f9ca01c76932a62n7v8f" Dec 11 08:25:28 crc kubenswrapper[4860]: I1211 08:25:28.283964 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bf285af3-cbc9-4e8f-bb3a-bc1d16a3ea54-util\") pod \"941ec9b01b7a72127ba64306553d6766ab71ba59f32f9ca01c76932a62n7v8f\" (UID: \"bf285af3-cbc9-4e8f-bb3a-bc1d16a3ea54\") " pod="openstack-operators/941ec9b01b7a72127ba64306553d6766ab71ba59f32f9ca01c76932a62n7v8f" Dec 11 08:25:28 crc kubenswrapper[4860]: I1211 08:25:28.284057 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bf285af3-cbc9-4e8f-bb3a-bc1d16a3ea54-bundle\") pod \"941ec9b01b7a72127ba64306553d6766ab71ba59f32f9ca01c76932a62n7v8f\" (UID: \"bf285af3-cbc9-4e8f-bb3a-bc1d16a3ea54\") " pod="openstack-operators/941ec9b01b7a72127ba64306553d6766ab71ba59f32f9ca01c76932a62n7v8f" Dec 11 08:25:28 crc kubenswrapper[4860]: I1211 08:25:28.313772 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tjfv5\" (UniqueName: \"kubernetes.io/projected/bf285af3-cbc9-4e8f-bb3a-bc1d16a3ea54-kube-api-access-tjfv5\") pod \"941ec9b01b7a72127ba64306553d6766ab71ba59f32f9ca01c76932a62n7v8f\" (UID: \"bf285af3-cbc9-4e8f-bb3a-bc1d16a3ea54\") " pod="openstack-operators/941ec9b01b7a72127ba64306553d6766ab71ba59f32f9ca01c76932a62n7v8f" Dec 11 08:25:28 crc kubenswrapper[4860]: I1211 08:25:28.605911 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/941ec9b01b7a72127ba64306553d6766ab71ba59f32f9ca01c76932a62n7v8f" Dec 11 08:25:28 crc kubenswrapper[4860]: I1211 08:25:28.948209 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ggdl7" event={"ID":"929e897f-23a2-48ef-ac90-79e420fdf77f","Type":"ContainerStarted","Data":"55fc174489509d0658bcb9a2d28bb67ae41f95a3444f5fc9c4592f7b8438417f"} Dec 11 08:25:29 crc kubenswrapper[4860]: I1211 08:25:29.050276 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-ggdl7" podStartSLOduration=2.617181161 podStartE2EDuration="5.05025212s" podCreationTimestamp="2025-12-11 08:25:24 +0000 UTC" firstStartedPulling="2025-12-11 08:25:25.918842263 +0000 UTC m=+858.647361318" lastFinishedPulling="2025-12-11 08:25:28.351913222 +0000 UTC m=+861.080432277" observedRunningTime="2025-12-11 08:25:28.970174629 +0000 UTC m=+861.698693704" watchObservedRunningTime="2025-12-11 08:25:29.05025212 +0000 UTC m=+861.778771175" Dec 11 08:25:29 crc kubenswrapper[4860]: I1211 08:25:29.051730 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/941ec9b01b7a72127ba64306553d6766ab71ba59f32f9ca01c76932a62n7v8f"] Dec 11 08:25:29 crc kubenswrapper[4860]: I1211 08:25:29.956785 4860 generic.go:334] "Generic (PLEG): container finished" podID="bf285af3-cbc9-4e8f-bb3a-bc1d16a3ea54" containerID="c6df1026a3e43576b6626013ee3c654e8028bd2bbc7d5908e369931116999472" exitCode=0 Dec 11 08:25:29 crc kubenswrapper[4860]: I1211 08:25:29.956886 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/941ec9b01b7a72127ba64306553d6766ab71ba59f32f9ca01c76932a62n7v8f" event={"ID":"bf285af3-cbc9-4e8f-bb3a-bc1d16a3ea54","Type":"ContainerDied","Data":"c6df1026a3e43576b6626013ee3c654e8028bd2bbc7d5908e369931116999472"} Dec 11 08:25:29 crc kubenswrapper[4860]: I1211 08:25:29.957307 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/941ec9b01b7a72127ba64306553d6766ab71ba59f32f9ca01c76932a62n7v8f" event={"ID":"bf285af3-cbc9-4e8f-bb3a-bc1d16a3ea54","Type":"ContainerStarted","Data":"43f4319087143d08f0f52612adecb1bd3fb2f625ddfed68cc4fed4fffe5f3023"} Dec 11 08:25:30 crc kubenswrapper[4860]: I1211 08:25:30.965553 4860 generic.go:334] "Generic (PLEG): container finished" podID="bf285af3-cbc9-4e8f-bb3a-bc1d16a3ea54" containerID="a601eac88119aa4cebd0a13f21ff911a2ba23d8cb2b4b3f74dee5dabc5bd2933" exitCode=0 Dec 11 08:25:30 crc kubenswrapper[4860]: I1211 08:25:30.965657 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/941ec9b01b7a72127ba64306553d6766ab71ba59f32f9ca01c76932a62n7v8f" event={"ID":"bf285af3-cbc9-4e8f-bb3a-bc1d16a3ea54","Type":"ContainerDied","Data":"a601eac88119aa4cebd0a13f21ff911a2ba23d8cb2b4b3f74dee5dabc5bd2933"} Dec 11 08:25:31 crc kubenswrapper[4860]: I1211 08:25:31.975763 4860 generic.go:334] "Generic (PLEG): container finished" podID="bf285af3-cbc9-4e8f-bb3a-bc1d16a3ea54" containerID="a921104b94d0fbf7031bb132c15219ebe13895107670b1673e0b09cbbe836696" exitCode=0 Dec 11 08:25:31 crc kubenswrapper[4860]: I1211 08:25:31.975839 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/941ec9b01b7a72127ba64306553d6766ab71ba59f32f9ca01c76932a62n7v8f" event={"ID":"bf285af3-cbc9-4e8f-bb3a-bc1d16a3ea54","Type":"ContainerDied","Data":"a921104b94d0fbf7031bb132c15219ebe13895107670b1673e0b09cbbe836696"} Dec 11 08:25:33 crc kubenswrapper[4860]: I1211 08:25:33.334933 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/941ec9b01b7a72127ba64306553d6766ab71ba59f32f9ca01c76932a62n7v8f" Dec 11 08:25:33 crc kubenswrapper[4860]: I1211 08:25:33.467989 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bf285af3-cbc9-4e8f-bb3a-bc1d16a3ea54-util\") pod \"bf285af3-cbc9-4e8f-bb3a-bc1d16a3ea54\" (UID: \"bf285af3-cbc9-4e8f-bb3a-bc1d16a3ea54\") " Dec 11 08:25:33 crc kubenswrapper[4860]: I1211 08:25:33.468071 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tjfv5\" (UniqueName: \"kubernetes.io/projected/bf285af3-cbc9-4e8f-bb3a-bc1d16a3ea54-kube-api-access-tjfv5\") pod \"bf285af3-cbc9-4e8f-bb3a-bc1d16a3ea54\" (UID: \"bf285af3-cbc9-4e8f-bb3a-bc1d16a3ea54\") " Dec 11 08:25:33 crc kubenswrapper[4860]: I1211 08:25:33.468165 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bf285af3-cbc9-4e8f-bb3a-bc1d16a3ea54-bundle\") pod \"bf285af3-cbc9-4e8f-bb3a-bc1d16a3ea54\" (UID: \"bf285af3-cbc9-4e8f-bb3a-bc1d16a3ea54\") " Dec 11 08:25:33 crc kubenswrapper[4860]: I1211 08:25:33.469316 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bf285af3-cbc9-4e8f-bb3a-bc1d16a3ea54-bundle" (OuterVolumeSpecName: "bundle") pod "bf285af3-cbc9-4e8f-bb3a-bc1d16a3ea54" (UID: "bf285af3-cbc9-4e8f-bb3a-bc1d16a3ea54"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:25:33 crc kubenswrapper[4860]: I1211 08:25:33.476599 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf285af3-cbc9-4e8f-bb3a-bc1d16a3ea54-kube-api-access-tjfv5" (OuterVolumeSpecName: "kube-api-access-tjfv5") pod "bf285af3-cbc9-4e8f-bb3a-bc1d16a3ea54" (UID: "bf285af3-cbc9-4e8f-bb3a-bc1d16a3ea54"). InnerVolumeSpecName "kube-api-access-tjfv5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:25:33 crc kubenswrapper[4860]: I1211 08:25:33.484109 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bf285af3-cbc9-4e8f-bb3a-bc1d16a3ea54-util" (OuterVolumeSpecName: "util") pod "bf285af3-cbc9-4e8f-bb3a-bc1d16a3ea54" (UID: "bf285af3-cbc9-4e8f-bb3a-bc1d16a3ea54"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:25:33 crc kubenswrapper[4860]: I1211 08:25:33.570056 4860 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bf285af3-cbc9-4e8f-bb3a-bc1d16a3ea54-util\") on node \"crc\" DevicePath \"\"" Dec 11 08:25:33 crc kubenswrapper[4860]: I1211 08:25:33.570100 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tjfv5\" (UniqueName: \"kubernetes.io/projected/bf285af3-cbc9-4e8f-bb3a-bc1d16a3ea54-kube-api-access-tjfv5\") on node \"crc\" DevicePath \"\"" Dec 11 08:25:33 crc kubenswrapper[4860]: I1211 08:25:33.570114 4860 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bf285af3-cbc9-4e8f-bb3a-bc1d16a3ea54-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 08:25:33 crc kubenswrapper[4860]: I1211 08:25:33.992276 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/941ec9b01b7a72127ba64306553d6766ab71ba59f32f9ca01c76932a62n7v8f" event={"ID":"bf285af3-cbc9-4e8f-bb3a-bc1d16a3ea54","Type":"ContainerDied","Data":"43f4319087143d08f0f52612adecb1bd3fb2f625ddfed68cc4fed4fffe5f3023"} Dec 11 08:25:33 crc kubenswrapper[4860]: I1211 08:25:33.992333 4860 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="43f4319087143d08f0f52612adecb1bd3fb2f625ddfed68cc4fed4fffe5f3023" Dec 11 08:25:33 crc kubenswrapper[4860]: I1211 08:25:33.992409 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/941ec9b01b7a72127ba64306553d6766ab71ba59f32f9ca01c76932a62n7v8f" Dec 11 08:25:34 crc kubenswrapper[4860]: I1211 08:25:34.650125 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-ggdl7" Dec 11 08:25:34 crc kubenswrapper[4860]: I1211 08:25:34.650967 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-ggdl7" Dec 11 08:25:34 crc kubenswrapper[4860]: I1211 08:25:34.696823 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-ggdl7" Dec 11 08:25:35 crc kubenswrapper[4860]: I1211 08:25:35.037998 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-ggdl7" Dec 11 08:25:36 crc kubenswrapper[4860]: I1211 08:25:36.743994 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-7d57466ffb-mg466"] Dec 11 08:25:36 crc kubenswrapper[4860]: E1211 08:25:36.744718 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf285af3-cbc9-4e8f-bb3a-bc1d16a3ea54" containerName="util" Dec 11 08:25:36 crc kubenswrapper[4860]: I1211 08:25:36.744734 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf285af3-cbc9-4e8f-bb3a-bc1d16a3ea54" containerName="util" Dec 11 08:25:36 crc kubenswrapper[4860]: E1211 08:25:36.744752 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf285af3-cbc9-4e8f-bb3a-bc1d16a3ea54" containerName="extract" Dec 11 08:25:36 crc kubenswrapper[4860]: I1211 08:25:36.744764 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf285af3-cbc9-4e8f-bb3a-bc1d16a3ea54" containerName="extract" Dec 11 08:25:36 crc kubenswrapper[4860]: E1211 08:25:36.744781 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf285af3-cbc9-4e8f-bb3a-bc1d16a3ea54" containerName="pull" Dec 11 08:25:36 crc kubenswrapper[4860]: I1211 08:25:36.744787 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf285af3-cbc9-4e8f-bb3a-bc1d16a3ea54" containerName="pull" Dec 11 08:25:36 crc kubenswrapper[4860]: I1211 08:25:36.744907 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf285af3-cbc9-4e8f-bb3a-bc1d16a3ea54" containerName="extract" Dec 11 08:25:36 crc kubenswrapper[4860]: I1211 08:25:36.745418 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-7d57466ffb-mg466" Dec 11 08:25:36 crc kubenswrapper[4860]: I1211 08:25:36.756398 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-5crkf" Dec 11 08:25:36 crc kubenswrapper[4860]: I1211 08:25:36.790109 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-7d57466ffb-mg466"] Dec 11 08:25:36 crc kubenswrapper[4860]: I1211 08:25:36.932864 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2ztjq\" (UniqueName: \"kubernetes.io/projected/f57d6313-d688-4a8d-a9cb-face8d0e0ef6-kube-api-access-2ztjq\") pod \"openstack-operator-controller-operator-7d57466ffb-mg466\" (UID: \"f57d6313-d688-4a8d-a9cb-face8d0e0ef6\") " pod="openstack-operators/openstack-operator-controller-operator-7d57466ffb-mg466" Dec 11 08:25:37 crc kubenswrapper[4860]: I1211 08:25:37.034687 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2ztjq\" (UniqueName: \"kubernetes.io/projected/f57d6313-d688-4a8d-a9cb-face8d0e0ef6-kube-api-access-2ztjq\") pod \"openstack-operator-controller-operator-7d57466ffb-mg466\" (UID: \"f57d6313-d688-4a8d-a9cb-face8d0e0ef6\") " pod="openstack-operators/openstack-operator-controller-operator-7d57466ffb-mg466" Dec 11 08:25:37 crc kubenswrapper[4860]: I1211 08:25:37.057101 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2ztjq\" (UniqueName: \"kubernetes.io/projected/f57d6313-d688-4a8d-a9cb-face8d0e0ef6-kube-api-access-2ztjq\") pod \"openstack-operator-controller-operator-7d57466ffb-mg466\" (UID: \"f57d6313-d688-4a8d-a9cb-face8d0e0ef6\") " pod="openstack-operators/openstack-operator-controller-operator-7d57466ffb-mg466" Dec 11 08:25:37 crc kubenswrapper[4860]: I1211 08:25:37.069528 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-7d57466ffb-mg466" Dec 11 08:25:37 crc kubenswrapper[4860]: I1211 08:25:37.704016 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-7d57466ffb-mg466"] Dec 11 08:25:37 crc kubenswrapper[4860]: W1211 08:25:37.712454 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf57d6313_d688_4a8d_a9cb_face8d0e0ef6.slice/crio-8147bb104bafcff27b3f3544e54990ef607d0b16936a5a6312e4fbbc307842a1 WatchSource:0}: Error finding container 8147bb104bafcff27b3f3544e54990ef607d0b16936a5a6312e4fbbc307842a1: Status 404 returned error can't find the container with id 8147bb104bafcff27b3f3544e54990ef607d0b16936a5a6312e4fbbc307842a1 Dec 11 08:25:38 crc kubenswrapper[4860]: I1211 08:25:38.021039 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-7d57466ffb-mg466" event={"ID":"f57d6313-d688-4a8d-a9cb-face8d0e0ef6","Type":"ContainerStarted","Data":"8147bb104bafcff27b3f3544e54990ef607d0b16936a5a6312e4fbbc307842a1"} Dec 11 08:25:38 crc kubenswrapper[4860]: I1211 08:25:38.297607 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-ggdl7"] Dec 11 08:25:38 crc kubenswrapper[4860]: I1211 08:25:38.298238 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-ggdl7" podUID="929e897f-23a2-48ef-ac90-79e420fdf77f" containerName="registry-server" containerID="cri-o://55fc174489509d0658bcb9a2d28bb67ae41f95a3444f5fc9c4592f7b8438417f" gracePeriod=2 Dec 11 08:25:38 crc kubenswrapper[4860]: I1211 08:25:38.506176 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-nkzr6"] Dec 11 08:25:38 crc kubenswrapper[4860]: I1211 08:25:38.507697 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nkzr6" Dec 11 08:25:38 crc kubenswrapper[4860]: I1211 08:25:38.528186 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nkzr6"] Dec 11 08:25:38 crc kubenswrapper[4860]: I1211 08:25:38.565538 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d7724a83-b148-4b65-839a-3d85ecbbc4a0-utilities\") pod \"community-operators-nkzr6\" (UID: \"d7724a83-b148-4b65-839a-3d85ecbbc4a0\") " pod="openshift-marketplace/community-operators-nkzr6" Dec 11 08:25:38 crc kubenswrapper[4860]: I1211 08:25:38.565592 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-76nqn\" (UniqueName: \"kubernetes.io/projected/d7724a83-b148-4b65-839a-3d85ecbbc4a0-kube-api-access-76nqn\") pod \"community-operators-nkzr6\" (UID: \"d7724a83-b148-4b65-839a-3d85ecbbc4a0\") " pod="openshift-marketplace/community-operators-nkzr6" Dec 11 08:25:38 crc kubenswrapper[4860]: I1211 08:25:38.565702 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d7724a83-b148-4b65-839a-3d85ecbbc4a0-catalog-content\") pod \"community-operators-nkzr6\" (UID: \"d7724a83-b148-4b65-839a-3d85ecbbc4a0\") " pod="openshift-marketplace/community-operators-nkzr6" Dec 11 08:25:38 crc kubenswrapper[4860]: I1211 08:25:38.667406 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d7724a83-b148-4b65-839a-3d85ecbbc4a0-catalog-content\") pod \"community-operators-nkzr6\" (UID: \"d7724a83-b148-4b65-839a-3d85ecbbc4a0\") " pod="openshift-marketplace/community-operators-nkzr6" Dec 11 08:25:38 crc kubenswrapper[4860]: I1211 08:25:38.667511 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d7724a83-b148-4b65-839a-3d85ecbbc4a0-utilities\") pod \"community-operators-nkzr6\" (UID: \"d7724a83-b148-4b65-839a-3d85ecbbc4a0\") " pod="openshift-marketplace/community-operators-nkzr6" Dec 11 08:25:38 crc kubenswrapper[4860]: I1211 08:25:38.667550 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-76nqn\" (UniqueName: \"kubernetes.io/projected/d7724a83-b148-4b65-839a-3d85ecbbc4a0-kube-api-access-76nqn\") pod \"community-operators-nkzr6\" (UID: \"d7724a83-b148-4b65-839a-3d85ecbbc4a0\") " pod="openshift-marketplace/community-operators-nkzr6" Dec 11 08:25:38 crc kubenswrapper[4860]: I1211 08:25:38.668316 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d7724a83-b148-4b65-839a-3d85ecbbc4a0-utilities\") pod \"community-operators-nkzr6\" (UID: \"d7724a83-b148-4b65-839a-3d85ecbbc4a0\") " pod="openshift-marketplace/community-operators-nkzr6" Dec 11 08:25:38 crc kubenswrapper[4860]: I1211 08:25:38.668346 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d7724a83-b148-4b65-839a-3d85ecbbc4a0-catalog-content\") pod \"community-operators-nkzr6\" (UID: \"d7724a83-b148-4b65-839a-3d85ecbbc4a0\") " pod="openshift-marketplace/community-operators-nkzr6" Dec 11 08:25:38 crc kubenswrapper[4860]: I1211 08:25:38.692079 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-76nqn\" (UniqueName: \"kubernetes.io/projected/d7724a83-b148-4b65-839a-3d85ecbbc4a0-kube-api-access-76nqn\") pod \"community-operators-nkzr6\" (UID: \"d7724a83-b148-4b65-839a-3d85ecbbc4a0\") " pod="openshift-marketplace/community-operators-nkzr6" Dec 11 08:25:38 crc kubenswrapper[4860]: I1211 08:25:38.823771 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nkzr6" Dec 11 08:25:39 crc kubenswrapper[4860]: I1211 08:25:39.345154 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nkzr6"] Dec 11 08:25:40 crc kubenswrapper[4860]: I1211 08:25:40.039606 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nkzr6" event={"ID":"d7724a83-b148-4b65-839a-3d85ecbbc4a0","Type":"ContainerStarted","Data":"5107e38a1e17f43e27b0b4837f02d1fef84f81de267c807f1e553743c165a83d"} Dec 11 08:25:40 crc kubenswrapper[4860]: I1211 08:25:40.043058 4860 generic.go:334] "Generic (PLEG): container finished" podID="929e897f-23a2-48ef-ac90-79e420fdf77f" containerID="55fc174489509d0658bcb9a2d28bb67ae41f95a3444f5fc9c4592f7b8438417f" exitCode=0 Dec 11 08:25:40 crc kubenswrapper[4860]: I1211 08:25:40.043103 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ggdl7" event={"ID":"929e897f-23a2-48ef-ac90-79e420fdf77f","Type":"ContainerDied","Data":"55fc174489509d0658bcb9a2d28bb67ae41f95a3444f5fc9c4592f7b8438417f"} Dec 11 08:25:40 crc kubenswrapper[4860]: I1211 08:25:40.709260 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ggdl7" Dec 11 08:25:40 crc kubenswrapper[4860]: I1211 08:25:40.901071 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8mdgv\" (UniqueName: \"kubernetes.io/projected/929e897f-23a2-48ef-ac90-79e420fdf77f-kube-api-access-8mdgv\") pod \"929e897f-23a2-48ef-ac90-79e420fdf77f\" (UID: \"929e897f-23a2-48ef-ac90-79e420fdf77f\") " Dec 11 08:25:40 crc kubenswrapper[4860]: I1211 08:25:40.901312 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/929e897f-23a2-48ef-ac90-79e420fdf77f-utilities\") pod \"929e897f-23a2-48ef-ac90-79e420fdf77f\" (UID: \"929e897f-23a2-48ef-ac90-79e420fdf77f\") " Dec 11 08:25:40 crc kubenswrapper[4860]: I1211 08:25:40.901382 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/929e897f-23a2-48ef-ac90-79e420fdf77f-catalog-content\") pod \"929e897f-23a2-48ef-ac90-79e420fdf77f\" (UID: \"929e897f-23a2-48ef-ac90-79e420fdf77f\") " Dec 11 08:25:40 crc kubenswrapper[4860]: I1211 08:25:40.902841 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/929e897f-23a2-48ef-ac90-79e420fdf77f-utilities" (OuterVolumeSpecName: "utilities") pod "929e897f-23a2-48ef-ac90-79e420fdf77f" (UID: "929e897f-23a2-48ef-ac90-79e420fdf77f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:25:40 crc kubenswrapper[4860]: I1211 08:25:40.909520 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/929e897f-23a2-48ef-ac90-79e420fdf77f-kube-api-access-8mdgv" (OuterVolumeSpecName: "kube-api-access-8mdgv") pod "929e897f-23a2-48ef-ac90-79e420fdf77f" (UID: "929e897f-23a2-48ef-ac90-79e420fdf77f"). InnerVolumeSpecName "kube-api-access-8mdgv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:25:40 crc kubenswrapper[4860]: I1211 08:25:40.958527 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/929e897f-23a2-48ef-ac90-79e420fdf77f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "929e897f-23a2-48ef-ac90-79e420fdf77f" (UID: "929e897f-23a2-48ef-ac90-79e420fdf77f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:25:41 crc kubenswrapper[4860]: I1211 08:25:41.003221 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8mdgv\" (UniqueName: \"kubernetes.io/projected/929e897f-23a2-48ef-ac90-79e420fdf77f-kube-api-access-8mdgv\") on node \"crc\" DevicePath \"\"" Dec 11 08:25:41 crc kubenswrapper[4860]: I1211 08:25:41.003267 4860 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/929e897f-23a2-48ef-ac90-79e420fdf77f-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 08:25:41 crc kubenswrapper[4860]: I1211 08:25:41.003277 4860 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/929e897f-23a2-48ef-ac90-79e420fdf77f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 08:25:41 crc kubenswrapper[4860]: I1211 08:25:41.064160 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ggdl7" event={"ID":"929e897f-23a2-48ef-ac90-79e420fdf77f","Type":"ContainerDied","Data":"cfe93820a7634f7c4f852a3b454629fa4d33ba20dda01a93a62c9a54633fa52e"} Dec 11 08:25:41 crc kubenswrapper[4860]: I1211 08:25:41.064248 4860 scope.go:117] "RemoveContainer" containerID="55fc174489509d0658bcb9a2d28bb67ae41f95a3444f5fc9c4592f7b8438417f" Dec 11 08:25:41 crc kubenswrapper[4860]: I1211 08:25:41.064457 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ggdl7" Dec 11 08:25:41 crc kubenswrapper[4860]: I1211 08:25:41.071374 4860 generic.go:334] "Generic (PLEG): container finished" podID="d7724a83-b148-4b65-839a-3d85ecbbc4a0" containerID="46344f92ce53b61ab1f7db7709de8b758ffb8bc3f32a47dab65465255bd1706f" exitCode=0 Dec 11 08:25:41 crc kubenswrapper[4860]: I1211 08:25:41.071425 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nkzr6" event={"ID":"d7724a83-b148-4b65-839a-3d85ecbbc4a0","Type":"ContainerDied","Data":"46344f92ce53b61ab1f7db7709de8b758ffb8bc3f32a47dab65465255bd1706f"} Dec 11 08:25:41 crc kubenswrapper[4860]: I1211 08:25:41.101894 4860 scope.go:117] "RemoveContainer" containerID="7147f9b1be91971b5ae8ea2b0a4371e07e6071a177847857fbde1fb070142170" Dec 11 08:25:41 crc kubenswrapper[4860]: I1211 08:25:41.116711 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-ggdl7"] Dec 11 08:25:41 crc kubenswrapper[4860]: I1211 08:25:41.126621 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-ggdl7"] Dec 11 08:25:41 crc kubenswrapper[4860]: I1211 08:25:41.133722 4860 scope.go:117] "RemoveContainer" containerID="2049598cc557986568b4f1539b295cfa17c17ae4bf8d0ab99458d59cebe928c1" Dec 11 08:25:41 crc kubenswrapper[4860]: I1211 08:25:41.589551 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="929e897f-23a2-48ef-ac90-79e420fdf77f" path="/var/lib/kubelet/pods/929e897f-23a2-48ef-ac90-79e420fdf77f/volumes" Dec 11 08:25:45 crc kubenswrapper[4860]: I1211 08:25:45.115073 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-7d57466ffb-mg466" event={"ID":"f57d6313-d688-4a8d-a9cb-face8d0e0ef6","Type":"ContainerStarted","Data":"bbb286289065a9e673d46f3cfb43759c4319db1183bb1aad33a54b62acf04b1d"} Dec 11 08:25:45 crc kubenswrapper[4860]: I1211 08:25:45.115709 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-7d57466ffb-mg466" Dec 11 08:25:45 crc kubenswrapper[4860]: I1211 08:25:45.120376 4860 generic.go:334] "Generic (PLEG): container finished" podID="d7724a83-b148-4b65-839a-3d85ecbbc4a0" containerID="51dd5eee6540f7d3e36a7a502a1f1549c02a41b122885a0b0f302fedfb79b957" exitCode=0 Dec 11 08:25:45 crc kubenswrapper[4860]: I1211 08:25:45.120433 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nkzr6" event={"ID":"d7724a83-b148-4b65-839a-3d85ecbbc4a0","Type":"ContainerDied","Data":"51dd5eee6540f7d3e36a7a502a1f1549c02a41b122885a0b0f302fedfb79b957"} Dec 11 08:25:45 crc kubenswrapper[4860]: I1211 08:25:45.150723 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-7d57466ffb-mg466" podStartSLOduration=2.364394509 podStartE2EDuration="9.150697876s" podCreationTimestamp="2025-12-11 08:25:36 +0000 UTC" firstStartedPulling="2025-12-11 08:25:37.716256251 +0000 UTC m=+870.444775306" lastFinishedPulling="2025-12-11 08:25:44.502559618 +0000 UTC m=+877.231078673" observedRunningTime="2025-12-11 08:25:45.145155281 +0000 UTC m=+877.873674366" watchObservedRunningTime="2025-12-11 08:25:45.150697876 +0000 UTC m=+877.879216921" Dec 11 08:25:46 crc kubenswrapper[4860]: I1211 08:25:46.131633 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nkzr6" event={"ID":"d7724a83-b148-4b65-839a-3d85ecbbc4a0","Type":"ContainerStarted","Data":"2e5b3b5a9c5df0e9218c8ae77647ad1ee41ee767ff5e254d4adbefae795a48c6"} Dec 11 08:25:46 crc kubenswrapper[4860]: I1211 08:25:46.158313 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-nkzr6" podStartSLOduration=3.382286843 podStartE2EDuration="8.158279791s" podCreationTimestamp="2025-12-11 08:25:38 +0000 UTC" firstStartedPulling="2025-12-11 08:25:41.074582344 +0000 UTC m=+873.803101399" lastFinishedPulling="2025-12-11 08:25:45.850575292 +0000 UTC m=+878.579094347" observedRunningTime="2025-12-11 08:25:46.153369825 +0000 UTC m=+878.881888890" watchObservedRunningTime="2025-12-11 08:25:46.158279791 +0000 UTC m=+878.886798846" Dec 11 08:25:48 crc kubenswrapper[4860]: I1211 08:25:48.824878 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-nkzr6" Dec 11 08:25:48 crc kubenswrapper[4860]: I1211 08:25:48.825318 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-nkzr6" Dec 11 08:25:48 crc kubenswrapper[4860]: I1211 08:25:48.869028 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-nkzr6" Dec 11 08:25:57 crc kubenswrapper[4860]: I1211 08:25:57.072467 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-7d57466ffb-mg466" Dec 11 08:25:58 crc kubenswrapper[4860]: I1211 08:25:58.873994 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-nkzr6" Dec 11 08:25:58 crc kubenswrapper[4860]: I1211 08:25:58.930465 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nkzr6"] Dec 11 08:25:59 crc kubenswrapper[4860]: I1211 08:25:59.216772 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-nkzr6" podUID="d7724a83-b148-4b65-839a-3d85ecbbc4a0" containerName="registry-server" containerID="cri-o://2e5b3b5a9c5df0e9218c8ae77647ad1ee41ee767ff5e254d4adbefae795a48c6" gracePeriod=2 Dec 11 08:26:00 crc kubenswrapper[4860]: I1211 08:26:00.112505 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nkzr6" Dec 11 08:26:00 crc kubenswrapper[4860]: I1211 08:26:00.224795 4860 generic.go:334] "Generic (PLEG): container finished" podID="d7724a83-b148-4b65-839a-3d85ecbbc4a0" containerID="2e5b3b5a9c5df0e9218c8ae77647ad1ee41ee767ff5e254d4adbefae795a48c6" exitCode=0 Dec 11 08:26:00 crc kubenswrapper[4860]: I1211 08:26:00.224850 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nkzr6" event={"ID":"d7724a83-b148-4b65-839a-3d85ecbbc4a0","Type":"ContainerDied","Data":"2e5b3b5a9c5df0e9218c8ae77647ad1ee41ee767ff5e254d4adbefae795a48c6"} Dec 11 08:26:00 crc kubenswrapper[4860]: I1211 08:26:00.224856 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nkzr6" Dec 11 08:26:00 crc kubenswrapper[4860]: I1211 08:26:00.224889 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nkzr6" event={"ID":"d7724a83-b148-4b65-839a-3d85ecbbc4a0","Type":"ContainerDied","Data":"5107e38a1e17f43e27b0b4837f02d1fef84f81de267c807f1e553743c165a83d"} Dec 11 08:26:00 crc kubenswrapper[4860]: I1211 08:26:00.224911 4860 scope.go:117] "RemoveContainer" containerID="2e5b3b5a9c5df0e9218c8ae77647ad1ee41ee767ff5e254d4adbefae795a48c6" Dec 11 08:26:00 crc kubenswrapper[4860]: I1211 08:26:00.242249 4860 scope.go:117] "RemoveContainer" containerID="51dd5eee6540f7d3e36a7a502a1f1549c02a41b122885a0b0f302fedfb79b957" Dec 11 08:26:00 crc kubenswrapper[4860]: I1211 08:26:00.250882 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d7724a83-b148-4b65-839a-3d85ecbbc4a0-catalog-content\") pod \"d7724a83-b148-4b65-839a-3d85ecbbc4a0\" (UID: \"d7724a83-b148-4b65-839a-3d85ecbbc4a0\") " Dec 11 08:26:00 crc kubenswrapper[4860]: I1211 08:26:00.250929 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d7724a83-b148-4b65-839a-3d85ecbbc4a0-utilities\") pod \"d7724a83-b148-4b65-839a-3d85ecbbc4a0\" (UID: \"d7724a83-b148-4b65-839a-3d85ecbbc4a0\") " Dec 11 08:26:00 crc kubenswrapper[4860]: I1211 08:26:00.251069 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-76nqn\" (UniqueName: \"kubernetes.io/projected/d7724a83-b148-4b65-839a-3d85ecbbc4a0-kube-api-access-76nqn\") pod \"d7724a83-b148-4b65-839a-3d85ecbbc4a0\" (UID: \"d7724a83-b148-4b65-839a-3d85ecbbc4a0\") " Dec 11 08:26:00 crc kubenswrapper[4860]: I1211 08:26:00.251789 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d7724a83-b148-4b65-839a-3d85ecbbc4a0-utilities" (OuterVolumeSpecName: "utilities") pod "d7724a83-b148-4b65-839a-3d85ecbbc4a0" (UID: "d7724a83-b148-4b65-839a-3d85ecbbc4a0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:26:00 crc kubenswrapper[4860]: I1211 08:26:00.257321 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7724a83-b148-4b65-839a-3d85ecbbc4a0-kube-api-access-76nqn" (OuterVolumeSpecName: "kube-api-access-76nqn") pod "d7724a83-b148-4b65-839a-3d85ecbbc4a0" (UID: "d7724a83-b148-4b65-839a-3d85ecbbc4a0"). InnerVolumeSpecName "kube-api-access-76nqn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:26:00 crc kubenswrapper[4860]: I1211 08:26:00.262955 4860 scope.go:117] "RemoveContainer" containerID="46344f92ce53b61ab1f7db7709de8b758ffb8bc3f32a47dab65465255bd1706f" Dec 11 08:26:00 crc kubenswrapper[4860]: I1211 08:26:00.297054 4860 scope.go:117] "RemoveContainer" containerID="2e5b3b5a9c5df0e9218c8ae77647ad1ee41ee767ff5e254d4adbefae795a48c6" Dec 11 08:26:00 crc kubenswrapper[4860]: E1211 08:26:00.298306 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2e5b3b5a9c5df0e9218c8ae77647ad1ee41ee767ff5e254d4adbefae795a48c6\": container with ID starting with 2e5b3b5a9c5df0e9218c8ae77647ad1ee41ee767ff5e254d4adbefae795a48c6 not found: ID does not exist" containerID="2e5b3b5a9c5df0e9218c8ae77647ad1ee41ee767ff5e254d4adbefae795a48c6" Dec 11 08:26:00 crc kubenswrapper[4860]: I1211 08:26:00.299980 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e5b3b5a9c5df0e9218c8ae77647ad1ee41ee767ff5e254d4adbefae795a48c6"} err="failed to get container status \"2e5b3b5a9c5df0e9218c8ae77647ad1ee41ee767ff5e254d4adbefae795a48c6\": rpc error: code = NotFound desc = could not find container \"2e5b3b5a9c5df0e9218c8ae77647ad1ee41ee767ff5e254d4adbefae795a48c6\": container with ID starting with 2e5b3b5a9c5df0e9218c8ae77647ad1ee41ee767ff5e254d4adbefae795a48c6 not found: ID does not exist" Dec 11 08:26:00 crc kubenswrapper[4860]: I1211 08:26:00.300138 4860 scope.go:117] "RemoveContainer" containerID="51dd5eee6540f7d3e36a7a502a1f1549c02a41b122885a0b0f302fedfb79b957" Dec 11 08:26:00 crc kubenswrapper[4860]: E1211 08:26:00.301844 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"51dd5eee6540f7d3e36a7a502a1f1549c02a41b122885a0b0f302fedfb79b957\": container with ID starting with 51dd5eee6540f7d3e36a7a502a1f1549c02a41b122885a0b0f302fedfb79b957 not found: ID does not exist" containerID="51dd5eee6540f7d3e36a7a502a1f1549c02a41b122885a0b0f302fedfb79b957" Dec 11 08:26:00 crc kubenswrapper[4860]: I1211 08:26:00.301909 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"51dd5eee6540f7d3e36a7a502a1f1549c02a41b122885a0b0f302fedfb79b957"} err="failed to get container status \"51dd5eee6540f7d3e36a7a502a1f1549c02a41b122885a0b0f302fedfb79b957\": rpc error: code = NotFound desc = could not find container \"51dd5eee6540f7d3e36a7a502a1f1549c02a41b122885a0b0f302fedfb79b957\": container with ID starting with 51dd5eee6540f7d3e36a7a502a1f1549c02a41b122885a0b0f302fedfb79b957 not found: ID does not exist" Dec 11 08:26:00 crc kubenswrapper[4860]: I1211 08:26:00.301965 4860 scope.go:117] "RemoveContainer" containerID="46344f92ce53b61ab1f7db7709de8b758ffb8bc3f32a47dab65465255bd1706f" Dec 11 08:26:00 crc kubenswrapper[4860]: E1211 08:26:00.302286 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"46344f92ce53b61ab1f7db7709de8b758ffb8bc3f32a47dab65465255bd1706f\": container with ID starting with 46344f92ce53b61ab1f7db7709de8b758ffb8bc3f32a47dab65465255bd1706f not found: ID does not exist" containerID="46344f92ce53b61ab1f7db7709de8b758ffb8bc3f32a47dab65465255bd1706f" Dec 11 08:26:00 crc kubenswrapper[4860]: I1211 08:26:00.302381 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46344f92ce53b61ab1f7db7709de8b758ffb8bc3f32a47dab65465255bd1706f"} err="failed to get container status \"46344f92ce53b61ab1f7db7709de8b758ffb8bc3f32a47dab65465255bd1706f\": rpc error: code = NotFound desc = could not find container \"46344f92ce53b61ab1f7db7709de8b758ffb8bc3f32a47dab65465255bd1706f\": container with ID starting with 46344f92ce53b61ab1f7db7709de8b758ffb8bc3f32a47dab65465255bd1706f not found: ID does not exist" Dec 11 08:26:00 crc kubenswrapper[4860]: I1211 08:26:00.304883 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d7724a83-b148-4b65-839a-3d85ecbbc4a0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d7724a83-b148-4b65-839a-3d85ecbbc4a0" (UID: "d7724a83-b148-4b65-839a-3d85ecbbc4a0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:26:00 crc kubenswrapper[4860]: I1211 08:26:00.353112 4860 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d7724a83-b148-4b65-839a-3d85ecbbc4a0-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 08:26:00 crc kubenswrapper[4860]: I1211 08:26:00.353190 4860 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d7724a83-b148-4b65-839a-3d85ecbbc4a0-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 08:26:00 crc kubenswrapper[4860]: I1211 08:26:00.353216 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-76nqn\" (UniqueName: \"kubernetes.io/projected/d7724a83-b148-4b65-839a-3d85ecbbc4a0-kube-api-access-76nqn\") on node \"crc\" DevicePath \"\"" Dec 11 08:26:00 crc kubenswrapper[4860]: I1211 08:26:00.560324 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nkzr6"] Dec 11 08:26:00 crc kubenswrapper[4860]: I1211 08:26:00.564597 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-nkzr6"] Dec 11 08:26:01 crc kubenswrapper[4860]: I1211 08:26:01.588166 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d7724a83-b148-4b65-839a-3d85ecbbc4a0" path="/var/lib/kubelet/pods/d7724a83-b148-4b65-839a-3d85ecbbc4a0/volumes" Dec 11 08:26:08 crc kubenswrapper[4860]: I1211 08:26:08.328546 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-2425r"] Dec 11 08:26:08 crc kubenswrapper[4860]: E1211 08:26:08.329369 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7724a83-b148-4b65-839a-3d85ecbbc4a0" containerName="registry-server" Dec 11 08:26:08 crc kubenswrapper[4860]: I1211 08:26:08.329387 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7724a83-b148-4b65-839a-3d85ecbbc4a0" containerName="registry-server" Dec 11 08:26:08 crc kubenswrapper[4860]: E1211 08:26:08.329397 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="929e897f-23a2-48ef-ac90-79e420fdf77f" containerName="extract-utilities" Dec 11 08:26:08 crc kubenswrapper[4860]: I1211 08:26:08.329405 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="929e897f-23a2-48ef-ac90-79e420fdf77f" containerName="extract-utilities" Dec 11 08:26:08 crc kubenswrapper[4860]: E1211 08:26:08.329422 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="929e897f-23a2-48ef-ac90-79e420fdf77f" containerName="extract-content" Dec 11 08:26:08 crc kubenswrapper[4860]: I1211 08:26:08.329429 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="929e897f-23a2-48ef-ac90-79e420fdf77f" containerName="extract-content" Dec 11 08:26:08 crc kubenswrapper[4860]: E1211 08:26:08.329438 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="929e897f-23a2-48ef-ac90-79e420fdf77f" containerName="registry-server" Dec 11 08:26:08 crc kubenswrapper[4860]: I1211 08:26:08.329444 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="929e897f-23a2-48ef-ac90-79e420fdf77f" containerName="registry-server" Dec 11 08:26:08 crc kubenswrapper[4860]: E1211 08:26:08.329452 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7724a83-b148-4b65-839a-3d85ecbbc4a0" containerName="extract-utilities" Dec 11 08:26:08 crc kubenswrapper[4860]: I1211 08:26:08.329458 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7724a83-b148-4b65-839a-3d85ecbbc4a0" containerName="extract-utilities" Dec 11 08:26:08 crc kubenswrapper[4860]: E1211 08:26:08.329469 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7724a83-b148-4b65-839a-3d85ecbbc4a0" containerName="extract-content" Dec 11 08:26:08 crc kubenswrapper[4860]: I1211 08:26:08.329475 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7724a83-b148-4b65-839a-3d85ecbbc4a0" containerName="extract-content" Dec 11 08:26:08 crc kubenswrapper[4860]: I1211 08:26:08.329594 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7724a83-b148-4b65-839a-3d85ecbbc4a0" containerName="registry-server" Dec 11 08:26:08 crc kubenswrapper[4860]: I1211 08:26:08.329608 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="929e897f-23a2-48ef-ac90-79e420fdf77f" containerName="registry-server" Dec 11 08:26:08 crc kubenswrapper[4860]: I1211 08:26:08.330601 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2425r" Dec 11 08:26:08 crc kubenswrapper[4860]: I1211 08:26:08.358627 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2425r"] Dec 11 08:26:08 crc kubenswrapper[4860]: I1211 08:26:08.380757 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/16fb4ab2-190c-4a33-8ba1-f9cd5edaaff3-catalog-content\") pod \"redhat-marketplace-2425r\" (UID: \"16fb4ab2-190c-4a33-8ba1-f9cd5edaaff3\") " pod="openshift-marketplace/redhat-marketplace-2425r" Dec 11 08:26:08 crc kubenswrapper[4860]: I1211 08:26:08.380855 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pfk4r\" (UniqueName: \"kubernetes.io/projected/16fb4ab2-190c-4a33-8ba1-f9cd5edaaff3-kube-api-access-pfk4r\") pod \"redhat-marketplace-2425r\" (UID: \"16fb4ab2-190c-4a33-8ba1-f9cd5edaaff3\") " pod="openshift-marketplace/redhat-marketplace-2425r" Dec 11 08:26:08 crc kubenswrapper[4860]: I1211 08:26:08.381183 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/16fb4ab2-190c-4a33-8ba1-f9cd5edaaff3-utilities\") pod \"redhat-marketplace-2425r\" (UID: \"16fb4ab2-190c-4a33-8ba1-f9cd5edaaff3\") " pod="openshift-marketplace/redhat-marketplace-2425r" Dec 11 08:26:08 crc kubenswrapper[4860]: I1211 08:26:08.482846 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/16fb4ab2-190c-4a33-8ba1-f9cd5edaaff3-utilities\") pod \"redhat-marketplace-2425r\" (UID: \"16fb4ab2-190c-4a33-8ba1-f9cd5edaaff3\") " pod="openshift-marketplace/redhat-marketplace-2425r" Dec 11 08:26:08 crc kubenswrapper[4860]: I1211 08:26:08.483009 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/16fb4ab2-190c-4a33-8ba1-f9cd5edaaff3-catalog-content\") pod \"redhat-marketplace-2425r\" (UID: \"16fb4ab2-190c-4a33-8ba1-f9cd5edaaff3\") " pod="openshift-marketplace/redhat-marketplace-2425r" Dec 11 08:26:08 crc kubenswrapper[4860]: I1211 08:26:08.483072 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pfk4r\" (UniqueName: \"kubernetes.io/projected/16fb4ab2-190c-4a33-8ba1-f9cd5edaaff3-kube-api-access-pfk4r\") pod \"redhat-marketplace-2425r\" (UID: \"16fb4ab2-190c-4a33-8ba1-f9cd5edaaff3\") " pod="openshift-marketplace/redhat-marketplace-2425r" Dec 11 08:26:08 crc kubenswrapper[4860]: I1211 08:26:08.483849 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/16fb4ab2-190c-4a33-8ba1-f9cd5edaaff3-utilities\") pod \"redhat-marketplace-2425r\" (UID: \"16fb4ab2-190c-4a33-8ba1-f9cd5edaaff3\") " pod="openshift-marketplace/redhat-marketplace-2425r" Dec 11 08:26:08 crc kubenswrapper[4860]: I1211 08:26:08.483975 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/16fb4ab2-190c-4a33-8ba1-f9cd5edaaff3-catalog-content\") pod \"redhat-marketplace-2425r\" (UID: \"16fb4ab2-190c-4a33-8ba1-f9cd5edaaff3\") " pod="openshift-marketplace/redhat-marketplace-2425r" Dec 11 08:26:08 crc kubenswrapper[4860]: I1211 08:26:08.507818 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pfk4r\" (UniqueName: \"kubernetes.io/projected/16fb4ab2-190c-4a33-8ba1-f9cd5edaaff3-kube-api-access-pfk4r\") pod \"redhat-marketplace-2425r\" (UID: \"16fb4ab2-190c-4a33-8ba1-f9cd5edaaff3\") " pod="openshift-marketplace/redhat-marketplace-2425r" Dec 11 08:26:08 crc kubenswrapper[4860]: I1211 08:26:08.650726 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2425r" Dec 11 08:26:08 crc kubenswrapper[4860]: I1211 08:26:08.938089 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2425r"] Dec 11 08:26:09 crc kubenswrapper[4860]: I1211 08:26:09.285337 4860 generic.go:334] "Generic (PLEG): container finished" podID="16fb4ab2-190c-4a33-8ba1-f9cd5edaaff3" containerID="6fb1acd3dcd395508a11361e36797af7415d63f52aac61bc11319cf68de84664" exitCode=0 Dec 11 08:26:09 crc kubenswrapper[4860]: I1211 08:26:09.285401 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2425r" event={"ID":"16fb4ab2-190c-4a33-8ba1-f9cd5edaaff3","Type":"ContainerDied","Data":"6fb1acd3dcd395508a11361e36797af7415d63f52aac61bc11319cf68de84664"} Dec 11 08:26:09 crc kubenswrapper[4860]: I1211 08:26:09.285438 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2425r" event={"ID":"16fb4ab2-190c-4a33-8ba1-f9cd5edaaff3","Type":"ContainerStarted","Data":"e61abd02a3bb306d6cc6fc6c51606cdc3d963e90dda8cd22af18d5f7fb63fdfb"} Dec 11 08:26:11 crc kubenswrapper[4860]: I1211 08:26:11.303413 4860 generic.go:334] "Generic (PLEG): container finished" podID="16fb4ab2-190c-4a33-8ba1-f9cd5edaaff3" containerID="0f5fda4090970b2ecd060ce1226877e892c204b6e116470c09d46001ffb4cace" exitCode=0 Dec 11 08:26:11 crc kubenswrapper[4860]: I1211 08:26:11.303506 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2425r" event={"ID":"16fb4ab2-190c-4a33-8ba1-f9cd5edaaff3","Type":"ContainerDied","Data":"0f5fda4090970b2ecd060ce1226877e892c204b6e116470c09d46001ffb4cace"} Dec 11 08:26:12 crc kubenswrapper[4860]: I1211 08:26:12.321129 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2425r" event={"ID":"16fb4ab2-190c-4a33-8ba1-f9cd5edaaff3","Type":"ContainerStarted","Data":"e35bb40199af54a9bbd197edd80583cfb35a628ea0965a63590700958977af63"} Dec 11 08:26:12 crc kubenswrapper[4860]: I1211 08:26:12.344848 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-2425r" podStartSLOduration=1.668028979 podStartE2EDuration="4.344823352s" podCreationTimestamp="2025-12-11 08:26:08 +0000 UTC" firstStartedPulling="2025-12-11 08:26:09.287466207 +0000 UTC m=+902.015985252" lastFinishedPulling="2025-12-11 08:26:11.96426057 +0000 UTC m=+904.692779625" observedRunningTime="2025-12-11 08:26:12.343264024 +0000 UTC m=+905.071783079" watchObservedRunningTime="2025-12-11 08:26:12.344823352 +0000 UTC m=+905.073342407" Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.017514 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-bshq2"] Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.019899 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-bshq2" Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.023698 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-vlc6d" Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.029197 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6c677c69b-zrrf6"] Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.030578 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-zrrf6" Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.032610 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-wgq2c" Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.043112 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-bshq2"] Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.060586 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-697fb699cf-6t74n"] Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.061994 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-6t74n" Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.065206 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6c677c69b-zrrf6"] Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.066030 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-dl7kp" Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.101491 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-5697bb5779-kgzwt"] Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.102874 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-kgzwt" Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.106111 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-6mbk7" Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.107278 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-697fb699cf-6t74n"] Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.114390 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-5697bb5779-kgzwt"] Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.143963 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-lrhhk"] Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.145577 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-lrhhk" Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.148529 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-f97g6" Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.154263 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rgnth\" (UniqueName: \"kubernetes.io/projected/64f09777-fca3-412f-98f8-5cd189cb9fbd-kube-api-access-rgnth\") pod \"barbican-operator-controller-manager-7d9dfd778-bshq2\" (UID: \"64f09777-fca3-412f-98f8-5cd189cb9fbd\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-bshq2" Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.154358 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zf7j6\" (UniqueName: \"kubernetes.io/projected/323ad77f-a0d6-461d-b483-2ffd74ebb9b7-kube-api-access-zf7j6\") pod \"cinder-operator-controller-manager-6c677c69b-zrrf6\" (UID: \"323ad77f-a0d6-461d-b483-2ffd74ebb9b7\") " pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-zrrf6" Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.154461 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gbxtz\" (UniqueName: \"kubernetes.io/projected/a3bf483c-6438-4b7e-abb8-447b9b146644-kube-api-access-gbxtz\") pod \"designate-operator-controller-manager-697fb699cf-6t74n\" (UID: \"a3bf483c-6438-4b7e-abb8-447b9b146644\") " pod="openstack-operators/designate-operator-controller-manager-697fb699cf-6t74n" Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.167817 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-lrhhk"] Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.196022 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-65fv4"] Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.197726 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-65fv4" Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.202301 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-78d48bff9d-sq5gd"] Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.207720 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-flhq5" Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.209012 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-sq5gd" Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.215120 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-dmhff" Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.215955 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-65fv4"] Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.217927 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.235840 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-967d97867-5wvld"] Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.239062 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-967d97867-5wvld" Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.246372 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-vqwdq" Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.253376 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-78d48bff9d-sq5gd"] Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.258872 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gbxtz\" (UniqueName: \"kubernetes.io/projected/a3bf483c-6438-4b7e-abb8-447b9b146644-kube-api-access-gbxtz\") pod \"designate-operator-controller-manager-697fb699cf-6t74n\" (UID: \"a3bf483c-6438-4b7e-abb8-447b9b146644\") " pod="openstack-operators/designate-operator-controller-manager-697fb699cf-6t74n" Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.258999 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rgnth\" (UniqueName: \"kubernetes.io/projected/64f09777-fca3-412f-98f8-5cd189cb9fbd-kube-api-access-rgnth\") pod \"barbican-operator-controller-manager-7d9dfd778-bshq2\" (UID: \"64f09777-fca3-412f-98f8-5cd189cb9fbd\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-bshq2" Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.259048 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xl5z4\" (UniqueName: \"kubernetes.io/projected/e00c360d-20a4-4e2e-81ea-e4ffa1bc62fc-kube-api-access-xl5z4\") pod \"glance-operator-controller-manager-5697bb5779-kgzwt\" (UID: \"e00c360d-20a4-4e2e-81ea-e4ffa1bc62fc\") " pod="openstack-operators/glance-operator-controller-manager-5697bb5779-kgzwt" Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.259085 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6gzml\" (UniqueName: \"kubernetes.io/projected/505d980f-cfcb-42b2-876e-a4730abe7ea6-kube-api-access-6gzml\") pod \"heat-operator-controller-manager-5f64f6f8bb-lrhhk\" (UID: \"505d980f-cfcb-42b2-876e-a4730abe7ea6\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-lrhhk" Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.259113 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zf7j6\" (UniqueName: \"kubernetes.io/projected/323ad77f-a0d6-461d-b483-2ffd74ebb9b7-kube-api-access-zf7j6\") pod \"cinder-operator-controller-manager-6c677c69b-zrrf6\" (UID: \"323ad77f-a0d6-461d-b483-2ffd74ebb9b7\") " pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-zrrf6" Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.265411 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-967d97867-5wvld"] Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.294226 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-5b5fd79c9c-lpm8h"] Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.296751 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-lpm8h" Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.303273 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-2wpmd" Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.307965 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-6bpb9"] Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.310854 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-6bpb9" Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.316985 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-5b5fd79c9c-lpm8h"] Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.320720 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-2jvz5" Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.324311 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zf7j6\" (UniqueName: \"kubernetes.io/projected/323ad77f-a0d6-461d-b483-2ffd74ebb9b7-kube-api-access-zf7j6\") pod \"cinder-operator-controller-manager-6c677c69b-zrrf6\" (UID: \"323ad77f-a0d6-461d-b483-2ffd74ebb9b7\") " pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-zrrf6" Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.347385 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gbxtz\" (UniqueName: \"kubernetes.io/projected/a3bf483c-6438-4b7e-abb8-447b9b146644-kube-api-access-gbxtz\") pod \"designate-operator-controller-manager-697fb699cf-6t74n\" (UID: \"a3bf483c-6438-4b7e-abb8-447b9b146644\") " pod="openstack-operators/designate-operator-controller-manager-697fb699cf-6t74n" Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.353850 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-6bpb9"] Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.374224 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fq2rx\" (UniqueName: \"kubernetes.io/projected/91bb3ff9-fd19-4c79-bac4-621cb114c783-kube-api-access-fq2rx\") pod \"ironic-operator-controller-manager-967d97867-5wvld\" (UID: \"91bb3ff9-fd19-4c79-bac4-621cb114c783\") " pod="openstack-operators/ironic-operator-controller-manager-967d97867-5wvld" Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.374940 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5z78q\" (UniqueName: \"kubernetes.io/projected/bb231e77-e19b-41ed-bf7a-ae354e1089ec-kube-api-access-5z78q\") pod \"horizon-operator-controller-manager-68c6d99b8f-65fv4\" (UID: \"bb231e77-e19b-41ed-bf7a-ae354e1089ec\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-65fv4" Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.375122 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xl5z4\" (UniqueName: \"kubernetes.io/projected/e00c360d-20a4-4e2e-81ea-e4ffa1bc62fc-kube-api-access-xl5z4\") pod \"glance-operator-controller-manager-5697bb5779-kgzwt\" (UID: \"e00c360d-20a4-4e2e-81ea-e4ffa1bc62fc\") " pod="openstack-operators/glance-operator-controller-manager-5697bb5779-kgzwt" Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.375179 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7f1d9cd2-1e0c-4c4d-86a6-b91a38eadba2-cert\") pod \"infra-operator-controller-manager-78d48bff9d-sq5gd\" (UID: \"7f1d9cd2-1e0c-4c4d-86a6-b91a38eadba2\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-sq5gd" Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.375206 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6gzml\" (UniqueName: \"kubernetes.io/projected/505d980f-cfcb-42b2-876e-a4730abe7ea6-kube-api-access-6gzml\") pod \"heat-operator-controller-manager-5f64f6f8bb-lrhhk\" (UID: \"505d980f-cfcb-42b2-876e-a4730abe7ea6\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-lrhhk" Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.375246 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n5wf8\" (UniqueName: \"kubernetes.io/projected/7f1d9cd2-1e0c-4c4d-86a6-b91a38eadba2-kube-api-access-n5wf8\") pod \"infra-operator-controller-manager-78d48bff9d-sq5gd\" (UID: \"7f1d9cd2-1e0c-4c4d-86a6-b91a38eadba2\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-sq5gd" Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.379632 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rgnth\" (UniqueName: \"kubernetes.io/projected/64f09777-fca3-412f-98f8-5cd189cb9fbd-kube-api-access-rgnth\") pod \"barbican-operator-controller-manager-7d9dfd778-bshq2\" (UID: \"64f09777-fca3-412f-98f8-5cd189cb9fbd\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-bshq2" Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.385174 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-79c8c4686c-4tw58"] Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.392703 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-4tw58" Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.443319 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-9p67l" Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.446260 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-6t74n" Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.446978 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-zrrf6" Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.474810 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xl5z4\" (UniqueName: \"kubernetes.io/projected/e00c360d-20a4-4e2e-81ea-e4ffa1bc62fc-kube-api-access-xl5z4\") pod \"glance-operator-controller-manager-5697bb5779-kgzwt\" (UID: \"e00c360d-20a4-4e2e-81ea-e4ffa1bc62fc\") " pod="openstack-operators/glance-operator-controller-manager-5697bb5779-kgzwt" Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.476742 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fq2rx\" (UniqueName: \"kubernetes.io/projected/91bb3ff9-fd19-4c79-bac4-621cb114c783-kube-api-access-fq2rx\") pod \"ironic-operator-controller-manager-967d97867-5wvld\" (UID: \"91bb3ff9-fd19-4c79-bac4-621cb114c783\") " pod="openstack-operators/ironic-operator-controller-manager-967d97867-5wvld" Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.476867 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5z78q\" (UniqueName: \"kubernetes.io/projected/bb231e77-e19b-41ed-bf7a-ae354e1089ec-kube-api-access-5z78q\") pod \"horizon-operator-controller-manager-68c6d99b8f-65fv4\" (UID: \"bb231e77-e19b-41ed-bf7a-ae354e1089ec\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-65fv4" Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.476929 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5ldm2\" (UniqueName: \"kubernetes.io/projected/d5e0079f-f500-407f-b06f-a610a19c741b-kube-api-access-5ldm2\") pod \"keystone-operator-controller-manager-7765d96ddf-6bpb9\" (UID: \"d5e0079f-f500-407f-b06f-a610a19c741b\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-6bpb9" Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.476974 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zgv9m\" (UniqueName: \"kubernetes.io/projected/5a589b13-7f29-40ae-9804-7c5c2bc3d745-kube-api-access-zgv9m\") pod \"mariadb-operator-controller-manager-79c8c4686c-4tw58\" (UID: \"5a589b13-7f29-40ae-9804-7c5c2bc3d745\") " pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-4tw58" Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.477027 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7f1d9cd2-1e0c-4c4d-86a6-b91a38eadba2-cert\") pod \"infra-operator-controller-manager-78d48bff9d-sq5gd\" (UID: \"7f1d9cd2-1e0c-4c4d-86a6-b91a38eadba2\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-sq5gd" Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.477070 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n5wf8\" (UniqueName: \"kubernetes.io/projected/7f1d9cd2-1e0c-4c4d-86a6-b91a38eadba2-kube-api-access-n5wf8\") pod \"infra-operator-controller-manager-78d48bff9d-sq5gd\" (UID: \"7f1d9cd2-1e0c-4c4d-86a6-b91a38eadba2\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-sq5gd" Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.477096 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rb8wj\" (UniqueName: \"kubernetes.io/projected/aae650b9-dcbc-482b-b49f-0a45651bb4b4-kube-api-access-rb8wj\") pod \"manila-operator-controller-manager-5b5fd79c9c-lpm8h\" (UID: \"aae650b9-dcbc-482b-b49f-0a45651bb4b4\") " pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-lpm8h" Dec 11 08:26:16 crc kubenswrapper[4860]: E1211 08:26:16.477870 4860 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 11 08:26:16 crc kubenswrapper[4860]: E1211 08:26:16.477945 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7f1d9cd2-1e0c-4c4d-86a6-b91a38eadba2-cert podName:7f1d9cd2-1e0c-4c4d-86a6-b91a38eadba2 nodeName:}" failed. No retries permitted until 2025-12-11 08:26:16.97792046 +0000 UTC m=+909.706439515 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/7f1d9cd2-1e0c-4c4d-86a6-b91a38eadba2-cert") pod "infra-operator-controller-manager-78d48bff9d-sq5gd" (UID: "7f1d9cd2-1e0c-4c4d-86a6-b91a38eadba2") : secret "infra-operator-webhook-server-cert" not found Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.488433 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-79c8c4686c-4tw58"] Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.503485 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6gzml\" (UniqueName: \"kubernetes.io/projected/505d980f-cfcb-42b2-876e-a4730abe7ea6-kube-api-access-6gzml\") pod \"heat-operator-controller-manager-5f64f6f8bb-lrhhk\" (UID: \"505d980f-cfcb-42b2-876e-a4730abe7ea6\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-lrhhk" Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.517748 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-ht7db"] Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.519319 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-ht7db" Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.527507 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-4ns6s" Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.541296 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5z78q\" (UniqueName: \"kubernetes.io/projected/bb231e77-e19b-41ed-bf7a-ae354e1089ec-kube-api-access-5z78q\") pod \"horizon-operator-controller-manager-68c6d99b8f-65fv4\" (UID: \"bb231e77-e19b-41ed-bf7a-ae354e1089ec\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-65fv4" Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.553353 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n5wf8\" (UniqueName: \"kubernetes.io/projected/7f1d9cd2-1e0c-4c4d-86a6-b91a38eadba2-kube-api-access-n5wf8\") pod \"infra-operator-controller-manager-78d48bff9d-sq5gd\" (UID: \"7f1d9cd2-1e0c-4c4d-86a6-b91a38eadba2\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-sq5gd" Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.557545 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fq2rx\" (UniqueName: \"kubernetes.io/projected/91bb3ff9-fd19-4c79-bac4-621cb114c783-kube-api-access-fq2rx\") pod \"ironic-operator-controller-manager-967d97867-5wvld\" (UID: \"91bb3ff9-fd19-4c79-bac4-621cb114c783\") " pod="openstack-operators/ironic-operator-controller-manager-967d97867-5wvld" Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.557670 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-zrhjm"] Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.559137 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-zrhjm" Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.576433 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-6lzb8" Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.577976 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-967d97867-5wvld" Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.578215 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rb8wj\" (UniqueName: \"kubernetes.io/projected/aae650b9-dcbc-482b-b49f-0a45651bb4b4-kube-api-access-rb8wj\") pod \"manila-operator-controller-manager-5b5fd79c9c-lpm8h\" (UID: \"aae650b9-dcbc-482b-b49f-0a45651bb4b4\") " pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-lpm8h" Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.578313 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5ldm2\" (UniqueName: \"kubernetes.io/projected/d5e0079f-f500-407f-b06f-a610a19c741b-kube-api-access-5ldm2\") pod \"keystone-operator-controller-manager-7765d96ddf-6bpb9\" (UID: \"d5e0079f-f500-407f-b06f-a610a19c741b\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-6bpb9" Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.578348 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zgv9m\" (UniqueName: \"kubernetes.io/projected/5a589b13-7f29-40ae-9804-7c5c2bc3d745-kube-api-access-zgv9m\") pod \"mariadb-operator-controller-manager-79c8c4686c-4tw58\" (UID: \"5a589b13-7f29-40ae-9804-7c5c2bc3d745\") " pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-4tw58" Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.594857 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-ht7db"] Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.634283 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zgv9m\" (UniqueName: \"kubernetes.io/projected/5a589b13-7f29-40ae-9804-7c5c2bc3d745-kube-api-access-zgv9m\") pod \"mariadb-operator-controller-manager-79c8c4686c-4tw58\" (UID: \"5a589b13-7f29-40ae-9804-7c5c2bc3d745\") " pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-4tw58" Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.634391 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-zrhjm"] Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.636187 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5ldm2\" (UniqueName: \"kubernetes.io/projected/d5e0079f-f500-407f-b06f-a610a19c741b-kube-api-access-5ldm2\") pod \"keystone-operator-controller-manager-7765d96ddf-6bpb9\" (UID: \"d5e0079f-f500-407f-b06f-a610a19c741b\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-6bpb9" Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.647119 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-6bpb9" Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.666047 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-bshq2" Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.667518 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rb8wj\" (UniqueName: \"kubernetes.io/projected/aae650b9-dcbc-482b-b49f-0a45651bb4b4-kube-api-access-rb8wj\") pod \"manila-operator-controller-manager-5b5fd79c9c-lpm8h\" (UID: \"aae650b9-dcbc-482b-b49f-0a45651bb4b4\") " pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-lpm8h" Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.680996 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6rvhf\" (UniqueName: \"kubernetes.io/projected/dc79abc2-7b6c-46c4-b9b4-a9c76696a8b2-kube-api-access-6rvhf\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-ht7db\" (UID: \"dc79abc2-7b6c-46c4-b9b4-a9c76696a8b2\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-ht7db" Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.681593 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cs4kf\" (UniqueName: \"kubernetes.io/projected/5aff136d-8f48-4cb7-8ddc-72ce966607b7-kube-api-access-cs4kf\") pod \"nova-operator-controller-manager-697bc559fc-zrhjm\" (UID: \"5aff136d-8f48-4cb7-8ddc-72ce966607b7\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-zrhjm" Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.683066 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-4tw58" Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.721979 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-bvs87"] Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.724077 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-bvs87" Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.731505 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-g6gt6" Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.734962 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-kgzwt" Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.758925 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fgp9c8"] Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.760420 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fgp9c8" Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.771171 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-5qqkz" Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.781845 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.784472 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cs4kf\" (UniqueName: \"kubernetes.io/projected/5aff136d-8f48-4cb7-8ddc-72ce966607b7-kube-api-access-cs4kf\") pod \"nova-operator-controller-manager-697bc559fc-zrhjm\" (UID: \"5aff136d-8f48-4cb7-8ddc-72ce966607b7\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-zrhjm" Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.784662 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6rvhf\" (UniqueName: \"kubernetes.io/projected/dc79abc2-7b6c-46c4-b9b4-a9c76696a8b2-kube-api-access-6rvhf\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-ht7db\" (UID: \"dc79abc2-7b6c-46c4-b9b4-a9c76696a8b2\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-ht7db" Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.791563 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-lrhhk" Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.835192 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-65fv4" Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.854085 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cs4kf\" (UniqueName: \"kubernetes.io/projected/5aff136d-8f48-4cb7-8ddc-72ce966607b7-kube-api-access-cs4kf\") pod \"nova-operator-controller-manager-697bc559fc-zrhjm\" (UID: \"5aff136d-8f48-4cb7-8ddc-72ce966607b7\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-zrhjm" Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.868812 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-2qmxt"] Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.870263 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-2qmxt" Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.877003 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-qppr5" Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.877732 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-lpm8h" Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.879013 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6rvhf\" (UniqueName: \"kubernetes.io/projected/dc79abc2-7b6c-46c4-b9b4-a9c76696a8b2-kube-api-access-6rvhf\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-ht7db\" (UID: \"dc79abc2-7b6c-46c4-b9b4-a9c76696a8b2\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-ht7db" Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.884794 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-bvs87"] Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.896399 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fgp9c8"] Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.896516 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/04044c38-9aa4-4439-bad4-cf687e00ed5a-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879fgp9c8\" (UID: \"04044c38-9aa4-4439-bad4-cf687e00ed5a\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fgp9c8" Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.898392 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-ht7db" Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.899369 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wq9zg\" (UniqueName: \"kubernetes.io/projected/689923af-60de-49e9-bbed-bebaa63adc4b-kube-api-access-wq9zg\") pod \"octavia-operator-controller-manager-998648c74-bvs87\" (UID: \"689923af-60de-49e9-bbed-bebaa63adc4b\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-bvs87" Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.899517 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t4jt4\" (UniqueName: \"kubernetes.io/projected/04044c38-9aa4-4439-bad4-cf687e00ed5a-kube-api-access-t4jt4\") pod \"openstack-baremetal-operator-controller-manager-84b575879fgp9c8\" (UID: \"04044c38-9aa4-4439-bad4-cf687e00ed5a\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fgp9c8" Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.905419 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-nvt56"] Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.910492 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-zrhjm" Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.917749 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-nvt56" Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.918492 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-nvt56"] Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.920198 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-zbfrm" Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.927584 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-2qmxt"] Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.943742 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-58d5ff84df-hrpn8"] Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.945223 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-hrpn8" Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.950695 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-r5nfd" Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.956409 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-9d58d64bc-p7dkp"] Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.961221 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-p7dkp" Dec 11 08:26:16 crc kubenswrapper[4860]: I1211 08:26:16.964800 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-5ppht" Dec 11 08:26:17 crc kubenswrapper[4860]: I1211 08:26:17.000908 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7f1d9cd2-1e0c-4c4d-86a6-b91a38eadba2-cert\") pod \"infra-operator-controller-manager-78d48bff9d-sq5gd\" (UID: \"7f1d9cd2-1e0c-4c4d-86a6-b91a38eadba2\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-sq5gd" Dec 11 08:26:17 crc kubenswrapper[4860]: I1211 08:26:17.000993 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jdhtm\" (UniqueName: \"kubernetes.io/projected/99f7e97f-7cea-4785-b95e-133adf238ac3-kube-api-access-jdhtm\") pod \"ovn-operator-controller-manager-b6456fdb6-nvt56\" (UID: \"99f7e97f-7cea-4785-b95e-133adf238ac3\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-nvt56" Dec 11 08:26:17 crc kubenswrapper[4860]: I1211 08:26:17.001061 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t4jt4\" (UniqueName: \"kubernetes.io/projected/04044c38-9aa4-4439-bad4-cf687e00ed5a-kube-api-access-t4jt4\") pod \"openstack-baremetal-operator-controller-manager-84b575879fgp9c8\" (UID: \"04044c38-9aa4-4439-bad4-cf687e00ed5a\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fgp9c8" Dec 11 08:26:17 crc kubenswrapper[4860]: I1211 08:26:17.001095 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/04044c38-9aa4-4439-bad4-cf687e00ed5a-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879fgp9c8\" (UID: \"04044c38-9aa4-4439-bad4-cf687e00ed5a\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fgp9c8" Dec 11 08:26:17 crc kubenswrapper[4860]: I1211 08:26:17.001149 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wq9zg\" (UniqueName: \"kubernetes.io/projected/689923af-60de-49e9-bbed-bebaa63adc4b-kube-api-access-wq9zg\") pod \"octavia-operator-controller-manager-998648c74-bvs87\" (UID: \"689923af-60de-49e9-bbed-bebaa63adc4b\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-bvs87" Dec 11 08:26:17 crc kubenswrapper[4860]: I1211 08:26:17.001173 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nnldp\" (UniqueName: \"kubernetes.io/projected/a52f8df6-7c28-4776-aca5-bd8b47e82fe8-kube-api-access-nnldp\") pod \"placement-operator-controller-manager-78f8948974-2qmxt\" (UID: \"a52f8df6-7c28-4776-aca5-bd8b47e82fe8\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-2qmxt" Dec 11 08:26:17 crc kubenswrapper[4860]: E1211 08:26:17.001365 4860 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 11 08:26:17 crc kubenswrapper[4860]: E1211 08:26:17.001424 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7f1d9cd2-1e0c-4c4d-86a6-b91a38eadba2-cert podName:7f1d9cd2-1e0c-4c4d-86a6-b91a38eadba2 nodeName:}" failed. No retries permitted until 2025-12-11 08:26:18.001401122 +0000 UTC m=+910.729920177 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/7f1d9cd2-1e0c-4c4d-86a6-b91a38eadba2-cert") pod "infra-operator-controller-manager-78d48bff9d-sq5gd" (UID: "7f1d9cd2-1e0c-4c4d-86a6-b91a38eadba2") : secret "infra-operator-webhook-server-cert" not found Dec 11 08:26:17 crc kubenswrapper[4860]: E1211 08:26:17.002414 4860 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 11 08:26:17 crc kubenswrapper[4860]: E1211 08:26:17.002516 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/04044c38-9aa4-4439-bad4-cf687e00ed5a-cert podName:04044c38-9aa4-4439-bad4-cf687e00ed5a nodeName:}" failed. No retries permitted until 2025-12-11 08:26:17.502487206 +0000 UTC m=+910.231006261 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/04044c38-9aa4-4439-bad4-cf687e00ed5a-cert") pod "openstack-baremetal-operator-controller-manager-84b575879fgp9c8" (UID: "04044c38-9aa4-4439-bad4-cf687e00ed5a") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 11 08:26:17 crc kubenswrapper[4860]: I1211 08:26:17.034824 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t4jt4\" (UniqueName: \"kubernetes.io/projected/04044c38-9aa4-4439-bad4-cf687e00ed5a-kube-api-access-t4jt4\") pod \"openstack-baremetal-operator-controller-manager-84b575879fgp9c8\" (UID: \"04044c38-9aa4-4439-bad4-cf687e00ed5a\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fgp9c8" Dec 11 08:26:17 crc kubenswrapper[4860]: I1211 08:26:17.035287 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wq9zg\" (UniqueName: \"kubernetes.io/projected/689923af-60de-49e9-bbed-bebaa63adc4b-kube-api-access-wq9zg\") pod \"octavia-operator-controller-manager-998648c74-bvs87\" (UID: \"689923af-60de-49e9-bbed-bebaa63adc4b\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-bvs87" Dec 11 08:26:17 crc kubenswrapper[4860]: I1211 08:26:17.096346 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-fh5vc"] Dec 11 08:26:17 crc kubenswrapper[4860]: I1211 08:26:17.100314 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-fh5vc" Dec 11 08:26:17 crc kubenswrapper[4860]: I1211 08:26:17.105525 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ldx9s\" (UniqueName: \"kubernetes.io/projected/211d9c9c-f584-43ca-8db1-7b81f6307c21-kube-api-access-ldx9s\") pod \"swift-operator-controller-manager-9d58d64bc-p7dkp\" (UID: \"211d9c9c-f584-43ca-8db1-7b81f6307c21\") " pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-p7dkp" Dec 11 08:26:17 crc kubenswrapper[4860]: I1211 08:26:17.105622 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5vl5s\" (UniqueName: \"kubernetes.io/projected/e36550af-98a6-49b3-9769-bd52d0da2838-kube-api-access-5vl5s\") pod \"telemetry-operator-controller-manager-58d5ff84df-hrpn8\" (UID: \"e36550af-98a6-49b3-9769-bd52d0da2838\") " pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-hrpn8" Dec 11 08:26:17 crc kubenswrapper[4860]: I1211 08:26:17.105715 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nnldp\" (UniqueName: \"kubernetes.io/projected/a52f8df6-7c28-4776-aca5-bd8b47e82fe8-kube-api-access-nnldp\") pod \"placement-operator-controller-manager-78f8948974-2qmxt\" (UID: \"a52f8df6-7c28-4776-aca5-bd8b47e82fe8\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-2qmxt" Dec 11 08:26:17 crc kubenswrapper[4860]: I1211 08:26:17.105798 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jdhtm\" (UniqueName: \"kubernetes.io/projected/99f7e97f-7cea-4785-b95e-133adf238ac3-kube-api-access-jdhtm\") pod \"ovn-operator-controller-manager-b6456fdb6-nvt56\" (UID: \"99f7e97f-7cea-4785-b95e-133adf238ac3\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-nvt56" Dec 11 08:26:17 crc kubenswrapper[4860]: I1211 08:26:17.108555 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-ft7pv" Dec 11 08:26:17 crc kubenswrapper[4860]: I1211 08:26:17.141800 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-58d5ff84df-hrpn8"] Dec 11 08:26:17 crc kubenswrapper[4860]: I1211 08:26:17.152588 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nnldp\" (UniqueName: \"kubernetes.io/projected/a52f8df6-7c28-4776-aca5-bd8b47e82fe8-kube-api-access-nnldp\") pod \"placement-operator-controller-manager-78f8948974-2qmxt\" (UID: \"a52f8df6-7c28-4776-aca5-bd8b47e82fe8\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-2qmxt" Dec 11 08:26:17 crc kubenswrapper[4860]: I1211 08:26:17.164593 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jdhtm\" (UniqueName: \"kubernetes.io/projected/99f7e97f-7cea-4785-b95e-133adf238ac3-kube-api-access-jdhtm\") pod \"ovn-operator-controller-manager-b6456fdb6-nvt56\" (UID: \"99f7e97f-7cea-4785-b95e-133adf238ac3\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-nvt56" Dec 11 08:26:17 crc kubenswrapper[4860]: I1211 08:26:17.175411 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-9d58d64bc-p7dkp"] Dec 11 08:26:17 crc kubenswrapper[4860]: I1211 08:26:17.207658 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b85p7\" (UniqueName: \"kubernetes.io/projected/b6c38bbc-80d0-449a-aab0-291078361ebd-kube-api-access-b85p7\") pod \"test-operator-controller-manager-5854674fcc-fh5vc\" (UID: \"b6c38bbc-80d0-449a-aab0-291078361ebd\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-fh5vc" Dec 11 08:26:17 crc kubenswrapper[4860]: I1211 08:26:17.208230 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ldx9s\" (UniqueName: \"kubernetes.io/projected/211d9c9c-f584-43ca-8db1-7b81f6307c21-kube-api-access-ldx9s\") pod \"swift-operator-controller-manager-9d58d64bc-p7dkp\" (UID: \"211d9c9c-f584-43ca-8db1-7b81f6307c21\") " pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-p7dkp" Dec 11 08:26:17 crc kubenswrapper[4860]: I1211 08:26:17.208295 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5vl5s\" (UniqueName: \"kubernetes.io/projected/e36550af-98a6-49b3-9769-bd52d0da2838-kube-api-access-5vl5s\") pod \"telemetry-operator-controller-manager-58d5ff84df-hrpn8\" (UID: \"e36550af-98a6-49b3-9769-bd52d0da2838\") " pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-hrpn8" Dec 11 08:26:17 crc kubenswrapper[4860]: I1211 08:26:17.237959 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5vl5s\" (UniqueName: \"kubernetes.io/projected/e36550af-98a6-49b3-9769-bd52d0da2838-kube-api-access-5vl5s\") pod \"telemetry-operator-controller-manager-58d5ff84df-hrpn8\" (UID: \"e36550af-98a6-49b3-9769-bd52d0da2838\") " pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-hrpn8" Dec 11 08:26:17 crc kubenswrapper[4860]: I1211 08:26:17.247966 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ldx9s\" (UniqueName: \"kubernetes.io/projected/211d9c9c-f584-43ca-8db1-7b81f6307c21-kube-api-access-ldx9s\") pod \"swift-operator-controller-manager-9d58d64bc-p7dkp\" (UID: \"211d9c9c-f584-43ca-8db1-7b81f6307c21\") " pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-p7dkp" Dec 11 08:26:17 crc kubenswrapper[4860]: I1211 08:26:17.248779 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-bvs87" Dec 11 08:26:17 crc kubenswrapper[4860]: I1211 08:26:17.305668 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-75944c9b7-ck9nc"] Dec 11 08:26:17 crc kubenswrapper[4860]: I1211 08:26:17.307203 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-75944c9b7-ck9nc" Dec 11 08:26:17 crc kubenswrapper[4860]: I1211 08:26:17.309791 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b85p7\" (UniqueName: \"kubernetes.io/projected/b6c38bbc-80d0-449a-aab0-291078361ebd-kube-api-access-b85p7\") pod \"test-operator-controller-manager-5854674fcc-fh5vc\" (UID: \"b6c38bbc-80d0-449a-aab0-291078361ebd\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-fh5vc" Dec 11 08:26:17 crc kubenswrapper[4860]: I1211 08:26:17.310769 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-78wvx" Dec 11 08:26:17 crc kubenswrapper[4860]: I1211 08:26:17.313075 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-fh5vc"] Dec 11 08:26:17 crc kubenswrapper[4860]: I1211 08:26:17.315794 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-2qmxt" Dec 11 08:26:17 crc kubenswrapper[4860]: I1211 08:26:17.319403 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-75944c9b7-ck9nc"] Dec 11 08:26:17 crc kubenswrapper[4860]: I1211 08:26:17.332311 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b85p7\" (UniqueName: \"kubernetes.io/projected/b6c38bbc-80d0-449a-aab0-291078361ebd-kube-api-access-b85p7\") pod \"test-operator-controller-manager-5854674fcc-fh5vc\" (UID: \"b6c38bbc-80d0-449a-aab0-291078361ebd\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-fh5vc" Dec 11 08:26:17 crc kubenswrapper[4860]: I1211 08:26:17.338186 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-nvt56" Dec 11 08:26:17 crc kubenswrapper[4860]: I1211 08:26:17.353710 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-6d784644c6-68pll"] Dec 11 08:26:17 crc kubenswrapper[4860]: I1211 08:26:17.355392 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-6d784644c6-68pll" Dec 11 08:26:17 crc kubenswrapper[4860]: I1211 08:26:17.358331 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-hrpn8" Dec 11 08:26:17 crc kubenswrapper[4860]: I1211 08:26:17.361766 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-6d784644c6-68pll"] Dec 11 08:26:17 crc kubenswrapper[4860]: I1211 08:26:17.364241 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Dec 11 08:26:17 crc kubenswrapper[4860]: I1211 08:26:17.364626 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Dec 11 08:26:17 crc kubenswrapper[4860]: I1211 08:26:17.366855 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-cx7mf" Dec 11 08:26:17 crc kubenswrapper[4860]: I1211 08:26:17.372723 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-p7dkp" Dec 11 08:26:17 crc kubenswrapper[4860]: I1211 08:26:17.381425 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-jsftr"] Dec 11 08:26:17 crc kubenswrapper[4860]: I1211 08:26:17.383042 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-jsftr" Dec 11 08:26:17 crc kubenswrapper[4860]: I1211 08:26:17.393463 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-wbnf6" Dec 11 08:26:17 crc kubenswrapper[4860]: I1211 08:26:17.406780 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-jsftr"] Dec 11 08:26:17 crc kubenswrapper[4860]: I1211 08:26:17.419371 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l5mvc\" (UniqueName: \"kubernetes.io/projected/8462b48d-cc3d-4714-9558-22818db99c08-kube-api-access-l5mvc\") pod \"watcher-operator-controller-manager-75944c9b7-ck9nc\" (UID: \"8462b48d-cc3d-4714-9558-22818db99c08\") " pod="openstack-operators/watcher-operator-controller-manager-75944c9b7-ck9nc" Dec 11 08:26:17 crc kubenswrapper[4860]: I1211 08:26:17.453996 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-fh5vc" Dec 11 08:26:17 crc kubenswrapper[4860]: I1211 08:26:17.527153 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l5mvc\" (UniqueName: \"kubernetes.io/projected/8462b48d-cc3d-4714-9558-22818db99c08-kube-api-access-l5mvc\") pod \"watcher-operator-controller-manager-75944c9b7-ck9nc\" (UID: \"8462b48d-cc3d-4714-9558-22818db99c08\") " pod="openstack-operators/watcher-operator-controller-manager-75944c9b7-ck9nc" Dec 11 08:26:17 crc kubenswrapper[4860]: I1211 08:26:17.528478 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ndttp\" (UniqueName: \"kubernetes.io/projected/22f3c8f5-1781-4ba2-8610-c9ada5539d92-kube-api-access-ndttp\") pod \"rabbitmq-cluster-operator-manager-668c99d594-jsftr\" (UID: \"22f3c8f5-1781-4ba2-8610-c9ada5539d92\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-jsftr" Dec 11 08:26:17 crc kubenswrapper[4860]: I1211 08:26:17.528536 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/4bbd2122-dfb6-486b-8196-375ffeb93527-webhook-certs\") pod \"openstack-operator-controller-manager-6d784644c6-68pll\" (UID: \"4bbd2122-dfb6-486b-8196-375ffeb93527\") " pod="openstack-operators/openstack-operator-controller-manager-6d784644c6-68pll" Dec 11 08:26:17 crc kubenswrapper[4860]: I1211 08:26:17.528837 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2848x\" (UniqueName: \"kubernetes.io/projected/4bbd2122-dfb6-486b-8196-375ffeb93527-kube-api-access-2848x\") pod \"openstack-operator-controller-manager-6d784644c6-68pll\" (UID: \"4bbd2122-dfb6-486b-8196-375ffeb93527\") " pod="openstack-operators/openstack-operator-controller-manager-6d784644c6-68pll" Dec 11 08:26:17 crc kubenswrapper[4860]: I1211 08:26:17.528931 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/04044c38-9aa4-4439-bad4-cf687e00ed5a-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879fgp9c8\" (UID: \"04044c38-9aa4-4439-bad4-cf687e00ed5a\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fgp9c8" Dec 11 08:26:17 crc kubenswrapper[4860]: I1211 08:26:17.529018 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4bbd2122-dfb6-486b-8196-375ffeb93527-metrics-certs\") pod \"openstack-operator-controller-manager-6d784644c6-68pll\" (UID: \"4bbd2122-dfb6-486b-8196-375ffeb93527\") " pod="openstack-operators/openstack-operator-controller-manager-6d784644c6-68pll" Dec 11 08:26:17 crc kubenswrapper[4860]: E1211 08:26:17.529314 4860 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 11 08:26:17 crc kubenswrapper[4860]: E1211 08:26:17.529369 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/04044c38-9aa4-4439-bad4-cf687e00ed5a-cert podName:04044c38-9aa4-4439-bad4-cf687e00ed5a nodeName:}" failed. No retries permitted until 2025-12-11 08:26:18.529349855 +0000 UTC m=+911.257868910 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/04044c38-9aa4-4439-bad4-cf687e00ed5a-cert") pod "openstack-baremetal-operator-controller-manager-84b575879fgp9c8" (UID: "04044c38-9aa4-4439-bad4-cf687e00ed5a") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 11 08:26:17 crc kubenswrapper[4860]: I1211 08:26:17.573915 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l5mvc\" (UniqueName: \"kubernetes.io/projected/8462b48d-cc3d-4714-9558-22818db99c08-kube-api-access-l5mvc\") pod \"watcher-operator-controller-manager-75944c9b7-ck9nc\" (UID: \"8462b48d-cc3d-4714-9558-22818db99c08\") " pod="openstack-operators/watcher-operator-controller-manager-75944c9b7-ck9nc" Dec 11 08:26:17 crc kubenswrapper[4860]: I1211 08:26:17.630846 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2848x\" (UniqueName: \"kubernetes.io/projected/4bbd2122-dfb6-486b-8196-375ffeb93527-kube-api-access-2848x\") pod \"openstack-operator-controller-manager-6d784644c6-68pll\" (UID: \"4bbd2122-dfb6-486b-8196-375ffeb93527\") " pod="openstack-operators/openstack-operator-controller-manager-6d784644c6-68pll" Dec 11 08:26:17 crc kubenswrapper[4860]: I1211 08:26:17.631391 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4bbd2122-dfb6-486b-8196-375ffeb93527-metrics-certs\") pod \"openstack-operator-controller-manager-6d784644c6-68pll\" (UID: \"4bbd2122-dfb6-486b-8196-375ffeb93527\") " pod="openstack-operators/openstack-operator-controller-manager-6d784644c6-68pll" Dec 11 08:26:17 crc kubenswrapper[4860]: I1211 08:26:17.631467 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ndttp\" (UniqueName: \"kubernetes.io/projected/22f3c8f5-1781-4ba2-8610-c9ada5539d92-kube-api-access-ndttp\") pod \"rabbitmq-cluster-operator-manager-668c99d594-jsftr\" (UID: \"22f3c8f5-1781-4ba2-8610-c9ada5539d92\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-jsftr" Dec 11 08:26:17 crc kubenswrapper[4860]: I1211 08:26:17.631496 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/4bbd2122-dfb6-486b-8196-375ffeb93527-webhook-certs\") pod \"openstack-operator-controller-manager-6d784644c6-68pll\" (UID: \"4bbd2122-dfb6-486b-8196-375ffeb93527\") " pod="openstack-operators/openstack-operator-controller-manager-6d784644c6-68pll" Dec 11 08:26:17 crc kubenswrapper[4860]: E1211 08:26:17.631731 4860 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 11 08:26:17 crc kubenswrapper[4860]: E1211 08:26:17.631797 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4bbd2122-dfb6-486b-8196-375ffeb93527-webhook-certs podName:4bbd2122-dfb6-486b-8196-375ffeb93527 nodeName:}" failed. No retries permitted until 2025-12-11 08:26:18.131776161 +0000 UTC m=+910.860295216 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/4bbd2122-dfb6-486b-8196-375ffeb93527-webhook-certs") pod "openstack-operator-controller-manager-6d784644c6-68pll" (UID: "4bbd2122-dfb6-486b-8196-375ffeb93527") : secret "webhook-server-cert" not found Dec 11 08:26:17 crc kubenswrapper[4860]: E1211 08:26:17.632263 4860 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 11 08:26:17 crc kubenswrapper[4860]: E1211 08:26:17.632289 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4bbd2122-dfb6-486b-8196-375ffeb93527-metrics-certs podName:4bbd2122-dfb6-486b-8196-375ffeb93527 nodeName:}" failed. No retries permitted until 2025-12-11 08:26:18.132281897 +0000 UTC m=+910.860800952 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/4bbd2122-dfb6-486b-8196-375ffeb93527-metrics-certs") pod "openstack-operator-controller-manager-6d784644c6-68pll" (UID: "4bbd2122-dfb6-486b-8196-375ffeb93527") : secret "metrics-server-cert" not found Dec 11 08:26:17 crc kubenswrapper[4860]: I1211 08:26:17.653075 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2848x\" (UniqueName: \"kubernetes.io/projected/4bbd2122-dfb6-486b-8196-375ffeb93527-kube-api-access-2848x\") pod \"openstack-operator-controller-manager-6d784644c6-68pll\" (UID: \"4bbd2122-dfb6-486b-8196-375ffeb93527\") " pod="openstack-operators/openstack-operator-controller-manager-6d784644c6-68pll" Dec 11 08:26:17 crc kubenswrapper[4860]: I1211 08:26:17.654060 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ndttp\" (UniqueName: \"kubernetes.io/projected/22f3c8f5-1781-4ba2-8610-c9ada5539d92-kube-api-access-ndttp\") pod \"rabbitmq-cluster-operator-manager-668c99d594-jsftr\" (UID: \"22f3c8f5-1781-4ba2-8610-c9ada5539d92\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-jsftr" Dec 11 08:26:17 crc kubenswrapper[4860]: I1211 08:26:17.690248 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-75944c9b7-ck9nc" Dec 11 08:26:17 crc kubenswrapper[4860]: I1211 08:26:17.725808 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-jsftr" Dec 11 08:26:17 crc kubenswrapper[4860]: I1211 08:26:17.755290 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6c677c69b-zrrf6"] Dec 11 08:26:17 crc kubenswrapper[4860]: I1211 08:26:17.775372 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-697fb699cf-6t74n"] Dec 11 08:26:17 crc kubenswrapper[4860]: W1211 08:26:17.790524 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda3bf483c_6438_4b7e_abb8_447b9b146644.slice/crio-cb7410bd24fbf1840164f2008ab4d43e3855d798a3cdc7e3c11a640409b40fc4 WatchSource:0}: Error finding container cb7410bd24fbf1840164f2008ab4d43e3855d798a3cdc7e3c11a640409b40fc4: Status 404 returned error can't find the container with id cb7410bd24fbf1840164f2008ab4d43e3855d798a3cdc7e3c11a640409b40fc4 Dec 11 08:26:18 crc kubenswrapper[4860]: I1211 08:26:18.038971 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7f1d9cd2-1e0c-4c4d-86a6-b91a38eadba2-cert\") pod \"infra-operator-controller-manager-78d48bff9d-sq5gd\" (UID: \"7f1d9cd2-1e0c-4c4d-86a6-b91a38eadba2\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-sq5gd" Dec 11 08:26:18 crc kubenswrapper[4860]: E1211 08:26:18.039306 4860 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 11 08:26:18 crc kubenswrapper[4860]: E1211 08:26:18.039698 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7f1d9cd2-1e0c-4c4d-86a6-b91a38eadba2-cert podName:7f1d9cd2-1e0c-4c4d-86a6-b91a38eadba2 nodeName:}" failed. No retries permitted until 2025-12-11 08:26:20.039670262 +0000 UTC m=+912.768189317 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/7f1d9cd2-1e0c-4c4d-86a6-b91a38eadba2-cert") pod "infra-operator-controller-manager-78d48bff9d-sq5gd" (UID: "7f1d9cd2-1e0c-4c4d-86a6-b91a38eadba2") : secret "infra-operator-webhook-server-cert" not found Dec 11 08:26:18 crc kubenswrapper[4860]: I1211 08:26:18.109276 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-zrhjm"] Dec 11 08:26:18 crc kubenswrapper[4860]: W1211 08:26:18.122230 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5a589b13_7f29_40ae_9804_7c5c2bc3d745.slice/crio-65dd893162ca510b67561fc54b9cc251ee82311125d91d979550d829e999f9f9 WatchSource:0}: Error finding container 65dd893162ca510b67561fc54b9cc251ee82311125d91d979550d829e999f9f9: Status 404 returned error can't find the container with id 65dd893162ca510b67561fc54b9cc251ee82311125d91d979550d829e999f9f9 Dec 11 08:26:18 crc kubenswrapper[4860]: I1211 08:26:18.124019 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-5b5fd79c9c-lpm8h"] Dec 11 08:26:18 crc kubenswrapper[4860]: I1211 08:26:18.132921 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-79c8c4686c-4tw58"] Dec 11 08:26:18 crc kubenswrapper[4860]: W1211 08:26:18.136117 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5aff136d_8f48_4cb7_8ddc_72ce966607b7.slice/crio-a48de8d3d568b0151b1d51f6687650a783e9f8bd211092c1b1b96504eb32f5c5 WatchSource:0}: Error finding container a48de8d3d568b0151b1d51f6687650a783e9f8bd211092c1b1b96504eb32f5c5: Status 404 returned error can't find the container with id a48de8d3d568b0151b1d51f6687650a783e9f8bd211092c1b1b96504eb32f5c5 Dec 11 08:26:18 crc kubenswrapper[4860]: I1211 08:26:18.140590 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/4bbd2122-dfb6-486b-8196-375ffeb93527-webhook-certs\") pod \"openstack-operator-controller-manager-6d784644c6-68pll\" (UID: \"4bbd2122-dfb6-486b-8196-375ffeb93527\") " pod="openstack-operators/openstack-operator-controller-manager-6d784644c6-68pll" Dec 11 08:26:18 crc kubenswrapper[4860]: I1211 08:26:18.140705 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4bbd2122-dfb6-486b-8196-375ffeb93527-metrics-certs\") pod \"openstack-operator-controller-manager-6d784644c6-68pll\" (UID: \"4bbd2122-dfb6-486b-8196-375ffeb93527\") " pod="openstack-operators/openstack-operator-controller-manager-6d784644c6-68pll" Dec 11 08:26:18 crc kubenswrapper[4860]: E1211 08:26:18.140882 4860 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 11 08:26:18 crc kubenswrapper[4860]: E1211 08:26:18.140944 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4bbd2122-dfb6-486b-8196-375ffeb93527-metrics-certs podName:4bbd2122-dfb6-486b-8196-375ffeb93527 nodeName:}" failed. No retries permitted until 2025-12-11 08:26:19.140923223 +0000 UTC m=+911.869442278 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/4bbd2122-dfb6-486b-8196-375ffeb93527-metrics-certs") pod "openstack-operator-controller-manager-6d784644c6-68pll" (UID: "4bbd2122-dfb6-486b-8196-375ffeb93527") : secret "metrics-server-cert" not found Dec 11 08:26:18 crc kubenswrapper[4860]: E1211 08:26:18.141002 4860 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 11 08:26:18 crc kubenswrapper[4860]: E1211 08:26:18.141079 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4bbd2122-dfb6-486b-8196-375ffeb93527-webhook-certs podName:4bbd2122-dfb6-486b-8196-375ffeb93527 nodeName:}" failed. No retries permitted until 2025-12-11 08:26:19.141055577 +0000 UTC m=+911.869574632 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/4bbd2122-dfb6-486b-8196-375ffeb93527-webhook-certs") pod "openstack-operator-controller-manager-6d784644c6-68pll" (UID: "4bbd2122-dfb6-486b-8196-375ffeb93527") : secret "webhook-server-cert" not found Dec 11 08:26:18 crc kubenswrapper[4860]: I1211 08:26:18.142289 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-lrhhk"] Dec 11 08:26:18 crc kubenswrapper[4860]: W1211 08:26:18.147741 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod505d980f_cfcb_42b2_876e_a4730abe7ea6.slice/crio-b4ea37849dabbead5d85472ce2bd388fe8ad60c0e81d10f528c36066e4387806 WatchSource:0}: Error finding container b4ea37849dabbead5d85472ce2bd388fe8ad60c0e81d10f528c36066e4387806: Status 404 returned error can't find the container with id b4ea37849dabbead5d85472ce2bd388fe8ad60c0e81d10f528c36066e4387806 Dec 11 08:26:18 crc kubenswrapper[4860]: I1211 08:26:18.148889 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-5697bb5779-kgzwt"] Dec 11 08:26:18 crc kubenswrapper[4860]: I1211 08:26:18.158192 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-967d97867-5wvld"] Dec 11 08:26:18 crc kubenswrapper[4860]: W1211 08:26:18.160076 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod91bb3ff9_fd19_4c79_bac4_621cb114c783.slice/crio-7c07656126160d93b1d99fe632640dd89940f07b24b5f8d577a36b644f79b002 WatchSource:0}: Error finding container 7c07656126160d93b1d99fe632640dd89940f07b24b5f8d577a36b644f79b002: Status 404 returned error can't find the container with id 7c07656126160d93b1d99fe632640dd89940f07b24b5f8d577a36b644f79b002 Dec 11 08:26:18 crc kubenswrapper[4860]: I1211 08:26:18.170900 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-6bpb9"] Dec 11 08:26:18 crc kubenswrapper[4860]: I1211 08:26:18.182980 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-bshq2"] Dec 11 08:26:18 crc kubenswrapper[4860]: I1211 08:26:18.217862 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-ht7db"] Dec 11 08:26:18 crc kubenswrapper[4860]: W1211 08:26:18.229718 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod689923af_60de_49e9_bbed_bebaa63adc4b.slice/crio-51577831d800a8256d275689205751b1d7f908f7d054dfd806da190e6a9568e2 WatchSource:0}: Error finding container 51577831d800a8256d275689205751b1d7f908f7d054dfd806da190e6a9568e2: Status 404 returned error can't find the container with id 51577831d800a8256d275689205751b1d7f908f7d054dfd806da190e6a9568e2 Dec 11 08:26:18 crc kubenswrapper[4860]: I1211 08:26:18.231462 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-65fv4"] Dec 11 08:26:18 crc kubenswrapper[4860]: W1211 08:26:18.234532 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddc79abc2_7b6c_46c4_b9b4_a9c76696a8b2.slice/crio-a1210c53560e4871c75875b4ae7d9047aed860882b07252449eaa7bac1fb6bcb WatchSource:0}: Error finding container a1210c53560e4871c75875b4ae7d9047aed860882b07252449eaa7bac1fb6bcb: Status 404 returned error can't find the container with id a1210c53560e4871c75875b4ae7d9047aed860882b07252449eaa7bac1fb6bcb Dec 11 08:26:18 crc kubenswrapper[4860]: W1211 08:26:18.234913 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda52f8df6_7c28_4776_aca5_bd8b47e82fe8.slice/crio-4739997cbeccb54db2d0cec8d03165945c833188a77abe577361f2c981e5ca65 WatchSource:0}: Error finding container 4739997cbeccb54db2d0cec8d03165945c833188a77abe577361f2c981e5ca65: Status 404 returned error can't find the container with id 4739997cbeccb54db2d0cec8d03165945c833188a77abe577361f2c981e5ca65 Dec 11 08:26:18 crc kubenswrapper[4860]: E1211 08:26:18.237551 4860 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/neutron-operator@sha256:0b3fb69f35c151895d3dffd514974a9f9fe1c77c3bca69b78b81efb183cf4557,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-6rvhf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-5fdfd5b6b5-ht7db_openstack-operators(dc79abc2-7b6c-46c4-b9b4-a9c76696a8b2): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 11 08:26:18 crc kubenswrapper[4860]: E1211 08:26:18.238010 4860 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-nnldp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-78f8948974-2qmxt_openstack-operators(a52f8df6-7c28-4776-aca5-bd8b47e82fe8): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 11 08:26:18 crc kubenswrapper[4860]: E1211 08:26:18.241827 4860 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-nnldp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-78f8948974-2qmxt_openstack-operators(a52f8df6-7c28-4776-aca5-bd8b47e82fe8): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 11 08:26:18 crc kubenswrapper[4860]: E1211 08:26:18.241959 4860 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-6rvhf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-5fdfd5b6b5-ht7db_openstack-operators(dc79abc2-7b6c-46c4-b9b4-a9c76696a8b2): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 11 08:26:18 crc kubenswrapper[4860]: I1211 08:26:18.242538 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-2qmxt"] Dec 11 08:26:18 crc kubenswrapper[4860]: E1211 08:26:18.243032 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/placement-operator-controller-manager-78f8948974-2qmxt" podUID="a52f8df6-7c28-4776-aca5-bd8b47e82fe8" Dec 11 08:26:18 crc kubenswrapper[4860]: E1211 08:26:18.243132 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-ht7db" podUID="dc79abc2-7b6c-46c4-b9b4-a9c76696a8b2" Dec 11 08:26:18 crc kubenswrapper[4860]: I1211 08:26:18.249722 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-bvs87"] Dec 11 08:26:18 crc kubenswrapper[4860]: I1211 08:26:18.432005 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-zrhjm" event={"ID":"5aff136d-8f48-4cb7-8ddc-72ce966607b7","Type":"ContainerStarted","Data":"a48de8d3d568b0151b1d51f6687650a783e9f8bd211092c1b1b96504eb32f5c5"} Dec 11 08:26:18 crc kubenswrapper[4860]: I1211 08:26:18.436846 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-4tw58" event={"ID":"5a589b13-7f29-40ae-9804-7c5c2bc3d745","Type":"ContainerStarted","Data":"65dd893162ca510b67561fc54b9cc251ee82311125d91d979550d829e999f9f9"} Dec 11 08:26:18 crc kubenswrapper[4860]: I1211 08:26:18.439580 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-65fv4" event={"ID":"bb231e77-e19b-41ed-bf7a-ae354e1089ec","Type":"ContainerStarted","Data":"a666955d9c94f7e5f0e1778bb4d8fa1094abf28406d3c116a33e0fe79f678b75"} Dec 11 08:26:18 crc kubenswrapper[4860]: I1211 08:26:18.440777 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-lpm8h" event={"ID":"aae650b9-dcbc-482b-b49f-0a45651bb4b4","Type":"ContainerStarted","Data":"f0256584fe1e2641d3279425a699f741af3c84106b26acea8a45a6b707a733cd"} Dec 11 08:26:18 crc kubenswrapper[4860]: I1211 08:26:18.452745 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-zrrf6" event={"ID":"323ad77f-a0d6-461d-b483-2ffd74ebb9b7","Type":"ContainerStarted","Data":"e93ea5fcac6bc62065024e507d3888f8f50c52c41dec90d25f81bf8bda3ca443"} Dec 11 08:26:18 crc kubenswrapper[4860]: I1211 08:26:18.457929 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-6bpb9" event={"ID":"d5e0079f-f500-407f-b06f-a610a19c741b","Type":"ContainerStarted","Data":"ee14bb7c09c11ccd7261e95e936e28eb339b8d32618040b26f3edfcd6a6e91bd"} Dec 11 08:26:18 crc kubenswrapper[4860]: I1211 08:26:18.463958 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-lrhhk" event={"ID":"505d980f-cfcb-42b2-876e-a4730abe7ea6","Type":"ContainerStarted","Data":"b4ea37849dabbead5d85472ce2bd388fe8ad60c0e81d10f528c36066e4387806"} Dec 11 08:26:18 crc kubenswrapper[4860]: I1211 08:26:18.467066 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-967d97867-5wvld" event={"ID":"91bb3ff9-fd19-4c79-bac4-621cb114c783","Type":"ContainerStarted","Data":"7c07656126160d93b1d99fe632640dd89940f07b24b5f8d577a36b644f79b002"} Dec 11 08:26:18 crc kubenswrapper[4860]: I1211 08:26:18.469001 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-2qmxt" event={"ID":"a52f8df6-7c28-4776-aca5-bd8b47e82fe8","Type":"ContainerStarted","Data":"4739997cbeccb54db2d0cec8d03165945c833188a77abe577361f2c981e5ca65"} Dec 11 08:26:18 crc kubenswrapper[4860]: I1211 08:26:18.471486 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-6t74n" event={"ID":"a3bf483c-6438-4b7e-abb8-447b9b146644","Type":"ContainerStarted","Data":"cb7410bd24fbf1840164f2008ab4d43e3855d798a3cdc7e3c11a640409b40fc4"} Dec 11 08:26:18 crc kubenswrapper[4860]: I1211 08:26:18.481441 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-bvs87" event={"ID":"689923af-60de-49e9-bbed-bebaa63adc4b","Type":"ContainerStarted","Data":"51577831d800a8256d275689205751b1d7f908f7d054dfd806da190e6a9568e2"} Dec 11 08:26:18 crc kubenswrapper[4860]: I1211 08:26:18.481817 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-9d58d64bc-p7dkp"] Dec 11 08:26:18 crc kubenswrapper[4860]: E1211 08:26:18.515265 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/placement-operator-controller-manager-78f8948974-2qmxt" podUID="a52f8df6-7c28-4776-aca5-bd8b47e82fe8" Dec 11 08:26:18 crc kubenswrapper[4860]: I1211 08:26:18.515419 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-nvt56"] Dec 11 08:26:18 crc kubenswrapper[4860]: W1211 08:26:18.522685 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod99f7e97f_7cea_4785_b95e_133adf238ac3.slice/crio-0a7e8ac92f3b94a94c477767acf9b239d0b8de7138f4a78a4fcfeb816f77eda4 WatchSource:0}: Error finding container 0a7e8ac92f3b94a94c477767acf9b239d0b8de7138f4a78a4fcfeb816f77eda4: Status 404 returned error can't find the container with id 0a7e8ac92f3b94a94c477767acf9b239d0b8de7138f4a78a4fcfeb816f77eda4 Dec 11 08:26:18 crc kubenswrapper[4860]: W1211 08:26:18.525342 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod211d9c9c_f584_43ca_8db1_7b81f6307c21.slice/crio-93e1fa9807bdf18b59f3dbb06954b7711e398565ad81a6bfd1a6d8f684c738b8 WatchSource:0}: Error finding container 93e1fa9807bdf18b59f3dbb06954b7711e398565ad81a6bfd1a6d8f684c738b8: Status 404 returned error can't find the container with id 93e1fa9807bdf18b59f3dbb06954b7711e398565ad81a6bfd1a6d8f684c738b8 Dec 11 08:26:18 crc kubenswrapper[4860]: I1211 08:26:18.527764 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-fh5vc"] Dec 11 08:26:18 crc kubenswrapper[4860]: I1211 08:26:18.527896 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-kgzwt" event={"ID":"e00c360d-20a4-4e2e-81ea-e4ffa1bc62fc","Type":"ContainerStarted","Data":"a395eac6d6474bd4c5be1a793961971f26384e15e2e43516d339c4f6c780c035"} Dec 11 08:26:18 crc kubenswrapper[4860]: I1211 08:26:18.543935 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-ht7db" event={"ID":"dc79abc2-7b6c-46c4-b9b4-a9c76696a8b2","Type":"ContainerStarted","Data":"a1210c53560e4871c75875b4ae7d9047aed860882b07252449eaa7bac1fb6bcb"} Dec 11 08:26:18 crc kubenswrapper[4860]: E1211 08:26:18.546662 4860 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-b85p7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5854674fcc-fh5vc_openstack-operators(b6c38bbc-80d0-449a-aab0-291078361ebd): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 11 08:26:18 crc kubenswrapper[4860]: I1211 08:26:18.548218 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-75944c9b7-ck9nc"] Dec 11 08:26:18 crc kubenswrapper[4860]: E1211 08:26:18.549396 4860 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-b85p7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5854674fcc-fh5vc_openstack-operators(b6c38bbc-80d0-449a-aab0-291078361ebd): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 11 08:26:18 crc kubenswrapper[4860]: E1211 08:26:18.549442 4860 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-jdhtm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-b6456fdb6-nvt56_openstack-operators(99f7e97f-7cea-4785-b95e-133adf238ac3): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 11 08:26:18 crc kubenswrapper[4860]: E1211 08:26:18.550717 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-fh5vc" podUID="b6c38bbc-80d0-449a-aab0-291078361ebd" Dec 11 08:26:18 crc kubenswrapper[4860]: I1211 08:26:18.553946 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/04044c38-9aa4-4439-bad4-cf687e00ed5a-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879fgp9c8\" (UID: \"04044c38-9aa4-4439-bad4-cf687e00ed5a\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fgp9c8" Dec 11 08:26:18 crc kubenswrapper[4860]: E1211 08:26:18.555897 4860 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 11 08:26:18 crc kubenswrapper[4860]: E1211 08:26:18.555999 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/04044c38-9aa4-4439-bad4-cf687e00ed5a-cert podName:04044c38-9aa4-4439-bad4-cf687e00ed5a nodeName:}" failed. No retries permitted until 2025-12-11 08:26:20.555968919 +0000 UTC m=+913.284487974 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/04044c38-9aa4-4439-bad4-cf687e00ed5a-cert") pod "openstack-baremetal-operator-controller-manager-84b575879fgp9c8" (UID: "04044c38-9aa4-4439-bad4-cf687e00ed5a") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 11 08:26:18 crc kubenswrapper[4860]: W1211 08:26:18.563885 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8462b48d_cc3d_4714_9558_22818db99c08.slice/crio-d90e7e1766e5f17783b5ebc598fa6d1e14f3ae0798a45a8c649727089e9163d9 WatchSource:0}: Error finding container d90e7e1766e5f17783b5ebc598fa6d1e14f3ae0798a45a8c649727089e9163d9: Status 404 returned error can't find the container with id d90e7e1766e5f17783b5ebc598fa6d1e14f3ae0798a45a8c649727089e9163d9 Dec 11 08:26:18 crc kubenswrapper[4860]: I1211 08:26:18.566931 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-jsftr"] Dec 11 08:26:18 crc kubenswrapper[4860]: E1211 08:26:18.584380 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:0b3fb69f35c151895d3dffd514974a9f9fe1c77c3bca69b78b81efb183cf4557\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-ht7db" podUID="dc79abc2-7b6c-46c4-b9b4-a9c76696a8b2" Dec 11 08:26:18 crc kubenswrapper[4860]: E1211 08:26:18.585457 4860 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-ndttp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-jsftr_openstack-operators(22f3c8f5-1781-4ba2-8610-c9ada5539d92): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 11 08:26:18 crc kubenswrapper[4860]: E1211 08:26:18.587261 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-jsftr" podUID="22f3c8f5-1781-4ba2-8610-c9ada5539d92" Dec 11 08:26:18 crc kubenswrapper[4860]: I1211 08:26:18.587699 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-bshq2" event={"ID":"64f09777-fca3-412f-98f8-5cd189cb9fbd","Type":"ContainerStarted","Data":"249b47dec86516f7cb470273115ecb21f99852d344e0069b34dcd70e14ff9222"} Dec 11 08:26:18 crc kubenswrapper[4860]: I1211 08:26:18.592716 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-58d5ff84df-hrpn8"] Dec 11 08:26:18 crc kubenswrapper[4860]: E1211 08:26:18.594703 4860 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:961417d59f527d925ac48ff6a11de747d0493315e496e34dc83d76a1a1fff58a,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-l5mvc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-75944c9b7-ck9nc_openstack-operators(8462b48d-cc3d-4714-9558-22818db99c08): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 11 08:26:18 crc kubenswrapper[4860]: E1211 08:26:18.595371 4860 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:f27e732ec1faee765461bf137d9be81278b2fa39675019a73622755e1e610b6f,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5vl5s,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-58d5ff84df-hrpn8_openstack-operators(e36550af-98a6-49b3-9769-bd52d0da2838): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 11 08:26:18 crc kubenswrapper[4860]: E1211 08:26:18.596764 4860 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-l5mvc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-75944c9b7-ck9nc_openstack-operators(8462b48d-cc3d-4714-9558-22818db99c08): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 11 08:26:18 crc kubenswrapper[4860]: E1211 08:26:18.596928 4860 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5vl5s,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-58d5ff84df-hrpn8_openstack-operators(e36550af-98a6-49b3-9769-bd52d0da2838): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 11 08:26:18 crc kubenswrapper[4860]: E1211 08:26:18.598848 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-hrpn8" podUID="e36550af-98a6-49b3-9769-bd52d0da2838" Dec 11 08:26:18 crc kubenswrapper[4860]: E1211 08:26:18.598909 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/watcher-operator-controller-manager-75944c9b7-ck9nc" podUID="8462b48d-cc3d-4714-9558-22818db99c08" Dec 11 08:26:18 crc kubenswrapper[4860]: I1211 08:26:18.653102 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-2425r" Dec 11 08:26:18 crc kubenswrapper[4860]: I1211 08:26:18.653236 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-2425r" Dec 11 08:26:18 crc kubenswrapper[4860]: I1211 08:26:18.703563 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-2425r" Dec 11 08:26:19 crc kubenswrapper[4860]: I1211 08:26:19.170178 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/4bbd2122-dfb6-486b-8196-375ffeb93527-webhook-certs\") pod \"openstack-operator-controller-manager-6d784644c6-68pll\" (UID: \"4bbd2122-dfb6-486b-8196-375ffeb93527\") " pod="openstack-operators/openstack-operator-controller-manager-6d784644c6-68pll" Dec 11 08:26:19 crc kubenswrapper[4860]: I1211 08:26:19.170331 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4bbd2122-dfb6-486b-8196-375ffeb93527-metrics-certs\") pod \"openstack-operator-controller-manager-6d784644c6-68pll\" (UID: \"4bbd2122-dfb6-486b-8196-375ffeb93527\") " pod="openstack-operators/openstack-operator-controller-manager-6d784644c6-68pll" Dec 11 08:26:19 crc kubenswrapper[4860]: E1211 08:26:19.170482 4860 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 11 08:26:19 crc kubenswrapper[4860]: E1211 08:26:19.170636 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4bbd2122-dfb6-486b-8196-375ffeb93527-webhook-certs podName:4bbd2122-dfb6-486b-8196-375ffeb93527 nodeName:}" failed. No retries permitted until 2025-12-11 08:26:21.170601323 +0000 UTC m=+913.899120388 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/4bbd2122-dfb6-486b-8196-375ffeb93527-webhook-certs") pod "openstack-operator-controller-manager-6d784644c6-68pll" (UID: "4bbd2122-dfb6-486b-8196-375ffeb93527") : secret "webhook-server-cert" not found Dec 11 08:26:19 crc kubenswrapper[4860]: E1211 08:26:19.170680 4860 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 11 08:26:19 crc kubenswrapper[4860]: E1211 08:26:19.170864 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4bbd2122-dfb6-486b-8196-375ffeb93527-metrics-certs podName:4bbd2122-dfb6-486b-8196-375ffeb93527 nodeName:}" failed. No retries permitted until 2025-12-11 08:26:21.17080489 +0000 UTC m=+913.899324125 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/4bbd2122-dfb6-486b-8196-375ffeb93527-metrics-certs") pod "openstack-operator-controller-manager-6d784644c6-68pll" (UID: "4bbd2122-dfb6-486b-8196-375ffeb93527") : secret "metrics-server-cert" not found Dec 11 08:26:19 crc kubenswrapper[4860]: I1211 08:26:19.596138 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-nvt56" event={"ID":"99f7e97f-7cea-4785-b95e-133adf238ac3","Type":"ContainerStarted","Data":"0a7e8ac92f3b94a94c477767acf9b239d0b8de7138f4a78a4fcfeb816f77eda4"} Dec 11 08:26:19 crc kubenswrapper[4860]: I1211 08:26:19.596368 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-fh5vc" event={"ID":"b6c38bbc-80d0-449a-aab0-291078361ebd","Type":"ContainerStarted","Data":"51d55e87428569577746f0ce2cf07c7331c4cf671afbe6cd61b4e444583e4b42"} Dec 11 08:26:19 crc kubenswrapper[4860]: I1211 08:26:19.597937 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-hrpn8" event={"ID":"e36550af-98a6-49b3-9769-bd52d0da2838","Type":"ContainerStarted","Data":"5f0d5d21987e9aaff7f8fa5341fa1357fc7f44bf99d2b030cdc21e746c8c90d4"} Dec 11 08:26:19 crc kubenswrapper[4860]: E1211 08:26:19.599601 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-fh5vc" podUID="b6c38bbc-80d0-449a-aab0-291078361ebd" Dec 11 08:26:19 crc kubenswrapper[4860]: I1211 08:26:19.600779 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-jsftr" event={"ID":"22f3c8f5-1781-4ba2-8610-c9ada5539d92","Type":"ContainerStarted","Data":"cc8a61db51baed8cad86de05b7bedbd6231372df64047ba491c2e17df9440aaf"} Dec 11 08:26:19 crc kubenswrapper[4860]: E1211 08:26:19.601039 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:f27e732ec1faee765461bf137d9be81278b2fa39675019a73622755e1e610b6f\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-hrpn8" podUID="e36550af-98a6-49b3-9769-bd52d0da2838" Dec 11 08:26:19 crc kubenswrapper[4860]: I1211 08:26:19.602321 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-p7dkp" event={"ID":"211d9c9c-f584-43ca-8db1-7b81f6307c21","Type":"ContainerStarted","Data":"93e1fa9807bdf18b59f3dbb06954b7711e398565ad81a6bfd1a6d8f684c738b8"} Dec 11 08:26:19 crc kubenswrapper[4860]: E1211 08:26:19.602548 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-jsftr" podUID="22f3c8f5-1781-4ba2-8610-c9ada5539d92" Dec 11 08:26:19 crc kubenswrapper[4860]: I1211 08:26:19.603749 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-75944c9b7-ck9nc" event={"ID":"8462b48d-cc3d-4714-9558-22818db99c08","Type":"ContainerStarted","Data":"d90e7e1766e5f17783b5ebc598fa6d1e14f3ae0798a45a8c649727089e9163d9"} Dec 11 08:26:19 crc kubenswrapper[4860]: E1211 08:26:19.605932 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:0b3fb69f35c151895d3dffd514974a9f9fe1c77c3bca69b78b81efb183cf4557\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-ht7db" podUID="dc79abc2-7b6c-46c4-b9b4-a9c76696a8b2" Dec 11 08:26:19 crc kubenswrapper[4860]: E1211 08:26:19.606399 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:961417d59f527d925ac48ff6a11de747d0493315e496e34dc83d76a1a1fff58a\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/watcher-operator-controller-manager-75944c9b7-ck9nc" podUID="8462b48d-cc3d-4714-9558-22818db99c08" Dec 11 08:26:19 crc kubenswrapper[4860]: E1211 08:26:19.606717 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/placement-operator-controller-manager-78f8948974-2qmxt" podUID="a52f8df6-7c28-4776-aca5-bd8b47e82fe8" Dec 11 08:26:19 crc kubenswrapper[4860]: I1211 08:26:19.670867 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-2425r" Dec 11 08:26:19 crc kubenswrapper[4860]: I1211 08:26:19.764336 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2425r"] Dec 11 08:26:20 crc kubenswrapper[4860]: I1211 08:26:20.095156 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7f1d9cd2-1e0c-4c4d-86a6-b91a38eadba2-cert\") pod \"infra-operator-controller-manager-78d48bff9d-sq5gd\" (UID: \"7f1d9cd2-1e0c-4c4d-86a6-b91a38eadba2\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-sq5gd" Dec 11 08:26:20 crc kubenswrapper[4860]: E1211 08:26:20.095319 4860 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 11 08:26:20 crc kubenswrapper[4860]: E1211 08:26:20.095399 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7f1d9cd2-1e0c-4c4d-86a6-b91a38eadba2-cert podName:7f1d9cd2-1e0c-4c4d-86a6-b91a38eadba2 nodeName:}" failed. No retries permitted until 2025-12-11 08:26:24.095378797 +0000 UTC m=+916.823897852 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/7f1d9cd2-1e0c-4c4d-86a6-b91a38eadba2-cert") pod "infra-operator-controller-manager-78d48bff9d-sq5gd" (UID: "7f1d9cd2-1e0c-4c4d-86a6-b91a38eadba2") : secret "infra-operator-webhook-server-cert" not found Dec 11 08:26:20 crc kubenswrapper[4860]: I1211 08:26:20.605739 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/04044c38-9aa4-4439-bad4-cf687e00ed5a-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879fgp9c8\" (UID: \"04044c38-9aa4-4439-bad4-cf687e00ed5a\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fgp9c8" Dec 11 08:26:20 crc kubenswrapper[4860]: E1211 08:26:20.605983 4860 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 11 08:26:20 crc kubenswrapper[4860]: E1211 08:26:20.606397 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/04044c38-9aa4-4439-bad4-cf687e00ed5a-cert podName:04044c38-9aa4-4439-bad4-cf687e00ed5a nodeName:}" failed. No retries permitted until 2025-12-11 08:26:24.606372856 +0000 UTC m=+917.334891911 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/04044c38-9aa4-4439-bad4-cf687e00ed5a-cert") pod "openstack-baremetal-operator-controller-manager-84b575879fgp9c8" (UID: "04044c38-9aa4-4439-bad4-cf687e00ed5a") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 11 08:26:20 crc kubenswrapper[4860]: E1211 08:26:20.623295 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-jsftr" podUID="22f3c8f5-1781-4ba2-8610-c9ada5539d92" Dec 11 08:26:20 crc kubenswrapper[4860]: E1211 08:26:20.625492 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:f27e732ec1faee765461bf137d9be81278b2fa39675019a73622755e1e610b6f\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-hrpn8" podUID="e36550af-98a6-49b3-9769-bd52d0da2838" Dec 11 08:26:20 crc kubenswrapper[4860]: E1211 08:26:20.625551 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-fh5vc" podUID="b6c38bbc-80d0-449a-aab0-291078361ebd" Dec 11 08:26:20 crc kubenswrapper[4860]: E1211 08:26:20.625975 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:961417d59f527d925ac48ff6a11de747d0493315e496e34dc83d76a1a1fff58a\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/watcher-operator-controller-manager-75944c9b7-ck9nc" podUID="8462b48d-cc3d-4714-9558-22818db99c08" Dec 11 08:26:21 crc kubenswrapper[4860]: I1211 08:26:21.217206 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/4bbd2122-dfb6-486b-8196-375ffeb93527-webhook-certs\") pod \"openstack-operator-controller-manager-6d784644c6-68pll\" (UID: \"4bbd2122-dfb6-486b-8196-375ffeb93527\") " pod="openstack-operators/openstack-operator-controller-manager-6d784644c6-68pll" Dec 11 08:26:21 crc kubenswrapper[4860]: E1211 08:26:21.217494 4860 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 11 08:26:21 crc kubenswrapper[4860]: I1211 08:26:21.217535 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4bbd2122-dfb6-486b-8196-375ffeb93527-metrics-certs\") pod \"openstack-operator-controller-manager-6d784644c6-68pll\" (UID: \"4bbd2122-dfb6-486b-8196-375ffeb93527\") " pod="openstack-operators/openstack-operator-controller-manager-6d784644c6-68pll" Dec 11 08:26:21 crc kubenswrapper[4860]: E1211 08:26:21.217610 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4bbd2122-dfb6-486b-8196-375ffeb93527-webhook-certs podName:4bbd2122-dfb6-486b-8196-375ffeb93527 nodeName:}" failed. No retries permitted until 2025-12-11 08:26:25.217561692 +0000 UTC m=+917.946080747 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/4bbd2122-dfb6-486b-8196-375ffeb93527-webhook-certs") pod "openstack-operator-controller-manager-6d784644c6-68pll" (UID: "4bbd2122-dfb6-486b-8196-375ffeb93527") : secret "webhook-server-cert" not found Dec 11 08:26:21 crc kubenswrapper[4860]: E1211 08:26:21.217934 4860 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 11 08:26:21 crc kubenswrapper[4860]: E1211 08:26:21.218077 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4bbd2122-dfb6-486b-8196-375ffeb93527-metrics-certs podName:4bbd2122-dfb6-486b-8196-375ffeb93527 nodeName:}" failed. No retries permitted until 2025-12-11 08:26:25.218037266 +0000 UTC m=+917.946556501 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/4bbd2122-dfb6-486b-8196-375ffeb93527-metrics-certs") pod "openstack-operator-controller-manager-6d784644c6-68pll" (UID: "4bbd2122-dfb6-486b-8196-375ffeb93527") : secret "metrics-server-cert" not found Dec 11 08:26:21 crc kubenswrapper[4860]: I1211 08:26:21.632191 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-2425r" podUID="16fb4ab2-190c-4a33-8ba1-f9cd5edaaff3" containerName="registry-server" containerID="cri-o://e35bb40199af54a9bbd197edd80583cfb35a628ea0965a63590700958977af63" gracePeriod=2 Dec 11 08:26:22 crc kubenswrapper[4860]: I1211 08:26:22.650352 4860 generic.go:334] "Generic (PLEG): container finished" podID="16fb4ab2-190c-4a33-8ba1-f9cd5edaaff3" containerID="e35bb40199af54a9bbd197edd80583cfb35a628ea0965a63590700958977af63" exitCode=0 Dec 11 08:26:22 crc kubenswrapper[4860]: I1211 08:26:22.650441 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2425r" event={"ID":"16fb4ab2-190c-4a33-8ba1-f9cd5edaaff3","Type":"ContainerDied","Data":"e35bb40199af54a9bbd197edd80583cfb35a628ea0965a63590700958977af63"} Dec 11 08:26:24 crc kubenswrapper[4860]: I1211 08:26:24.177966 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7f1d9cd2-1e0c-4c4d-86a6-b91a38eadba2-cert\") pod \"infra-operator-controller-manager-78d48bff9d-sq5gd\" (UID: \"7f1d9cd2-1e0c-4c4d-86a6-b91a38eadba2\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-sq5gd" Dec 11 08:26:24 crc kubenswrapper[4860]: E1211 08:26:24.178160 4860 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 11 08:26:24 crc kubenswrapper[4860]: E1211 08:26:24.178250 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7f1d9cd2-1e0c-4c4d-86a6-b91a38eadba2-cert podName:7f1d9cd2-1e0c-4c4d-86a6-b91a38eadba2 nodeName:}" failed. No retries permitted until 2025-12-11 08:26:32.178223115 +0000 UTC m=+924.906742170 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/7f1d9cd2-1e0c-4c4d-86a6-b91a38eadba2-cert") pod "infra-operator-controller-manager-78d48bff9d-sq5gd" (UID: "7f1d9cd2-1e0c-4c4d-86a6-b91a38eadba2") : secret "infra-operator-webhook-server-cert" not found Dec 11 08:26:24 crc kubenswrapper[4860]: I1211 08:26:24.687553 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/04044c38-9aa4-4439-bad4-cf687e00ed5a-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879fgp9c8\" (UID: \"04044c38-9aa4-4439-bad4-cf687e00ed5a\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fgp9c8" Dec 11 08:26:24 crc kubenswrapper[4860]: E1211 08:26:24.687799 4860 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 11 08:26:24 crc kubenswrapper[4860]: E1211 08:26:24.687907 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/04044c38-9aa4-4439-bad4-cf687e00ed5a-cert podName:04044c38-9aa4-4439-bad4-cf687e00ed5a nodeName:}" failed. No retries permitted until 2025-12-11 08:26:32.687881862 +0000 UTC m=+925.416400917 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/04044c38-9aa4-4439-bad4-cf687e00ed5a-cert") pod "openstack-baremetal-operator-controller-manager-84b575879fgp9c8" (UID: "04044c38-9aa4-4439-bad4-cf687e00ed5a") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 11 08:26:25 crc kubenswrapper[4860]: I1211 08:26:25.297242 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4bbd2122-dfb6-486b-8196-375ffeb93527-metrics-certs\") pod \"openstack-operator-controller-manager-6d784644c6-68pll\" (UID: \"4bbd2122-dfb6-486b-8196-375ffeb93527\") " pod="openstack-operators/openstack-operator-controller-manager-6d784644c6-68pll" Dec 11 08:26:25 crc kubenswrapper[4860]: I1211 08:26:25.297368 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/4bbd2122-dfb6-486b-8196-375ffeb93527-webhook-certs\") pod \"openstack-operator-controller-manager-6d784644c6-68pll\" (UID: \"4bbd2122-dfb6-486b-8196-375ffeb93527\") " pod="openstack-operators/openstack-operator-controller-manager-6d784644c6-68pll" Dec 11 08:26:25 crc kubenswrapper[4860]: E1211 08:26:25.297527 4860 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 11 08:26:25 crc kubenswrapper[4860]: E1211 08:26:25.297575 4860 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 11 08:26:25 crc kubenswrapper[4860]: E1211 08:26:25.297690 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4bbd2122-dfb6-486b-8196-375ffeb93527-metrics-certs podName:4bbd2122-dfb6-486b-8196-375ffeb93527 nodeName:}" failed. No retries permitted until 2025-12-11 08:26:33.297635912 +0000 UTC m=+926.026155147 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/4bbd2122-dfb6-486b-8196-375ffeb93527-metrics-certs") pod "openstack-operator-controller-manager-6d784644c6-68pll" (UID: "4bbd2122-dfb6-486b-8196-375ffeb93527") : secret "metrics-server-cert" not found Dec 11 08:26:25 crc kubenswrapper[4860]: E1211 08:26:25.297717 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4bbd2122-dfb6-486b-8196-375ffeb93527-webhook-certs podName:4bbd2122-dfb6-486b-8196-375ffeb93527 nodeName:}" failed. No retries permitted until 2025-12-11 08:26:33.297708494 +0000 UTC m=+926.026227749 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/4bbd2122-dfb6-486b-8196-375ffeb93527-webhook-certs") pod "openstack-operator-controller-manager-6d784644c6-68pll" (UID: "4bbd2122-dfb6-486b-8196-375ffeb93527") : secret "webhook-server-cert" not found Dec 11 08:26:28 crc kubenswrapper[4860]: E1211 08:26:28.653662 4860 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of e35bb40199af54a9bbd197edd80583cfb35a628ea0965a63590700958977af63 is running failed: container process not found" containerID="e35bb40199af54a9bbd197edd80583cfb35a628ea0965a63590700958977af63" cmd=["grpc_health_probe","-addr=:50051"] Dec 11 08:26:28 crc kubenswrapper[4860]: E1211 08:26:28.655629 4860 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of e35bb40199af54a9bbd197edd80583cfb35a628ea0965a63590700958977af63 is running failed: container process not found" containerID="e35bb40199af54a9bbd197edd80583cfb35a628ea0965a63590700958977af63" cmd=["grpc_health_probe","-addr=:50051"] Dec 11 08:26:28 crc kubenswrapper[4860]: E1211 08:26:28.656081 4860 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of e35bb40199af54a9bbd197edd80583cfb35a628ea0965a63590700958977af63 is running failed: container process not found" containerID="e35bb40199af54a9bbd197edd80583cfb35a628ea0965a63590700958977af63" cmd=["grpc_health_probe","-addr=:50051"] Dec 11 08:26:28 crc kubenswrapper[4860]: E1211 08:26:28.656813 4860 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of e35bb40199af54a9bbd197edd80583cfb35a628ea0965a63590700958977af63 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/redhat-marketplace-2425r" podUID="16fb4ab2-190c-4a33-8ba1-f9cd5edaaff3" containerName="registry-server" Dec 11 08:26:32 crc kubenswrapper[4860]: I1211 08:26:32.224778 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7f1d9cd2-1e0c-4c4d-86a6-b91a38eadba2-cert\") pod \"infra-operator-controller-manager-78d48bff9d-sq5gd\" (UID: \"7f1d9cd2-1e0c-4c4d-86a6-b91a38eadba2\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-sq5gd" Dec 11 08:26:32 crc kubenswrapper[4860]: I1211 08:26:32.238755 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7f1d9cd2-1e0c-4c4d-86a6-b91a38eadba2-cert\") pod \"infra-operator-controller-manager-78d48bff9d-sq5gd\" (UID: \"7f1d9cd2-1e0c-4c4d-86a6-b91a38eadba2\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-sq5gd" Dec 11 08:26:32 crc kubenswrapper[4860]: E1211 08:26:32.416881 4860 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/swift-operator@sha256:3aa109bb973253ae9dcf339b9b65abbd1176cdb4be672c93e538a5f113816991" Dec 11 08:26:32 crc kubenswrapper[4860]: E1211 08:26:32.417201 4860 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:3aa109bb973253ae9dcf339b9b65abbd1176cdb4be672c93e538a5f113816991,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-ldx9s,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-9d58d64bc-p7dkp_openstack-operators(211d9c9c-f584-43ca-8db1-7b81f6307c21): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 11 08:26:32 crc kubenswrapper[4860]: I1211 08:26:32.448405 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-sq5gd" Dec 11 08:26:32 crc kubenswrapper[4860]: I1211 08:26:32.735212 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/04044c38-9aa4-4439-bad4-cf687e00ed5a-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879fgp9c8\" (UID: \"04044c38-9aa4-4439-bad4-cf687e00ed5a\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fgp9c8" Dec 11 08:26:32 crc kubenswrapper[4860]: I1211 08:26:32.753048 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/04044c38-9aa4-4439-bad4-cf687e00ed5a-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879fgp9c8\" (UID: \"04044c38-9aa4-4439-bad4-cf687e00ed5a\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fgp9c8" Dec 11 08:26:32 crc kubenswrapper[4860]: I1211 08:26:32.879846 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fgp9c8" Dec 11 08:26:33 crc kubenswrapper[4860]: E1211 08:26:33.274846 4860 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/glance-operator@sha256:5370dc4a8e776923eec00bb50cbdb2e390e9dde50be26bdc04a216bd2d6b5027" Dec 11 08:26:33 crc kubenswrapper[4860]: E1211 08:26:33.275087 4860 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/glance-operator@sha256:5370dc4a8e776923eec00bb50cbdb2e390e9dde50be26bdc04a216bd2d6b5027,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-xl5z4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-operator-controller-manager-5697bb5779-kgzwt_openstack-operators(e00c360d-20a4-4e2e-81ea-e4ffa1bc62fc): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 11 08:26:33 crc kubenswrapper[4860]: I1211 08:26:33.338047 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2425r" Dec 11 08:26:33 crc kubenswrapper[4860]: I1211 08:26:33.345230 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/4bbd2122-dfb6-486b-8196-375ffeb93527-webhook-certs\") pod \"openstack-operator-controller-manager-6d784644c6-68pll\" (UID: \"4bbd2122-dfb6-486b-8196-375ffeb93527\") " pod="openstack-operators/openstack-operator-controller-manager-6d784644c6-68pll" Dec 11 08:26:33 crc kubenswrapper[4860]: I1211 08:26:33.345327 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4bbd2122-dfb6-486b-8196-375ffeb93527-metrics-certs\") pod \"openstack-operator-controller-manager-6d784644c6-68pll\" (UID: \"4bbd2122-dfb6-486b-8196-375ffeb93527\") " pod="openstack-operators/openstack-operator-controller-manager-6d784644c6-68pll" Dec 11 08:26:33 crc kubenswrapper[4860]: E1211 08:26:33.345479 4860 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 11 08:26:33 crc kubenswrapper[4860]: E1211 08:26:33.345616 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4bbd2122-dfb6-486b-8196-375ffeb93527-webhook-certs podName:4bbd2122-dfb6-486b-8196-375ffeb93527 nodeName:}" failed. No retries permitted until 2025-12-11 08:26:49.345585549 +0000 UTC m=+942.074104764 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/4bbd2122-dfb6-486b-8196-375ffeb93527-webhook-certs") pod "openstack-operator-controller-manager-6d784644c6-68pll" (UID: "4bbd2122-dfb6-486b-8196-375ffeb93527") : secret "webhook-server-cert" not found Dec 11 08:26:33 crc kubenswrapper[4860]: I1211 08:26:33.350968 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4bbd2122-dfb6-486b-8196-375ffeb93527-metrics-certs\") pod \"openstack-operator-controller-manager-6d784644c6-68pll\" (UID: \"4bbd2122-dfb6-486b-8196-375ffeb93527\") " pod="openstack-operators/openstack-operator-controller-manager-6d784644c6-68pll" Dec 11 08:26:33 crc kubenswrapper[4860]: I1211 08:26:33.448238 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/16fb4ab2-190c-4a33-8ba1-f9cd5edaaff3-utilities\") pod \"16fb4ab2-190c-4a33-8ba1-f9cd5edaaff3\" (UID: \"16fb4ab2-190c-4a33-8ba1-f9cd5edaaff3\") " Dec 11 08:26:33 crc kubenswrapper[4860]: I1211 08:26:33.448422 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pfk4r\" (UniqueName: \"kubernetes.io/projected/16fb4ab2-190c-4a33-8ba1-f9cd5edaaff3-kube-api-access-pfk4r\") pod \"16fb4ab2-190c-4a33-8ba1-f9cd5edaaff3\" (UID: \"16fb4ab2-190c-4a33-8ba1-f9cd5edaaff3\") " Dec 11 08:26:33 crc kubenswrapper[4860]: I1211 08:26:33.448467 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/16fb4ab2-190c-4a33-8ba1-f9cd5edaaff3-catalog-content\") pod \"16fb4ab2-190c-4a33-8ba1-f9cd5edaaff3\" (UID: \"16fb4ab2-190c-4a33-8ba1-f9cd5edaaff3\") " Dec 11 08:26:33 crc kubenswrapper[4860]: I1211 08:26:33.450172 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/16fb4ab2-190c-4a33-8ba1-f9cd5edaaff3-utilities" (OuterVolumeSpecName: "utilities") pod "16fb4ab2-190c-4a33-8ba1-f9cd5edaaff3" (UID: "16fb4ab2-190c-4a33-8ba1-f9cd5edaaff3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:26:33 crc kubenswrapper[4860]: I1211 08:26:33.452807 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/16fb4ab2-190c-4a33-8ba1-f9cd5edaaff3-kube-api-access-pfk4r" (OuterVolumeSpecName: "kube-api-access-pfk4r") pod "16fb4ab2-190c-4a33-8ba1-f9cd5edaaff3" (UID: "16fb4ab2-190c-4a33-8ba1-f9cd5edaaff3"). InnerVolumeSpecName "kube-api-access-pfk4r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:26:33 crc kubenswrapper[4860]: I1211 08:26:33.470447 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/16fb4ab2-190c-4a33-8ba1-f9cd5edaaff3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "16fb4ab2-190c-4a33-8ba1-f9cd5edaaff3" (UID: "16fb4ab2-190c-4a33-8ba1-f9cd5edaaff3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:26:33 crc kubenswrapper[4860]: I1211 08:26:33.549874 4860 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/16fb4ab2-190c-4a33-8ba1-f9cd5edaaff3-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 08:26:33 crc kubenswrapper[4860]: I1211 08:26:33.549919 4860 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/16fb4ab2-190c-4a33-8ba1-f9cd5edaaff3-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 08:26:33 crc kubenswrapper[4860]: I1211 08:26:33.549930 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pfk4r\" (UniqueName: \"kubernetes.io/projected/16fb4ab2-190c-4a33-8ba1-f9cd5edaaff3-kube-api-access-pfk4r\") on node \"crc\" DevicePath \"\"" Dec 11 08:26:33 crc kubenswrapper[4860]: I1211 08:26:33.741980 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2425r" event={"ID":"16fb4ab2-190c-4a33-8ba1-f9cd5edaaff3","Type":"ContainerDied","Data":"e61abd02a3bb306d6cc6fc6c51606cdc3d963e90dda8cd22af18d5f7fb63fdfb"} Dec 11 08:26:33 crc kubenswrapper[4860]: I1211 08:26:33.742059 4860 scope.go:117] "RemoveContainer" containerID="e35bb40199af54a9bbd197edd80583cfb35a628ea0965a63590700958977af63" Dec 11 08:26:33 crc kubenswrapper[4860]: I1211 08:26:33.742232 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2425r" Dec 11 08:26:33 crc kubenswrapper[4860]: I1211 08:26:33.775571 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2425r"] Dec 11 08:26:33 crc kubenswrapper[4860]: I1211 08:26:33.783690 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-2425r"] Dec 11 08:26:34 crc kubenswrapper[4860]: E1211 08:26:34.007773 4860 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670" Dec 11 08:26:34 crc kubenswrapper[4860]: E1211 08:26:34.008138 4860 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-cs4kf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-697bc559fc-zrhjm_openstack-operators(5aff136d-8f48-4cb7-8ddc-72ce966607b7): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 11 08:26:34 crc kubenswrapper[4860]: E1211 08:26:34.499779 4860 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7" Dec 11 08:26:34 crc kubenswrapper[4860]: E1211 08:26:34.500104 4860 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5ldm2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-7765d96ddf-6bpb9_openstack-operators(d5e0079f-f500-407f-b06f-a610a19c741b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 11 08:26:35 crc kubenswrapper[4860]: I1211 08:26:35.595801 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="16fb4ab2-190c-4a33-8ba1-f9cd5edaaff3" path="/var/lib/kubelet/pods/16fb4ab2-190c-4a33-8ba1-f9cd5edaaff3/volumes" Dec 11 08:26:41 crc kubenswrapper[4860]: I1211 08:26:41.763909 4860 scope.go:117] "RemoveContainer" containerID="0f5fda4090970b2ecd060ce1226877e892c204b6e116470c09d46001ffb4cace" Dec 11 08:26:41 crc kubenswrapper[4860]: I1211 08:26:41.995400 4860 scope.go:117] "RemoveContainer" containerID="6fb1acd3dcd395508a11361e36797af7415d63f52aac61bc11319cf68de84664" Dec 11 08:26:42 crc kubenswrapper[4860]: I1211 08:26:42.544956 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-78d48bff9d-sq5gd"] Dec 11 08:26:42 crc kubenswrapper[4860]: I1211 08:26:42.840666 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fgp9c8"] Dec 11 08:26:43 crc kubenswrapper[4860]: E1211 08:26:43.120534 4860 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 11 08:26:43 crc kubenswrapper[4860]: E1211 08:26:43.120785 4860 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-jdhtm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-b6456fdb6-nvt56_openstack-operators(99f7e97f-7cea-4785-b95e-133adf238ac3): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 11 08:26:43 crc kubenswrapper[4860]: E1211 08:26:43.122024 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"]" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-nvt56" podUID="99f7e97f-7cea-4785-b95e-133adf238ac3" Dec 11 08:26:43 crc kubenswrapper[4860]: I1211 08:26:43.849558 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fgp9c8" event={"ID":"04044c38-9aa4-4439-bad4-cf687e00ed5a","Type":"ContainerStarted","Data":"e58d1e761440d7b0f68afb95120c9d23ad5a0aebf3e18190f66f2da9c951774c"} Dec 11 08:26:43 crc kubenswrapper[4860]: I1211 08:26:43.851792 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-sq5gd" event={"ID":"7f1d9cd2-1e0c-4c4d-86a6-b91a38eadba2","Type":"ContainerStarted","Data":"f4fee6709da1fd33170cfe9d4cd7c42e2372ce3c843b80328dba0c5f68ddb375"} Dec 11 08:26:46 crc kubenswrapper[4860]: I1211 08:26:46.883442 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-967d97867-5wvld" event={"ID":"91bb3ff9-fd19-4c79-bac4-621cb114c783","Type":"ContainerStarted","Data":"1d160c487d6e9119a8ac68e6857488d091970c39384e1d3af19937c3a21230bf"} Dec 11 08:26:47 crc kubenswrapper[4860]: I1211 08:26:47.918362 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-zrrf6" event={"ID":"323ad77f-a0d6-461d-b483-2ffd74ebb9b7","Type":"ContainerStarted","Data":"c6a866e094fd379be08e9085ce4e57c2e6b5a96ebe02fee796267a70486bed19"} Dec 11 08:26:47 crc kubenswrapper[4860]: I1211 08:26:47.930815 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-65fv4" event={"ID":"bb231e77-e19b-41ed-bf7a-ae354e1089ec","Type":"ContainerStarted","Data":"835e28b7a4bcdea9522d2ddd0dc32c11fa592a3980e0976900e6572c790eb55f"} Dec 11 08:26:48 crc kubenswrapper[4860]: I1211 08:26:48.943232 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-lpm8h" event={"ID":"aae650b9-dcbc-482b-b49f-0a45651bb4b4","Type":"ContainerStarted","Data":"cd8c79b699ad1f3636401e3089d6a2e65f53dbc504f2c4162df7fb9ea38cf543"} Dec 11 08:26:48 crc kubenswrapper[4860]: I1211 08:26:48.947337 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-bshq2" event={"ID":"64f09777-fca3-412f-98f8-5cd189cb9fbd","Type":"ContainerStarted","Data":"73a6e9c173ab6db8408fcb5385c85e90b768c9aae40066e791c51f2921f0044d"} Dec 11 08:26:48 crc kubenswrapper[4860]: I1211 08:26:48.949048 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-lrhhk" event={"ID":"505d980f-cfcb-42b2-876e-a4730abe7ea6","Type":"ContainerStarted","Data":"4e65fe5e7057ab7bf71c1e67c24b51cdf826d3c3a1322e6b6626ddd1422a8d27"} Dec 11 08:26:48 crc kubenswrapper[4860]: I1211 08:26:48.951990 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-bvs87" event={"ID":"689923af-60de-49e9-bbed-bebaa63adc4b","Type":"ContainerStarted","Data":"c12653dc9c32b1dcd9c74c6d96bb8e88a94a64891bbb36e52549752f4e2dba20"} Dec 11 08:26:49 crc kubenswrapper[4860]: I1211 08:26:49.362240 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/4bbd2122-dfb6-486b-8196-375ffeb93527-webhook-certs\") pod \"openstack-operator-controller-manager-6d784644c6-68pll\" (UID: \"4bbd2122-dfb6-486b-8196-375ffeb93527\") " pod="openstack-operators/openstack-operator-controller-manager-6d784644c6-68pll" Dec 11 08:26:49 crc kubenswrapper[4860]: I1211 08:26:49.370634 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/4bbd2122-dfb6-486b-8196-375ffeb93527-webhook-certs\") pod \"openstack-operator-controller-manager-6d784644c6-68pll\" (UID: \"4bbd2122-dfb6-486b-8196-375ffeb93527\") " pod="openstack-operators/openstack-operator-controller-manager-6d784644c6-68pll" Dec 11 08:26:49 crc kubenswrapper[4860]: I1211 08:26:49.514287 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-6d784644c6-68pll" Dec 11 08:26:49 crc kubenswrapper[4860]: I1211 08:26:49.961309 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-4tw58" event={"ID":"5a589b13-7f29-40ae-9804-7c5c2bc3d745","Type":"ContainerStarted","Data":"f08792826a8ed768e4acf1c1e2089022a37d8ee19a84de39a4a3c7758ac337c7"} Dec 11 08:26:50 crc kubenswrapper[4860]: I1211 08:26:50.969743 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-6t74n" event={"ID":"a3bf483c-6438-4b7e-abb8-447b9b146644","Type":"ContainerStarted","Data":"1270f0f59e5c5ef879fc990348e2c2fee07e9ba4555225a0d7474c8b7807ccff"} Dec 11 08:26:52 crc kubenswrapper[4860]: I1211 08:26:52.339497 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-6d784644c6-68pll"] Dec 11 08:26:52 crc kubenswrapper[4860]: W1211 08:26:52.408123 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4bbd2122_dfb6_486b_8196_375ffeb93527.slice/crio-36b9b50f9bd55f7fa850efcb33e6cabdf1b42d3850af7ad878d02251f6abbbe2 WatchSource:0}: Error finding container 36b9b50f9bd55f7fa850efcb33e6cabdf1b42d3850af7ad878d02251f6abbbe2: Status 404 returned error can't find the container with id 36b9b50f9bd55f7fa850efcb33e6cabdf1b42d3850af7ad878d02251f6abbbe2 Dec 11 08:26:53 crc kubenswrapper[4860]: I1211 08:26:53.014155 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-2qmxt" event={"ID":"a52f8df6-7c28-4776-aca5-bd8b47e82fe8","Type":"ContainerStarted","Data":"0e21a1dd18a5a5373d48801ccd3956435bdd4fe899985c4cd1b6cdbae4195551"} Dec 11 08:26:53 crc kubenswrapper[4860]: I1211 08:26:53.035998 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-sq5gd" event={"ID":"7f1d9cd2-1e0c-4c4d-86a6-b91a38eadba2","Type":"ContainerStarted","Data":"646d2aa815171a5bf71ac8392790b8cb034800651e91cdda41999236a297ee25"} Dec 11 08:26:53 crc kubenswrapper[4860]: I1211 08:26:53.070448 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-967d97867-5wvld" event={"ID":"91bb3ff9-fd19-4c79-bac4-621cb114c783","Type":"ContainerStarted","Data":"25889c87a7fe63df723ba501fa31fd2398c93db67d7c4686e1569d9a37340eb6"} Dec 11 08:26:53 crc kubenswrapper[4860]: I1211 08:26:53.072886 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-967d97867-5wvld" Dec 11 08:26:53 crc kubenswrapper[4860]: I1211 08:26:53.078346 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-967d97867-5wvld" Dec 11 08:26:53 crc kubenswrapper[4860]: I1211 08:26:53.087919 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-jsftr" event={"ID":"22f3c8f5-1781-4ba2-8610-c9ada5539d92","Type":"ContainerStarted","Data":"2e21fdd250df03fcbc49e3c765e6c5172e7b9b47bb8fc9e0e37992644c2ee773"} Dec 11 08:26:53 crc kubenswrapper[4860]: I1211 08:26:53.109085 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fgp9c8" event={"ID":"04044c38-9aa4-4439-bad4-cf687e00ed5a","Type":"ContainerStarted","Data":"0cdd66000c8fc022fe2e39b396c423a76aac7ef49218bd8b16f76ca3fa034307"} Dec 11 08:26:53 crc kubenswrapper[4860]: I1211 08:26:53.132882 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-ht7db" event={"ID":"dc79abc2-7b6c-46c4-b9b4-a9c76696a8b2","Type":"ContainerStarted","Data":"262a165fe0cbe224867d2a5409b33372b2149605491679bd362b39895a5ec7f9"} Dec 11 08:26:53 crc kubenswrapper[4860]: I1211 08:26:53.150607 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-75944c9b7-ck9nc" event={"ID":"8462b48d-cc3d-4714-9558-22818db99c08","Type":"ContainerStarted","Data":"6b8aafad0198330b706b65f67c99a8b0b15f577320b2ffa79b7c8f8847637736"} Dec 11 08:26:53 crc kubenswrapper[4860]: I1211 08:26:53.175458 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-967d97867-5wvld" podStartSLOduration=13.448865524 podStartE2EDuration="37.175419588s" podCreationTimestamp="2025-12-11 08:26:16 +0000 UTC" firstStartedPulling="2025-12-11 08:26:18.1659144 +0000 UTC m=+910.894433675" lastFinishedPulling="2025-12-11 08:26:41.892468684 +0000 UTC m=+934.620987739" observedRunningTime="2025-12-11 08:26:53.161429008 +0000 UTC m=+945.889948063" watchObservedRunningTime="2025-12-11 08:26:53.175419588 +0000 UTC m=+945.903938643" Dec 11 08:26:53 crc kubenswrapper[4860]: I1211 08:26:53.175983 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-fh5vc" event={"ID":"b6c38bbc-80d0-449a-aab0-291078361ebd","Type":"ContainerStarted","Data":"60ecc99b65df0d485bc7229c01dedd77b80eccdb12e6719758d32813f99f6eb4"} Dec 11 08:26:53 crc kubenswrapper[4860]: I1211 08:26:53.188867 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-hrpn8" event={"ID":"e36550af-98a6-49b3-9769-bd52d0da2838","Type":"ContainerStarted","Data":"f3388cc33b865b6327c9a6c83ee6e2af048c156e12b382108cb559ea81526496"} Dec 11 08:26:53 crc kubenswrapper[4860]: I1211 08:26:53.207842 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-6d784644c6-68pll" event={"ID":"4bbd2122-dfb6-486b-8196-375ffeb93527","Type":"ContainerStarted","Data":"36b9b50f9bd55f7fa850efcb33e6cabdf1b42d3850af7ad878d02251f6abbbe2"} Dec 11 08:26:53 crc kubenswrapper[4860]: I1211 08:26:53.215242 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-jsftr" podStartSLOduration=3.700573329 podStartE2EDuration="37.215219222s" podCreationTimestamp="2025-12-11 08:26:16 +0000 UTC" firstStartedPulling="2025-12-11 08:26:18.584274581 +0000 UTC m=+911.312793636" lastFinishedPulling="2025-12-11 08:26:52.098920474 +0000 UTC m=+944.827439529" observedRunningTime="2025-12-11 08:26:53.214431307 +0000 UTC m=+945.942950362" watchObservedRunningTime="2025-12-11 08:26:53.215219222 +0000 UTC m=+945.943738267" Dec 11 08:26:53 crc kubenswrapper[4860]: E1211 08:26:53.645753 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-zrhjm" podUID="5aff136d-8f48-4cb7-8ddc-72ce966607b7" Dec 11 08:26:53 crc kubenswrapper[4860]: E1211 08:26:53.664285 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-kgzwt" podUID="e00c360d-20a4-4e2e-81ea-e4ffa1bc62fc" Dec 11 08:26:53 crc kubenswrapper[4860]: E1211 08:26:53.664881 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-6bpb9" podUID="d5e0079f-f500-407f-b06f-a610a19c741b" Dec 11 08:26:54 crc kubenswrapper[4860]: I1211 08:26:54.248010 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-fh5vc" event={"ID":"b6c38bbc-80d0-449a-aab0-291078361ebd","Type":"ContainerStarted","Data":"915836e0b82a55800591aeb5e919c9e6b92dc39a0b653360aaa3295be2cbcf8b"} Dec 11 08:26:54 crc kubenswrapper[4860]: I1211 08:26:54.248826 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5854674fcc-fh5vc" Dec 11 08:26:54 crc kubenswrapper[4860]: I1211 08:26:54.259377 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-kgzwt" event={"ID":"e00c360d-20a4-4e2e-81ea-e4ffa1bc62fc","Type":"ContainerStarted","Data":"34f47fd12546192427de93592fb4ef2e550166ce9ff96dae344c9601f4d196ff"} Dec 11 08:26:54 crc kubenswrapper[4860]: I1211 08:26:54.273374 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-lpm8h" event={"ID":"aae650b9-dcbc-482b-b49f-0a45651bb4b4","Type":"ContainerStarted","Data":"aa27cdcd64ccd570d239c78d1925383fbbce11765436378faa1c19cad92e6223"} Dec 11 08:26:54 crc kubenswrapper[4860]: I1211 08:26:54.274333 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-lpm8h" Dec 11 08:26:54 crc kubenswrapper[4860]: E1211 08:26:54.276085 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-p7dkp" podUID="211d9c9c-f584-43ca-8db1-7b81f6307c21" Dec 11 08:26:54 crc kubenswrapper[4860]: I1211 08:26:54.300900 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-lpm8h" Dec 11 08:26:54 crc kubenswrapper[4860]: I1211 08:26:54.302252 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5854674fcc-fh5vc" podStartSLOduration=7.99259278 podStartE2EDuration="38.302229348s" podCreationTimestamp="2025-12-11 08:26:16 +0000 UTC" firstStartedPulling="2025-12-11 08:26:18.546421229 +0000 UTC m=+911.274940284" lastFinishedPulling="2025-12-11 08:26:48.856057797 +0000 UTC m=+941.584576852" observedRunningTime="2025-12-11 08:26:54.301298059 +0000 UTC m=+947.029817104" watchObservedRunningTime="2025-12-11 08:26:54.302229348 +0000 UTC m=+947.030748403" Dec 11 08:26:54 crc kubenswrapper[4860]: I1211 08:26:54.308757 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-6bpb9" event={"ID":"d5e0079f-f500-407f-b06f-a610a19c741b","Type":"ContainerStarted","Data":"ea92ecc72b14620c5d3ea2daa4aa158845a4721cb3f582b42da91e3b69d69b84"} Dec 11 08:26:54 crc kubenswrapper[4860]: I1211 08:26:54.323859 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-zrhjm" event={"ID":"5aff136d-8f48-4cb7-8ddc-72ce966607b7","Type":"ContainerStarted","Data":"55cdba31bffc42b3316248a934590a8702446cd6612f5014228cb3577852149c"} Dec 11 08:26:54 crc kubenswrapper[4860]: I1211 08:26:54.346375 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-lrhhk" event={"ID":"505d980f-cfcb-42b2-876e-a4730abe7ea6","Type":"ContainerStarted","Data":"93650ecbca77c93229a0e12618c8e6c937b6970dc808006a4cddc00e07b65275"} Dec 11 08:26:54 crc kubenswrapper[4860]: I1211 08:26:54.347412 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-lrhhk" Dec 11 08:26:54 crc kubenswrapper[4860]: I1211 08:26:54.363830 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-lpm8h" podStartSLOduration=14.544007506 podStartE2EDuration="38.363808098s" podCreationTimestamp="2025-12-11 08:26:16 +0000 UTC" firstStartedPulling="2025-12-11 08:26:18.149963897 +0000 UTC m=+910.878482942" lastFinishedPulling="2025-12-11 08:26:41.969764479 +0000 UTC m=+934.698283534" observedRunningTime="2025-12-11 08:26:54.333161183 +0000 UTC m=+947.061680248" watchObservedRunningTime="2025-12-11 08:26:54.363808098 +0000 UTC m=+947.092327153" Dec 11 08:26:54 crc kubenswrapper[4860]: I1211 08:26:54.370304 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-2qmxt" event={"ID":"a52f8df6-7c28-4776-aca5-bd8b47e82fe8","Type":"ContainerStarted","Data":"25cf70dacef404a11846d3727d2ad6d8835c1c0ddba66fe0ef518078763d3481"} Dec 11 08:26:54 crc kubenswrapper[4860]: I1211 08:26:54.371453 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-78f8948974-2qmxt" Dec 11 08:26:54 crc kubenswrapper[4860]: I1211 08:26:54.394428 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-lrhhk" Dec 11 08:26:54 crc kubenswrapper[4860]: I1211 08:26:54.407723 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-65fv4" event={"ID":"bb231e77-e19b-41ed-bf7a-ae354e1089ec","Type":"ContainerStarted","Data":"f0c0037b06ecb1d22e5f4f338c1f3f8661a184b208a6400c53b43697a6686d04"} Dec 11 08:26:54 crc kubenswrapper[4860]: I1211 08:26:54.408999 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-65fv4" Dec 11 08:26:54 crc kubenswrapper[4860]: I1211 08:26:54.444470 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-65fv4" Dec 11 08:26:54 crc kubenswrapper[4860]: I1211 08:26:54.451340 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-6d784644c6-68pll" event={"ID":"4bbd2122-dfb6-486b-8196-375ffeb93527","Type":"ContainerStarted","Data":"642fff4ab4d7e034da87615b39bc8d3e257cc6216bf8d679956a4481ecc17d83"} Dec 11 08:26:54 crc kubenswrapper[4860]: I1211 08:26:54.452329 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-6d784644c6-68pll" Dec 11 08:26:54 crc kubenswrapper[4860]: I1211 08:26:54.470227 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-lrhhk" podStartSLOduration=14.652590898 podStartE2EDuration="38.47020077s" podCreationTimestamp="2025-12-11 08:26:16 +0000 UTC" firstStartedPulling="2025-12-11 08:26:18.153972384 +0000 UTC m=+910.882491439" lastFinishedPulling="2025-12-11 08:26:41.971582226 +0000 UTC m=+934.700101311" observedRunningTime="2025-12-11 08:26:54.452244294 +0000 UTC m=+947.180763349" watchObservedRunningTime="2025-12-11 08:26:54.47020077 +0000 UTC m=+947.198719825" Dec 11 08:26:54 crc kubenswrapper[4860]: I1211 08:26:54.500021 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-bshq2" event={"ID":"64f09777-fca3-412f-98f8-5cd189cb9fbd","Type":"ContainerStarted","Data":"0be54578063888877058b15208da52fee8f50e8d81f74bbc702c634b8f4540d7"} Dec 11 08:26:54 crc kubenswrapper[4860]: I1211 08:26:54.501254 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-bshq2" Dec 11 08:26:54 crc kubenswrapper[4860]: I1211 08:26:54.522237 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-bshq2" Dec 11 08:26:54 crc kubenswrapper[4860]: I1211 08:26:54.530807 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-78f8948974-2qmxt" podStartSLOduration=9.059941154 podStartE2EDuration="38.530779358s" podCreationTimestamp="2025-12-11 08:26:16 +0000 UTC" firstStartedPulling="2025-12-11 08:26:18.237905118 +0000 UTC m=+910.966424173" lastFinishedPulling="2025-12-11 08:26:47.708743322 +0000 UTC m=+940.437262377" observedRunningTime="2025-12-11 08:26:54.519673549 +0000 UTC m=+947.248192604" watchObservedRunningTime="2025-12-11 08:26:54.530779358 +0000 UTC m=+947.259298413" Dec 11 08:26:54 crc kubenswrapper[4860]: I1211 08:26:54.545720 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-75944c9b7-ck9nc" Dec 11 08:26:54 crc kubenswrapper[4860]: I1211 08:26:54.563850 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-nvt56" event={"ID":"99f7e97f-7cea-4785-b95e-133adf238ac3","Type":"ContainerStarted","Data":"c746536367a9ac29fc522dfbb969ed758a48ddfe63f6c064b6abf830f3024697"} Dec 11 08:26:54 crc kubenswrapper[4860]: I1211 08:26:54.564936 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-nvt56" Dec 11 08:26:54 crc kubenswrapper[4860]: I1211 08:26:54.585821 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-4tw58" event={"ID":"5a589b13-7f29-40ae-9804-7c5c2bc3d745","Type":"ContainerStarted","Data":"7e41116b1a103c759bfefc34cb9feb85300215edaad429166a7dfb6348792224"} Dec 11 08:26:54 crc kubenswrapper[4860]: I1211 08:26:54.586381 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-4tw58" Dec 11 08:26:54 crc kubenswrapper[4860]: I1211 08:26:54.598381 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-4tw58" Dec 11 08:26:54 crc kubenswrapper[4860]: I1211 08:26:54.632851 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-65fv4" podStartSLOduration=14.967653475 podStartE2EDuration="38.632819514s" podCreationTimestamp="2025-12-11 08:26:16 +0000 UTC" firstStartedPulling="2025-12-11 08:26:18.225327802 +0000 UTC m=+910.953846857" lastFinishedPulling="2025-12-11 08:26:41.890493841 +0000 UTC m=+934.619012896" observedRunningTime="2025-12-11 08:26:54.581570319 +0000 UTC m=+947.310089394" watchObservedRunningTime="2025-12-11 08:26:54.632819514 +0000 UTC m=+947.361338569" Dec 11 08:26:54 crc kubenswrapper[4860]: I1211 08:26:54.665471 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-bshq2" podStartSLOduration=15.863231304 podStartE2EDuration="39.665444381s" podCreationTimestamp="2025-12-11 08:26:15 +0000 UTC" firstStartedPulling="2025-12-11 08:26:18.169864204 +0000 UTC m=+910.898383259" lastFinishedPulling="2025-12-11 08:26:41.972077281 +0000 UTC m=+934.700596336" observedRunningTime="2025-12-11 08:26:54.658373009 +0000 UTC m=+947.386892064" watchObservedRunningTime="2025-12-11 08:26:54.665444381 +0000 UTC m=+947.393963436" Dec 11 08:26:54 crc kubenswrapper[4860]: I1211 08:26:54.783488 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-6d784644c6-68pll" podStartSLOduration=38.78346244 podStartE2EDuration="38.78346244s" podCreationTimestamp="2025-12-11 08:26:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:26:54.76539715 +0000 UTC m=+947.493916205" watchObservedRunningTime="2025-12-11 08:26:54.78346244 +0000 UTC m=+947.511981495" Dec 11 08:26:54 crc kubenswrapper[4860]: I1211 08:26:54.831881 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-4tw58" podStartSLOduration=14.989441279 podStartE2EDuration="38.831853134s" podCreationTimestamp="2025-12-11 08:26:16 +0000 UTC" firstStartedPulling="2025-12-11 08:26:18.127799878 +0000 UTC m=+910.856318933" lastFinishedPulling="2025-12-11 08:26:41.970211733 +0000 UTC m=+934.698730788" observedRunningTime="2025-12-11 08:26:54.824950126 +0000 UTC m=+947.553469181" watchObservedRunningTime="2025-12-11 08:26:54.831853134 +0000 UTC m=+947.560372189" Dec 11 08:26:54 crc kubenswrapper[4860]: I1211 08:26:54.871016 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-75944c9b7-ck9nc" podStartSLOduration=9.75554049 podStartE2EDuration="38.870987577s" podCreationTimestamp="2025-12-11 08:26:16 +0000 UTC" firstStartedPulling="2025-12-11 08:26:18.594458002 +0000 UTC m=+911.322977057" lastFinishedPulling="2025-12-11 08:26:47.709905089 +0000 UTC m=+940.438424144" observedRunningTime="2025-12-11 08:26:54.864810892 +0000 UTC m=+947.593329947" watchObservedRunningTime="2025-12-11 08:26:54.870987577 +0000 UTC m=+947.599506632" Dec 11 08:26:54 crc kubenswrapper[4860]: I1211 08:26:54.896718 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-nvt56" podStartSLOduration=5.43492082 podStartE2EDuration="38.896690327s" podCreationTimestamp="2025-12-11 08:26:16 +0000 UTC" firstStartedPulling="2025-12-11 08:26:18.549304539 +0000 UTC m=+911.277823594" lastFinishedPulling="2025-12-11 08:26:52.011074046 +0000 UTC m=+944.739593101" observedRunningTime="2025-12-11 08:26:54.891308527 +0000 UTC m=+947.619827582" watchObservedRunningTime="2025-12-11 08:26:54.896690327 +0000 UTC m=+947.625209382" Dec 11 08:26:55 crc kubenswrapper[4860]: I1211 08:26:55.596697 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fgp9c8" event={"ID":"04044c38-9aa4-4439-bad4-cf687e00ed5a","Type":"ContainerStarted","Data":"b1c28dc40cadb57c8b40cf78b492b1987c6bddd569841c4d78f468901dde9f3c"} Dec 11 08:26:55 crc kubenswrapper[4860]: I1211 08:26:55.597187 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fgp9c8" Dec 11 08:26:55 crc kubenswrapper[4860]: I1211 08:26:55.599280 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-p7dkp" event={"ID":"211d9c9c-f584-43ca-8db1-7b81f6307c21","Type":"ContainerStarted","Data":"3d28e80efa3d9364849b8ddd36cdb6eeddba440d4d81e3a8ad2d2c556bb65b27"} Dec 11 08:26:55 crc kubenswrapper[4860]: I1211 08:26:55.605774 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-sq5gd" event={"ID":"7f1d9cd2-1e0c-4c4d-86a6-b91a38eadba2","Type":"ContainerStarted","Data":"1e672c28eaa5d6fe72878c89b8bf22b52be03b8e69b984eea20b9287863c0945"} Dec 11 08:26:55 crc kubenswrapper[4860]: I1211 08:26:55.606790 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-sq5gd" Dec 11 08:26:55 crc kubenswrapper[4860]: I1211 08:26:55.609194 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-bvs87" event={"ID":"689923af-60de-49e9-bbed-bebaa63adc4b","Type":"ContainerStarted","Data":"65590d839d3b30c902362d0529425e9433ba955e652e3876d9493d7555573d02"} Dec 11 08:26:55 crc kubenswrapper[4860]: I1211 08:26:55.609349 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-998648c74-bvs87" Dec 11 08:26:55 crc kubenswrapper[4860]: I1211 08:26:55.612889 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-nvt56" event={"ID":"99f7e97f-7cea-4785-b95e-133adf238ac3","Type":"ContainerStarted","Data":"b434f9ae3cb3cc45a872d019cb90a5fd03daf87c13001206c0a53d5055012892"} Dec 11 08:26:55 crc kubenswrapper[4860]: I1211 08:26:55.614826 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-998648c74-bvs87" Dec 11 08:26:55 crc kubenswrapper[4860]: I1211 08:26:55.615417 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-6t74n" event={"ID":"a3bf483c-6438-4b7e-abb8-447b9b146644","Type":"ContainerStarted","Data":"2bd85778a2bd2e30a56e865d5690cb8e3a4359f85830065a6aeb4d853edeb0f7"} Dec 11 08:26:55 crc kubenswrapper[4860]: I1211 08:26:55.615988 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-6t74n" Dec 11 08:26:55 crc kubenswrapper[4860]: I1211 08:26:55.618496 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-kgzwt" event={"ID":"e00c360d-20a4-4e2e-81ea-e4ffa1bc62fc","Type":"ContainerStarted","Data":"aa5a66022962a6c5c2d0f004abca68bf45eba58018c8b251bfc5a33342bdee6f"} Dec 11 08:26:55 crc kubenswrapper[4860]: I1211 08:26:55.619022 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-kgzwt" Dec 11 08:26:55 crc kubenswrapper[4860]: I1211 08:26:55.621389 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-6t74n" Dec 11 08:26:55 crc kubenswrapper[4860]: I1211 08:26:55.621761 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-hrpn8" event={"ID":"e36550af-98a6-49b3-9769-bd52d0da2838","Type":"ContainerStarted","Data":"a12a55ccf33f558e4ab8b02c0dcc67e4eb2829a6ff328962cb18585c8884b515"} Dec 11 08:26:55 crc kubenswrapper[4860]: I1211 08:26:55.621980 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-hrpn8" Dec 11 08:26:55 crc kubenswrapper[4860]: I1211 08:26:55.624734 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-ht7db" event={"ID":"dc79abc2-7b6c-46c4-b9b4-a9c76696a8b2","Type":"ContainerStarted","Data":"1d54d77b2b35403cd698288f6145eb1293e9f8efa76d26c1201f70e3fc15c09e"} Dec 11 08:26:55 crc kubenswrapper[4860]: I1211 08:26:55.625300 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-ht7db" Dec 11 08:26:55 crc kubenswrapper[4860]: I1211 08:26:55.628026 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-6bpb9" event={"ID":"d5e0079f-f500-407f-b06f-a610a19c741b","Type":"ContainerStarted","Data":"60b34b41d2eb1a21d19b6a0b951c9108bf9bb2e47f4ec3b7f1ea4369f506ebf5"} Dec 11 08:26:55 crc kubenswrapper[4860]: I1211 08:26:55.630855 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-zrhjm" event={"ID":"5aff136d-8f48-4cb7-8ddc-72ce966607b7","Type":"ContainerStarted","Data":"71b0fd09898370bcfb452fab6d6bae03a454511345b1f77cf78dad699e1861c5"} Dec 11 08:26:55 crc kubenswrapper[4860]: I1211 08:26:55.631664 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-zrhjm" Dec 11 08:26:55 crc kubenswrapper[4860]: I1211 08:26:55.637919 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-75944c9b7-ck9nc" event={"ID":"8462b48d-cc3d-4714-9558-22818db99c08","Type":"ContainerStarted","Data":"1235f83a09b4c0144864c7a0ac7c25c61f20c2bc5e2ba4eacb4a3d6edead7567"} Dec 11 08:26:55 crc kubenswrapper[4860]: I1211 08:26:55.641354 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-zrrf6" event={"ID":"323ad77f-a0d6-461d-b483-2ffd74ebb9b7","Type":"ContainerStarted","Data":"0183f782f9fa7d6bba6028c99e027650855aa2dc24f2117f6175a95c174e9ea3"} Dec 11 08:26:55 crc kubenswrapper[4860]: I1211 08:26:55.642573 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-zrrf6" Dec 11 08:26:55 crc kubenswrapper[4860]: I1211 08:26:55.645561 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-zrrf6" Dec 11 08:26:55 crc kubenswrapper[4860]: I1211 08:26:55.653270 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fgp9c8" podStartSLOduration=31.523149486 podStartE2EDuration="39.653241021s" podCreationTimestamp="2025-12-11 08:26:16 +0000 UTC" firstStartedPulling="2025-12-11 08:26:43.132008015 +0000 UTC m=+935.860527070" lastFinishedPulling="2025-12-11 08:26:51.26209955 +0000 UTC m=+943.990618605" observedRunningTime="2025-12-11 08:26:55.652312942 +0000 UTC m=+948.380832027" watchObservedRunningTime="2025-12-11 08:26:55.653241021 +0000 UTC m=+948.381760066" Dec 11 08:26:55 crc kubenswrapper[4860]: I1211 08:26:55.679606 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-zrhjm" podStartSLOduration=2.629956869 podStartE2EDuration="39.679584271s" podCreationTimestamp="2025-12-11 08:26:16 +0000 UTC" firstStartedPulling="2025-12-11 08:26:18.141723217 +0000 UTC m=+910.870242272" lastFinishedPulling="2025-12-11 08:26:55.191350619 +0000 UTC m=+947.919869674" observedRunningTime="2025-12-11 08:26:55.674369847 +0000 UTC m=+948.402888892" watchObservedRunningTime="2025-12-11 08:26:55.679584271 +0000 UTC m=+948.408103326" Dec 11 08:26:55 crc kubenswrapper[4860]: I1211 08:26:55.701277 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-hrpn8" podStartSLOduration=7.037514488 podStartE2EDuration="39.701249684s" podCreationTimestamp="2025-12-11 08:26:16 +0000 UTC" firstStartedPulling="2025-12-11 08:26:18.595229556 +0000 UTC m=+911.323748611" lastFinishedPulling="2025-12-11 08:26:51.258964752 +0000 UTC m=+943.987483807" observedRunningTime="2025-12-11 08:26:55.70016669 +0000 UTC m=+948.428685755" watchObservedRunningTime="2025-12-11 08:26:55.701249684 +0000 UTC m=+948.429768739" Dec 11 08:26:55 crc kubenswrapper[4860]: I1211 08:26:55.811873 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-6bpb9" podStartSLOduration=2.9424069040000003 podStartE2EDuration="39.811841878s" podCreationTimestamp="2025-12-11 08:26:16 +0000 UTC" firstStartedPulling="2025-12-11 08:26:18.176438822 +0000 UTC m=+910.904957877" lastFinishedPulling="2025-12-11 08:26:55.045873796 +0000 UTC m=+947.774392851" observedRunningTime="2025-12-11 08:26:55.778324482 +0000 UTC m=+948.506843557" watchObservedRunningTime="2025-12-11 08:26:55.811841878 +0000 UTC m=+948.540360943" Dec 11 08:26:55 crc kubenswrapper[4860]: I1211 08:26:55.815695 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-ht7db" podStartSLOduration=6.793998435 podStartE2EDuration="39.815676218s" podCreationTimestamp="2025-12-11 08:26:16 +0000 UTC" firstStartedPulling="2025-12-11 08:26:18.237408352 +0000 UTC m=+910.965927407" lastFinishedPulling="2025-12-11 08:26:51.259086125 +0000 UTC m=+943.987605190" observedRunningTime="2025-12-11 08:26:55.810172975 +0000 UTC m=+948.538692030" watchObservedRunningTime="2025-12-11 08:26:55.815676218 +0000 UTC m=+948.544195283" Dec 11 08:26:55 crc kubenswrapper[4860]: I1211 08:26:55.843419 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-sq5gd" podStartSLOduration=30.941145359 podStartE2EDuration="39.843400882s" podCreationTimestamp="2025-12-11 08:26:16 +0000 UTC" firstStartedPulling="2025-12-11 08:26:43.127881084 +0000 UTC m=+935.856400139" lastFinishedPulling="2025-12-11 08:26:52.030136597 +0000 UTC m=+944.758655662" observedRunningTime="2025-12-11 08:26:55.839357165 +0000 UTC m=+948.567876230" watchObservedRunningTime="2025-12-11 08:26:55.843400882 +0000 UTC m=+948.571919937" Dec 11 08:26:55 crc kubenswrapper[4860]: I1211 08:26:55.861547 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-kgzwt" podStartSLOduration=3.191430549 podStartE2EDuration="39.861522793s" podCreationTimestamp="2025-12-11 08:26:16 +0000 UTC" firstStartedPulling="2025-12-11 08:26:18.177505926 +0000 UTC m=+910.906024981" lastFinishedPulling="2025-12-11 08:26:54.84759817 +0000 UTC m=+947.576117225" observedRunningTime="2025-12-11 08:26:55.858850289 +0000 UTC m=+948.587369364" watchObservedRunningTime="2025-12-11 08:26:55.861522793 +0000 UTC m=+948.590041848" Dec 11 08:26:55 crc kubenswrapper[4860]: I1211 08:26:55.884938 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-zrrf6" podStartSLOduration=15.783578149 podStartE2EDuration="39.8849026s" podCreationTimestamp="2025-12-11 08:26:16 +0000 UTC" firstStartedPulling="2025-12-11 08:26:17.773033692 +0000 UTC m=+910.501552747" lastFinishedPulling="2025-12-11 08:26:41.874358143 +0000 UTC m=+934.602877198" observedRunningTime="2025-12-11 08:26:55.876939538 +0000 UTC m=+948.605458593" watchObservedRunningTime="2025-12-11 08:26:55.8849026 +0000 UTC m=+948.613421645" Dec 11 08:26:55 crc kubenswrapper[4860]: I1211 08:26:55.959550 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-998648c74-bvs87" podStartSLOduration=16.305150401 podStartE2EDuration="39.959526811s" podCreationTimestamp="2025-12-11 08:26:16 +0000 UTC" firstStartedPulling="2025-12-11 08:26:18.236957308 +0000 UTC m=+910.965476353" lastFinishedPulling="2025-12-11 08:26:41.891333698 +0000 UTC m=+934.619852763" observedRunningTime="2025-12-11 08:26:55.938077665 +0000 UTC m=+948.666596740" watchObservedRunningTime="2025-12-11 08:26:55.959526811 +0000 UTC m=+948.688045866" Dec 11 08:26:55 crc kubenswrapper[4860]: I1211 08:26:55.973554 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-6t74n" podStartSLOduration=15.799479009 podStartE2EDuration="39.973509471s" podCreationTimestamp="2025-12-11 08:26:16 +0000 UTC" firstStartedPulling="2025-12-11 08:26:17.796300045 +0000 UTC m=+910.524819100" lastFinishedPulling="2025-12-11 08:26:41.970330507 +0000 UTC m=+934.698849562" observedRunningTime="2025-12-11 08:26:55.955505954 +0000 UTC m=+948.684025009" watchObservedRunningTime="2025-12-11 08:26:55.973509471 +0000 UTC m=+948.702028526" Dec 11 08:26:56 crc kubenswrapper[4860]: I1211 08:26:56.648161 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-6bpb9" Dec 11 08:26:56 crc kubenswrapper[4860]: I1211 08:26:56.650863 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-p7dkp" event={"ID":"211d9c9c-f584-43ca-8db1-7b81f6307c21","Type":"ContainerStarted","Data":"e598d8d07f95823e8822638fccdb16de7cdc89bbd0a6d497a856bbc3f9c34aba"} Dec 11 08:26:56 crc kubenswrapper[4860]: I1211 08:26:56.680729 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-p7dkp" podStartSLOduration=3.167836786 podStartE2EDuration="40.68069904s" podCreationTimestamp="2025-12-11 08:26:16 +0000 UTC" firstStartedPulling="2025-12-11 08:26:18.543689793 +0000 UTC m=+911.272208848" lastFinishedPulling="2025-12-11 08:26:56.056552047 +0000 UTC m=+948.785071102" observedRunningTime="2025-12-11 08:26:56.677583522 +0000 UTC m=+949.406102577" watchObservedRunningTime="2025-12-11 08:26:56.68069904 +0000 UTC m=+949.409218095" Dec 11 08:26:57 crc kubenswrapper[4860]: I1211 08:26:57.319368 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-78f8948974-2qmxt" Dec 11 08:26:57 crc kubenswrapper[4860]: I1211 08:26:57.362187 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-hrpn8" Dec 11 08:26:57 crc kubenswrapper[4860]: I1211 08:26:57.374098 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-p7dkp" Dec 11 08:26:57 crc kubenswrapper[4860]: I1211 08:26:57.457729 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5854674fcc-fh5vc" Dec 11 08:26:57 crc kubenswrapper[4860]: I1211 08:26:57.673218 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-ht7db" Dec 11 08:26:57 crc kubenswrapper[4860]: I1211 08:26:57.673663 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-sq5gd" Dec 11 08:26:57 crc kubenswrapper[4860]: I1211 08:26:57.700246 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-75944c9b7-ck9nc" Dec 11 08:26:59 crc kubenswrapper[4860]: I1211 08:26:59.520887 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-6d784644c6-68pll" Dec 11 08:27:02 crc kubenswrapper[4860]: I1211 08:27:02.887415 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fgp9c8" Dec 11 08:27:06 crc kubenswrapper[4860]: I1211 08:27:06.652341 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-6bpb9" Dec 11 08:27:06 crc kubenswrapper[4860]: I1211 08:27:06.739929 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-kgzwt" Dec 11 08:27:06 crc kubenswrapper[4860]: I1211 08:27:06.913851 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-zrhjm" Dec 11 08:27:07 crc kubenswrapper[4860]: I1211 08:27:07.341950 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-nvt56" Dec 11 08:27:07 crc kubenswrapper[4860]: I1211 08:27:07.377290 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-p7dkp" Dec 11 08:27:21 crc kubenswrapper[4860]: I1211 08:27:21.399088 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-w78r8"] Dec 11 08:27:21 crc kubenswrapper[4860]: E1211 08:27:21.400364 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16fb4ab2-190c-4a33-8ba1-f9cd5edaaff3" containerName="extract-content" Dec 11 08:27:21 crc kubenswrapper[4860]: I1211 08:27:21.400381 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="16fb4ab2-190c-4a33-8ba1-f9cd5edaaff3" containerName="extract-content" Dec 11 08:27:21 crc kubenswrapper[4860]: E1211 08:27:21.400408 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16fb4ab2-190c-4a33-8ba1-f9cd5edaaff3" containerName="extract-utilities" Dec 11 08:27:21 crc kubenswrapper[4860]: I1211 08:27:21.400415 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="16fb4ab2-190c-4a33-8ba1-f9cd5edaaff3" containerName="extract-utilities" Dec 11 08:27:21 crc kubenswrapper[4860]: E1211 08:27:21.400427 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16fb4ab2-190c-4a33-8ba1-f9cd5edaaff3" containerName="registry-server" Dec 11 08:27:21 crc kubenswrapper[4860]: I1211 08:27:21.400436 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="16fb4ab2-190c-4a33-8ba1-f9cd5edaaff3" containerName="registry-server" Dec 11 08:27:21 crc kubenswrapper[4860]: I1211 08:27:21.400594 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="16fb4ab2-190c-4a33-8ba1-f9cd5edaaff3" containerName="registry-server" Dec 11 08:27:21 crc kubenswrapper[4860]: I1211 08:27:21.401545 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-w78r8" Dec 11 08:27:21 crc kubenswrapper[4860]: I1211 08:27:21.404574 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Dec 11 08:27:21 crc kubenswrapper[4860]: I1211 08:27:21.405837 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Dec 11 08:27:21 crc kubenswrapper[4860]: I1211 08:27:21.406147 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Dec 11 08:27:21 crc kubenswrapper[4860]: I1211 08:27:21.407609 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-5rksw" Dec 11 08:27:21 crc kubenswrapper[4860]: I1211 08:27:21.420218 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-w78r8"] Dec 11 08:27:21 crc kubenswrapper[4860]: I1211 08:27:21.489921 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-wt4jr"] Dec 11 08:27:21 crc kubenswrapper[4860]: I1211 08:27:21.491424 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-wt4jr" Dec 11 08:27:21 crc kubenswrapper[4860]: I1211 08:27:21.493526 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Dec 11 08:27:21 crc kubenswrapper[4860]: I1211 08:27:21.507541 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-wt4jr"] Dec 11 08:27:21 crc kubenswrapper[4860]: I1211 08:27:21.525427 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/375a2a30-43d5-4557-be42-44a862c171c6-config\") pod \"dnsmasq-dns-675f4bcbfc-w78r8\" (UID: \"375a2a30-43d5-4557-be42-44a862c171c6\") " pod="openstack/dnsmasq-dns-675f4bcbfc-w78r8" Dec 11 08:27:21 crc kubenswrapper[4860]: I1211 08:27:21.525519 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zfgcm\" (UniqueName: \"kubernetes.io/projected/375a2a30-43d5-4557-be42-44a862c171c6-kube-api-access-zfgcm\") pod \"dnsmasq-dns-675f4bcbfc-w78r8\" (UID: \"375a2a30-43d5-4557-be42-44a862c171c6\") " pod="openstack/dnsmasq-dns-675f4bcbfc-w78r8" Dec 11 08:27:21 crc kubenswrapper[4860]: I1211 08:27:21.626902 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zfgcm\" (UniqueName: \"kubernetes.io/projected/375a2a30-43d5-4557-be42-44a862c171c6-kube-api-access-zfgcm\") pod \"dnsmasq-dns-675f4bcbfc-w78r8\" (UID: \"375a2a30-43d5-4557-be42-44a862c171c6\") " pod="openstack/dnsmasq-dns-675f4bcbfc-w78r8" Dec 11 08:27:21 crc kubenswrapper[4860]: I1211 08:27:21.627031 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4pfgv\" (UniqueName: \"kubernetes.io/projected/79fac2d1-5332-496d-86c9-b82df3c9209a-kube-api-access-4pfgv\") pod \"dnsmasq-dns-78dd6ddcc-wt4jr\" (UID: \"79fac2d1-5332-496d-86c9-b82df3c9209a\") " pod="openstack/dnsmasq-dns-78dd6ddcc-wt4jr" Dec 11 08:27:21 crc kubenswrapper[4860]: I1211 08:27:21.627076 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/375a2a30-43d5-4557-be42-44a862c171c6-config\") pod \"dnsmasq-dns-675f4bcbfc-w78r8\" (UID: \"375a2a30-43d5-4557-be42-44a862c171c6\") " pod="openstack/dnsmasq-dns-675f4bcbfc-w78r8" Dec 11 08:27:21 crc kubenswrapper[4860]: I1211 08:27:21.627103 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/79fac2d1-5332-496d-86c9-b82df3c9209a-config\") pod \"dnsmasq-dns-78dd6ddcc-wt4jr\" (UID: \"79fac2d1-5332-496d-86c9-b82df3c9209a\") " pod="openstack/dnsmasq-dns-78dd6ddcc-wt4jr" Dec 11 08:27:21 crc kubenswrapper[4860]: I1211 08:27:21.627170 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/79fac2d1-5332-496d-86c9-b82df3c9209a-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-wt4jr\" (UID: \"79fac2d1-5332-496d-86c9-b82df3c9209a\") " pod="openstack/dnsmasq-dns-78dd6ddcc-wt4jr" Dec 11 08:27:21 crc kubenswrapper[4860]: I1211 08:27:21.628330 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/375a2a30-43d5-4557-be42-44a862c171c6-config\") pod \"dnsmasq-dns-675f4bcbfc-w78r8\" (UID: \"375a2a30-43d5-4557-be42-44a862c171c6\") " pod="openstack/dnsmasq-dns-675f4bcbfc-w78r8" Dec 11 08:27:21 crc kubenswrapper[4860]: I1211 08:27:21.648732 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zfgcm\" (UniqueName: \"kubernetes.io/projected/375a2a30-43d5-4557-be42-44a862c171c6-kube-api-access-zfgcm\") pod \"dnsmasq-dns-675f4bcbfc-w78r8\" (UID: \"375a2a30-43d5-4557-be42-44a862c171c6\") " pod="openstack/dnsmasq-dns-675f4bcbfc-w78r8" Dec 11 08:27:21 crc kubenswrapper[4860]: I1211 08:27:21.724976 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-w78r8" Dec 11 08:27:21 crc kubenswrapper[4860]: I1211 08:27:21.735184 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4pfgv\" (UniqueName: \"kubernetes.io/projected/79fac2d1-5332-496d-86c9-b82df3c9209a-kube-api-access-4pfgv\") pod \"dnsmasq-dns-78dd6ddcc-wt4jr\" (UID: \"79fac2d1-5332-496d-86c9-b82df3c9209a\") " pod="openstack/dnsmasq-dns-78dd6ddcc-wt4jr" Dec 11 08:27:21 crc kubenswrapper[4860]: I1211 08:27:21.735261 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/79fac2d1-5332-496d-86c9-b82df3c9209a-config\") pod \"dnsmasq-dns-78dd6ddcc-wt4jr\" (UID: \"79fac2d1-5332-496d-86c9-b82df3c9209a\") " pod="openstack/dnsmasq-dns-78dd6ddcc-wt4jr" Dec 11 08:27:21 crc kubenswrapper[4860]: I1211 08:27:21.735301 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/79fac2d1-5332-496d-86c9-b82df3c9209a-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-wt4jr\" (UID: \"79fac2d1-5332-496d-86c9-b82df3c9209a\") " pod="openstack/dnsmasq-dns-78dd6ddcc-wt4jr" Dec 11 08:27:21 crc kubenswrapper[4860]: I1211 08:27:21.736390 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/79fac2d1-5332-496d-86c9-b82df3c9209a-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-wt4jr\" (UID: \"79fac2d1-5332-496d-86c9-b82df3c9209a\") " pod="openstack/dnsmasq-dns-78dd6ddcc-wt4jr" Dec 11 08:27:21 crc kubenswrapper[4860]: I1211 08:27:21.736972 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/79fac2d1-5332-496d-86c9-b82df3c9209a-config\") pod \"dnsmasq-dns-78dd6ddcc-wt4jr\" (UID: \"79fac2d1-5332-496d-86c9-b82df3c9209a\") " pod="openstack/dnsmasq-dns-78dd6ddcc-wt4jr" Dec 11 08:27:21 crc kubenswrapper[4860]: I1211 08:27:21.758843 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4pfgv\" (UniqueName: \"kubernetes.io/projected/79fac2d1-5332-496d-86c9-b82df3c9209a-kube-api-access-4pfgv\") pod \"dnsmasq-dns-78dd6ddcc-wt4jr\" (UID: \"79fac2d1-5332-496d-86c9-b82df3c9209a\") " pod="openstack/dnsmasq-dns-78dd6ddcc-wt4jr" Dec 11 08:27:21 crc kubenswrapper[4860]: I1211 08:27:21.809581 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-wt4jr" Dec 11 08:27:22 crc kubenswrapper[4860]: I1211 08:27:22.223507 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-w78r8"] Dec 11 08:27:22 crc kubenswrapper[4860]: W1211 08:27:22.233342 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod375a2a30_43d5_4557_be42_44a862c171c6.slice/crio-f7db3763bed5d831657c1f12305c76e0924691e1ecc7060cc05b12b83ce81d38 WatchSource:0}: Error finding container f7db3763bed5d831657c1f12305c76e0924691e1ecc7060cc05b12b83ce81d38: Status 404 returned error can't find the container with id f7db3763bed5d831657c1f12305c76e0924691e1ecc7060cc05b12b83ce81d38 Dec 11 08:27:22 crc kubenswrapper[4860]: I1211 08:27:22.245854 4860 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 11 08:27:22 crc kubenswrapper[4860]: I1211 08:27:22.345405 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-wt4jr"] Dec 11 08:27:22 crc kubenswrapper[4860]: W1211 08:27:22.347175 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod79fac2d1_5332_496d_86c9_b82df3c9209a.slice/crio-5a1c218decb2a91b10a6ae30a4c33885dfd804d66fd99ecc270e7728c34d5bf3 WatchSource:0}: Error finding container 5a1c218decb2a91b10a6ae30a4c33885dfd804d66fd99ecc270e7728c34d5bf3: Status 404 returned error can't find the container with id 5a1c218decb2a91b10a6ae30a4c33885dfd804d66fd99ecc270e7728c34d5bf3 Dec 11 08:27:22 crc kubenswrapper[4860]: I1211 08:27:22.887724 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-w78r8" event={"ID":"375a2a30-43d5-4557-be42-44a862c171c6","Type":"ContainerStarted","Data":"f7db3763bed5d831657c1f12305c76e0924691e1ecc7060cc05b12b83ce81d38"} Dec 11 08:27:22 crc kubenswrapper[4860]: I1211 08:27:22.890663 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-wt4jr" event={"ID":"79fac2d1-5332-496d-86c9-b82df3c9209a","Type":"ContainerStarted","Data":"5a1c218decb2a91b10a6ae30a4c33885dfd804d66fd99ecc270e7728c34d5bf3"} Dec 11 08:27:24 crc kubenswrapper[4860]: I1211 08:27:24.424031 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-w78r8"] Dec 11 08:27:24 crc kubenswrapper[4860]: I1211 08:27:24.452981 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-vsf77"] Dec 11 08:27:24 crc kubenswrapper[4860]: I1211 08:27:24.462421 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-vsf77" Dec 11 08:27:24 crc kubenswrapper[4860]: I1211 08:27:24.489679 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-vsf77"] Dec 11 08:27:24 crc kubenswrapper[4860]: I1211 08:27:24.595951 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5ea9b400-63b3-480e-880e-978b406fce50-config\") pod \"dnsmasq-dns-666b6646f7-vsf77\" (UID: \"5ea9b400-63b3-480e-880e-978b406fce50\") " pod="openstack/dnsmasq-dns-666b6646f7-vsf77" Dec 11 08:27:24 crc kubenswrapper[4860]: I1211 08:27:24.596025 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5ea9b400-63b3-480e-880e-978b406fce50-dns-svc\") pod \"dnsmasq-dns-666b6646f7-vsf77\" (UID: \"5ea9b400-63b3-480e-880e-978b406fce50\") " pod="openstack/dnsmasq-dns-666b6646f7-vsf77" Dec 11 08:27:24 crc kubenswrapper[4860]: I1211 08:27:24.596799 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rgh8v\" (UniqueName: \"kubernetes.io/projected/5ea9b400-63b3-480e-880e-978b406fce50-kube-api-access-rgh8v\") pod \"dnsmasq-dns-666b6646f7-vsf77\" (UID: \"5ea9b400-63b3-480e-880e-978b406fce50\") " pod="openstack/dnsmasq-dns-666b6646f7-vsf77" Dec 11 08:27:24 crc kubenswrapper[4860]: I1211 08:27:24.699450 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5ea9b400-63b3-480e-880e-978b406fce50-config\") pod \"dnsmasq-dns-666b6646f7-vsf77\" (UID: \"5ea9b400-63b3-480e-880e-978b406fce50\") " pod="openstack/dnsmasq-dns-666b6646f7-vsf77" Dec 11 08:27:24 crc kubenswrapper[4860]: I1211 08:27:24.699505 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5ea9b400-63b3-480e-880e-978b406fce50-dns-svc\") pod \"dnsmasq-dns-666b6646f7-vsf77\" (UID: \"5ea9b400-63b3-480e-880e-978b406fce50\") " pod="openstack/dnsmasq-dns-666b6646f7-vsf77" Dec 11 08:27:24 crc kubenswrapper[4860]: I1211 08:27:24.699625 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rgh8v\" (UniqueName: \"kubernetes.io/projected/5ea9b400-63b3-480e-880e-978b406fce50-kube-api-access-rgh8v\") pod \"dnsmasq-dns-666b6646f7-vsf77\" (UID: \"5ea9b400-63b3-480e-880e-978b406fce50\") " pod="openstack/dnsmasq-dns-666b6646f7-vsf77" Dec 11 08:27:24 crc kubenswrapper[4860]: I1211 08:27:24.701595 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5ea9b400-63b3-480e-880e-978b406fce50-dns-svc\") pod \"dnsmasq-dns-666b6646f7-vsf77\" (UID: \"5ea9b400-63b3-480e-880e-978b406fce50\") " pod="openstack/dnsmasq-dns-666b6646f7-vsf77" Dec 11 08:27:24 crc kubenswrapper[4860]: I1211 08:27:24.701970 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5ea9b400-63b3-480e-880e-978b406fce50-config\") pod \"dnsmasq-dns-666b6646f7-vsf77\" (UID: \"5ea9b400-63b3-480e-880e-978b406fce50\") " pod="openstack/dnsmasq-dns-666b6646f7-vsf77" Dec 11 08:27:24 crc kubenswrapper[4860]: I1211 08:27:24.764246 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rgh8v\" (UniqueName: \"kubernetes.io/projected/5ea9b400-63b3-480e-880e-978b406fce50-kube-api-access-rgh8v\") pod \"dnsmasq-dns-666b6646f7-vsf77\" (UID: \"5ea9b400-63b3-480e-880e-978b406fce50\") " pod="openstack/dnsmasq-dns-666b6646f7-vsf77" Dec 11 08:27:24 crc kubenswrapper[4860]: I1211 08:27:24.771428 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-wt4jr"] Dec 11 08:27:24 crc kubenswrapper[4860]: I1211 08:27:24.819969 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-vwqtx"] Dec 11 08:27:24 crc kubenswrapper[4860]: I1211 08:27:24.820132 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-vsf77" Dec 11 08:27:24 crc kubenswrapper[4860]: I1211 08:27:24.821761 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-vwqtx" Dec 11 08:27:24 crc kubenswrapper[4860]: I1211 08:27:24.833585 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-vwqtx"] Dec 11 08:27:25 crc kubenswrapper[4860]: I1211 08:27:25.004829 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-89cmp\" (UniqueName: \"kubernetes.io/projected/7e94e216-32ea-44d2-b7ae-f765ebbdd524-kube-api-access-89cmp\") pod \"dnsmasq-dns-57d769cc4f-vwqtx\" (UID: \"7e94e216-32ea-44d2-b7ae-f765ebbdd524\") " pod="openstack/dnsmasq-dns-57d769cc4f-vwqtx" Dec 11 08:27:25 crc kubenswrapper[4860]: I1211 08:27:25.006106 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7e94e216-32ea-44d2-b7ae-f765ebbdd524-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-vwqtx\" (UID: \"7e94e216-32ea-44d2-b7ae-f765ebbdd524\") " pod="openstack/dnsmasq-dns-57d769cc4f-vwqtx" Dec 11 08:27:25 crc kubenswrapper[4860]: I1211 08:27:25.006166 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e94e216-32ea-44d2-b7ae-f765ebbdd524-config\") pod \"dnsmasq-dns-57d769cc4f-vwqtx\" (UID: \"7e94e216-32ea-44d2-b7ae-f765ebbdd524\") " pod="openstack/dnsmasq-dns-57d769cc4f-vwqtx" Dec 11 08:27:25 crc kubenswrapper[4860]: I1211 08:27:25.108778 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e94e216-32ea-44d2-b7ae-f765ebbdd524-config\") pod \"dnsmasq-dns-57d769cc4f-vwqtx\" (UID: \"7e94e216-32ea-44d2-b7ae-f765ebbdd524\") " pod="openstack/dnsmasq-dns-57d769cc4f-vwqtx" Dec 11 08:27:25 crc kubenswrapper[4860]: I1211 08:27:25.108922 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-89cmp\" (UniqueName: \"kubernetes.io/projected/7e94e216-32ea-44d2-b7ae-f765ebbdd524-kube-api-access-89cmp\") pod \"dnsmasq-dns-57d769cc4f-vwqtx\" (UID: \"7e94e216-32ea-44d2-b7ae-f765ebbdd524\") " pod="openstack/dnsmasq-dns-57d769cc4f-vwqtx" Dec 11 08:27:25 crc kubenswrapper[4860]: I1211 08:27:25.108960 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7e94e216-32ea-44d2-b7ae-f765ebbdd524-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-vwqtx\" (UID: \"7e94e216-32ea-44d2-b7ae-f765ebbdd524\") " pod="openstack/dnsmasq-dns-57d769cc4f-vwqtx" Dec 11 08:27:25 crc kubenswrapper[4860]: I1211 08:27:25.110066 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7e94e216-32ea-44d2-b7ae-f765ebbdd524-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-vwqtx\" (UID: \"7e94e216-32ea-44d2-b7ae-f765ebbdd524\") " pod="openstack/dnsmasq-dns-57d769cc4f-vwqtx" Dec 11 08:27:25 crc kubenswrapper[4860]: I1211 08:27:25.110196 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e94e216-32ea-44d2-b7ae-f765ebbdd524-config\") pod \"dnsmasq-dns-57d769cc4f-vwqtx\" (UID: \"7e94e216-32ea-44d2-b7ae-f765ebbdd524\") " pod="openstack/dnsmasq-dns-57d769cc4f-vwqtx" Dec 11 08:27:25 crc kubenswrapper[4860]: I1211 08:27:25.135374 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-89cmp\" (UniqueName: \"kubernetes.io/projected/7e94e216-32ea-44d2-b7ae-f765ebbdd524-kube-api-access-89cmp\") pod \"dnsmasq-dns-57d769cc4f-vwqtx\" (UID: \"7e94e216-32ea-44d2-b7ae-f765ebbdd524\") " pod="openstack/dnsmasq-dns-57d769cc4f-vwqtx" Dec 11 08:27:25 crc kubenswrapper[4860]: I1211 08:27:25.163256 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-vwqtx" Dec 11 08:27:25 crc kubenswrapper[4860]: I1211 08:27:25.541773 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-vsf77"] Dec 11 08:27:25 crc kubenswrapper[4860]: W1211 08:27:25.574141 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5ea9b400_63b3_480e_880e_978b406fce50.slice/crio-4d3997c82d9e0064cf79f1a5f121041549d0feddf6308665af53a842c9d5a66c WatchSource:0}: Error finding container 4d3997c82d9e0064cf79f1a5f121041549d0feddf6308665af53a842c9d5a66c: Status 404 returned error can't find the container with id 4d3997c82d9e0064cf79f1a5f121041549d0feddf6308665af53a842c9d5a66c Dec 11 08:27:25 crc kubenswrapper[4860]: I1211 08:27:25.639821 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 11 08:27:25 crc kubenswrapper[4860]: I1211 08:27:25.646170 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 11 08:27:25 crc kubenswrapper[4860]: I1211 08:27:25.648516 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 11 08:27:25 crc kubenswrapper[4860]: I1211 08:27:25.649240 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-6vxjr" Dec 11 08:27:25 crc kubenswrapper[4860]: I1211 08:27:25.649404 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 11 08:27:25 crc kubenswrapper[4860]: I1211 08:27:25.649556 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 11 08:27:25 crc kubenswrapper[4860]: I1211 08:27:25.649707 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 11 08:27:25 crc kubenswrapper[4860]: I1211 08:27:25.649814 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 11 08:27:25 crc kubenswrapper[4860]: I1211 08:27:25.650064 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 11 08:27:25 crc kubenswrapper[4860]: I1211 08:27:25.650324 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 11 08:27:25 crc kubenswrapper[4860]: I1211 08:27:25.676350 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-vwqtx"] Dec 11 08:27:25 crc kubenswrapper[4860]: W1211 08:27:25.687057 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7e94e216_32ea_44d2_b7ae_f765ebbdd524.slice/crio-31913ca3198c2b25c60b6218937620ced66fe11ea1da0f74d81684db7eb8693a WatchSource:0}: Error finding container 31913ca3198c2b25c60b6218937620ced66fe11ea1da0f74d81684db7eb8693a: Status 404 returned error can't find the container with id 31913ca3198c2b25c60b6218937620ced66fe11ea1da0f74d81684db7eb8693a Dec 11 08:27:25 crc kubenswrapper[4860]: I1211 08:27:25.847749 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/97634ec0-7953-4e38-a07e-95967ce874d7-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"97634ec0-7953-4e38-a07e-95967ce874d7\") " pod="openstack/rabbitmq-server-0" Dec 11 08:27:25 crc kubenswrapper[4860]: I1211 08:27:25.847842 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/97634ec0-7953-4e38-a07e-95967ce874d7-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"97634ec0-7953-4e38-a07e-95967ce874d7\") " pod="openstack/rabbitmq-server-0" Dec 11 08:27:25 crc kubenswrapper[4860]: I1211 08:27:25.847863 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/97634ec0-7953-4e38-a07e-95967ce874d7-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"97634ec0-7953-4e38-a07e-95967ce874d7\") " pod="openstack/rabbitmq-server-0" Dec 11 08:27:25 crc kubenswrapper[4860]: I1211 08:27:25.847888 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jbrbn\" (UniqueName: \"kubernetes.io/projected/97634ec0-7953-4e38-a07e-95967ce874d7-kube-api-access-jbrbn\") pod \"rabbitmq-server-0\" (UID: \"97634ec0-7953-4e38-a07e-95967ce874d7\") " pod="openstack/rabbitmq-server-0" Dec 11 08:27:25 crc kubenswrapper[4860]: I1211 08:27:25.848069 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/97634ec0-7953-4e38-a07e-95967ce874d7-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"97634ec0-7953-4e38-a07e-95967ce874d7\") " pod="openstack/rabbitmq-server-0" Dec 11 08:27:25 crc kubenswrapper[4860]: I1211 08:27:25.848140 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/97634ec0-7953-4e38-a07e-95967ce874d7-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"97634ec0-7953-4e38-a07e-95967ce874d7\") " pod="openstack/rabbitmq-server-0" Dec 11 08:27:25 crc kubenswrapper[4860]: I1211 08:27:25.848201 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/97634ec0-7953-4e38-a07e-95967ce874d7-config-data\") pod \"rabbitmq-server-0\" (UID: \"97634ec0-7953-4e38-a07e-95967ce874d7\") " pod="openstack/rabbitmq-server-0" Dec 11 08:27:25 crc kubenswrapper[4860]: I1211 08:27:25.848336 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/97634ec0-7953-4e38-a07e-95967ce874d7-server-conf\") pod \"rabbitmq-server-0\" (UID: \"97634ec0-7953-4e38-a07e-95967ce874d7\") " pod="openstack/rabbitmq-server-0" Dec 11 08:27:25 crc kubenswrapper[4860]: I1211 08:27:25.848414 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"97634ec0-7953-4e38-a07e-95967ce874d7\") " pod="openstack/rabbitmq-server-0" Dec 11 08:27:25 crc kubenswrapper[4860]: I1211 08:27:25.848476 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/97634ec0-7953-4e38-a07e-95967ce874d7-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"97634ec0-7953-4e38-a07e-95967ce874d7\") " pod="openstack/rabbitmq-server-0" Dec 11 08:27:25 crc kubenswrapper[4860]: I1211 08:27:25.848565 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/97634ec0-7953-4e38-a07e-95967ce874d7-pod-info\") pod \"rabbitmq-server-0\" (UID: \"97634ec0-7953-4e38-a07e-95967ce874d7\") " pod="openstack/rabbitmq-server-0" Dec 11 08:27:25 crc kubenswrapper[4860]: I1211 08:27:25.933156 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-vsf77" event={"ID":"5ea9b400-63b3-480e-880e-978b406fce50","Type":"ContainerStarted","Data":"4d3997c82d9e0064cf79f1a5f121041549d0feddf6308665af53a842c9d5a66c"} Dec 11 08:27:25 crc kubenswrapper[4860]: I1211 08:27:25.935585 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-vwqtx" event={"ID":"7e94e216-32ea-44d2-b7ae-f765ebbdd524","Type":"ContainerStarted","Data":"31913ca3198c2b25c60b6218937620ced66fe11ea1da0f74d81684db7eb8693a"} Dec 11 08:27:25 crc kubenswrapper[4860]: I1211 08:27:25.937864 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 11 08:27:25 crc kubenswrapper[4860]: I1211 08:27:25.946840 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:27:25 crc kubenswrapper[4860]: I1211 08:27:25.952253 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 11 08:27:25 crc kubenswrapper[4860]: I1211 08:27:25.952584 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 11 08:27:25 crc kubenswrapper[4860]: I1211 08:27:25.952729 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 11 08:27:25 crc kubenswrapper[4860]: I1211 08:27:25.952922 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 11 08:27:25 crc kubenswrapper[4860]: I1211 08:27:25.953059 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-vmvgn" Dec 11 08:27:25 crc kubenswrapper[4860]: I1211 08:27:25.953191 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 11 08:27:25 crc kubenswrapper[4860]: I1211 08:27:25.953324 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 11 08:27:25 crc kubenswrapper[4860]: I1211 08:27:25.958231 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/97634ec0-7953-4e38-a07e-95967ce874d7-pod-info\") pod \"rabbitmq-server-0\" (UID: \"97634ec0-7953-4e38-a07e-95967ce874d7\") " pod="openstack/rabbitmq-server-0" Dec 11 08:27:25 crc kubenswrapper[4860]: I1211 08:27:25.958292 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/97634ec0-7953-4e38-a07e-95967ce874d7-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"97634ec0-7953-4e38-a07e-95967ce874d7\") " pod="openstack/rabbitmq-server-0" Dec 11 08:27:25 crc kubenswrapper[4860]: I1211 08:27:25.958335 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/97634ec0-7953-4e38-a07e-95967ce874d7-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"97634ec0-7953-4e38-a07e-95967ce874d7\") " pod="openstack/rabbitmq-server-0" Dec 11 08:27:25 crc kubenswrapper[4860]: I1211 08:27:25.958352 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/97634ec0-7953-4e38-a07e-95967ce874d7-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"97634ec0-7953-4e38-a07e-95967ce874d7\") " pod="openstack/rabbitmq-server-0" Dec 11 08:27:25 crc kubenswrapper[4860]: I1211 08:27:25.958370 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jbrbn\" (UniqueName: \"kubernetes.io/projected/97634ec0-7953-4e38-a07e-95967ce874d7-kube-api-access-jbrbn\") pod \"rabbitmq-server-0\" (UID: \"97634ec0-7953-4e38-a07e-95967ce874d7\") " pod="openstack/rabbitmq-server-0" Dec 11 08:27:25 crc kubenswrapper[4860]: I1211 08:27:25.958402 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/97634ec0-7953-4e38-a07e-95967ce874d7-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"97634ec0-7953-4e38-a07e-95967ce874d7\") " pod="openstack/rabbitmq-server-0" Dec 11 08:27:25 crc kubenswrapper[4860]: I1211 08:27:25.958424 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/97634ec0-7953-4e38-a07e-95967ce874d7-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"97634ec0-7953-4e38-a07e-95967ce874d7\") " pod="openstack/rabbitmq-server-0" Dec 11 08:27:25 crc kubenswrapper[4860]: I1211 08:27:25.958447 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/97634ec0-7953-4e38-a07e-95967ce874d7-config-data\") pod \"rabbitmq-server-0\" (UID: \"97634ec0-7953-4e38-a07e-95967ce874d7\") " pod="openstack/rabbitmq-server-0" Dec 11 08:27:25 crc kubenswrapper[4860]: I1211 08:27:25.958480 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/97634ec0-7953-4e38-a07e-95967ce874d7-server-conf\") pod \"rabbitmq-server-0\" (UID: \"97634ec0-7953-4e38-a07e-95967ce874d7\") " pod="openstack/rabbitmq-server-0" Dec 11 08:27:25 crc kubenswrapper[4860]: I1211 08:27:25.958502 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"97634ec0-7953-4e38-a07e-95967ce874d7\") " pod="openstack/rabbitmq-server-0" Dec 11 08:27:25 crc kubenswrapper[4860]: I1211 08:27:25.958520 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/97634ec0-7953-4e38-a07e-95967ce874d7-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"97634ec0-7953-4e38-a07e-95967ce874d7\") " pod="openstack/rabbitmq-server-0" Dec 11 08:27:25 crc kubenswrapper[4860]: I1211 08:27:25.965074 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/97634ec0-7953-4e38-a07e-95967ce874d7-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"97634ec0-7953-4e38-a07e-95967ce874d7\") " pod="openstack/rabbitmq-server-0" Dec 11 08:27:25 crc kubenswrapper[4860]: I1211 08:27:25.971984 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/97634ec0-7953-4e38-a07e-95967ce874d7-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"97634ec0-7953-4e38-a07e-95967ce874d7\") " pod="openstack/rabbitmq-server-0" Dec 11 08:27:25 crc kubenswrapper[4860]: I1211 08:27:25.973664 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 11 08:27:25 crc kubenswrapper[4860]: I1211 08:27:25.974518 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/97634ec0-7953-4e38-a07e-95967ce874d7-server-conf\") pod \"rabbitmq-server-0\" (UID: \"97634ec0-7953-4e38-a07e-95967ce874d7\") " pod="openstack/rabbitmq-server-0" Dec 11 08:27:25 crc kubenswrapper[4860]: I1211 08:27:25.975127 4860 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"97634ec0-7953-4e38-a07e-95967ce874d7\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/rabbitmq-server-0" Dec 11 08:27:25 crc kubenswrapper[4860]: I1211 08:27:25.992663 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/97634ec0-7953-4e38-a07e-95967ce874d7-config-data\") pod \"rabbitmq-server-0\" (UID: \"97634ec0-7953-4e38-a07e-95967ce874d7\") " pod="openstack/rabbitmq-server-0" Dec 11 08:27:25 crc kubenswrapper[4860]: I1211 08:27:25.996923 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/97634ec0-7953-4e38-a07e-95967ce874d7-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"97634ec0-7953-4e38-a07e-95967ce874d7\") " pod="openstack/rabbitmq-server-0" Dec 11 08:27:26 crc kubenswrapper[4860]: I1211 08:27:26.003347 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/97634ec0-7953-4e38-a07e-95967ce874d7-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"97634ec0-7953-4e38-a07e-95967ce874d7\") " pod="openstack/rabbitmq-server-0" Dec 11 08:27:26 crc kubenswrapper[4860]: I1211 08:27:26.005248 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/97634ec0-7953-4e38-a07e-95967ce874d7-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"97634ec0-7953-4e38-a07e-95967ce874d7\") " pod="openstack/rabbitmq-server-0" Dec 11 08:27:26 crc kubenswrapper[4860]: I1211 08:27:26.005589 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/97634ec0-7953-4e38-a07e-95967ce874d7-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"97634ec0-7953-4e38-a07e-95967ce874d7\") " pod="openstack/rabbitmq-server-0" Dec 11 08:27:26 crc kubenswrapper[4860]: I1211 08:27:26.018300 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/97634ec0-7953-4e38-a07e-95967ce874d7-pod-info\") pod \"rabbitmq-server-0\" (UID: \"97634ec0-7953-4e38-a07e-95967ce874d7\") " pod="openstack/rabbitmq-server-0" Dec 11 08:27:26 crc kubenswrapper[4860]: I1211 08:27:26.033782 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"97634ec0-7953-4e38-a07e-95967ce874d7\") " pod="openstack/rabbitmq-server-0" Dec 11 08:27:26 crc kubenswrapper[4860]: I1211 08:27:26.037311 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jbrbn\" (UniqueName: \"kubernetes.io/projected/97634ec0-7953-4e38-a07e-95967ce874d7-kube-api-access-jbrbn\") pod \"rabbitmq-server-0\" (UID: \"97634ec0-7953-4e38-a07e-95967ce874d7\") " pod="openstack/rabbitmq-server-0" Dec 11 08:27:26 crc kubenswrapper[4860]: I1211 08:27:26.060256 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a004cb7b-ee08-44c0-9dc7-d0b69509fe13-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"a004cb7b-ee08-44c0-9dc7-d0b69509fe13\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:27:26 crc kubenswrapper[4860]: I1211 08:27:26.060340 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a004cb7b-ee08-44c0-9dc7-d0b69509fe13-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"a004cb7b-ee08-44c0-9dc7-d0b69509fe13\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:27:26 crc kubenswrapper[4860]: I1211 08:27:26.060374 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a004cb7b-ee08-44c0-9dc7-d0b69509fe13-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"a004cb7b-ee08-44c0-9dc7-d0b69509fe13\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:27:26 crc kubenswrapper[4860]: I1211 08:27:26.060479 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a004cb7b-ee08-44c0-9dc7-d0b69509fe13-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a004cb7b-ee08-44c0-9dc7-d0b69509fe13\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:27:26 crc kubenswrapper[4860]: I1211 08:27:26.060529 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a004cb7b-ee08-44c0-9dc7-d0b69509fe13-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"a004cb7b-ee08-44c0-9dc7-d0b69509fe13\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:27:26 crc kubenswrapper[4860]: I1211 08:27:26.060584 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a004cb7b-ee08-44c0-9dc7-d0b69509fe13-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"a004cb7b-ee08-44c0-9dc7-d0b69509fe13\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:27:26 crc kubenswrapper[4860]: I1211 08:27:26.060855 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a004cb7b-ee08-44c0-9dc7-d0b69509fe13-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"a004cb7b-ee08-44c0-9dc7-d0b69509fe13\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:27:26 crc kubenswrapper[4860]: I1211 08:27:26.061141 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a004cb7b-ee08-44c0-9dc7-d0b69509fe13-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a004cb7b-ee08-44c0-9dc7-d0b69509fe13\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:27:26 crc kubenswrapper[4860]: I1211 08:27:26.061230 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a004cb7b-ee08-44c0-9dc7-d0b69509fe13-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"a004cb7b-ee08-44c0-9dc7-d0b69509fe13\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:27:26 crc kubenswrapper[4860]: I1211 08:27:26.062449 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hkscb\" (UniqueName: \"kubernetes.io/projected/a004cb7b-ee08-44c0-9dc7-d0b69509fe13-kube-api-access-hkscb\") pod \"rabbitmq-cell1-server-0\" (UID: \"a004cb7b-ee08-44c0-9dc7-d0b69509fe13\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:27:26 crc kubenswrapper[4860]: I1211 08:27:26.062522 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"a004cb7b-ee08-44c0-9dc7-d0b69509fe13\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:27:26 crc kubenswrapper[4860]: I1211 08:27:26.167836 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a004cb7b-ee08-44c0-9dc7-d0b69509fe13-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"a004cb7b-ee08-44c0-9dc7-d0b69509fe13\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:27:26 crc kubenswrapper[4860]: I1211 08:27:26.167913 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hkscb\" (UniqueName: \"kubernetes.io/projected/a004cb7b-ee08-44c0-9dc7-d0b69509fe13-kube-api-access-hkscb\") pod \"rabbitmq-cell1-server-0\" (UID: \"a004cb7b-ee08-44c0-9dc7-d0b69509fe13\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:27:26 crc kubenswrapper[4860]: I1211 08:27:26.168007 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"a004cb7b-ee08-44c0-9dc7-d0b69509fe13\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:27:26 crc kubenswrapper[4860]: I1211 08:27:26.168106 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a004cb7b-ee08-44c0-9dc7-d0b69509fe13-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"a004cb7b-ee08-44c0-9dc7-d0b69509fe13\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:27:26 crc kubenswrapper[4860]: I1211 08:27:26.168138 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a004cb7b-ee08-44c0-9dc7-d0b69509fe13-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"a004cb7b-ee08-44c0-9dc7-d0b69509fe13\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:27:26 crc kubenswrapper[4860]: I1211 08:27:26.168158 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a004cb7b-ee08-44c0-9dc7-d0b69509fe13-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"a004cb7b-ee08-44c0-9dc7-d0b69509fe13\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:27:26 crc kubenswrapper[4860]: I1211 08:27:26.168217 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a004cb7b-ee08-44c0-9dc7-d0b69509fe13-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a004cb7b-ee08-44c0-9dc7-d0b69509fe13\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:27:26 crc kubenswrapper[4860]: I1211 08:27:26.168255 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a004cb7b-ee08-44c0-9dc7-d0b69509fe13-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"a004cb7b-ee08-44c0-9dc7-d0b69509fe13\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:27:26 crc kubenswrapper[4860]: I1211 08:27:26.168317 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a004cb7b-ee08-44c0-9dc7-d0b69509fe13-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"a004cb7b-ee08-44c0-9dc7-d0b69509fe13\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:27:26 crc kubenswrapper[4860]: I1211 08:27:26.168362 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a004cb7b-ee08-44c0-9dc7-d0b69509fe13-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"a004cb7b-ee08-44c0-9dc7-d0b69509fe13\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:27:26 crc kubenswrapper[4860]: I1211 08:27:26.168405 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a004cb7b-ee08-44c0-9dc7-d0b69509fe13-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a004cb7b-ee08-44c0-9dc7-d0b69509fe13\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:27:26 crc kubenswrapper[4860]: I1211 08:27:26.168954 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a004cb7b-ee08-44c0-9dc7-d0b69509fe13-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"a004cb7b-ee08-44c0-9dc7-d0b69509fe13\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:27:26 crc kubenswrapper[4860]: I1211 08:27:26.169117 4860 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"a004cb7b-ee08-44c0-9dc7-d0b69509fe13\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:27:26 crc kubenswrapper[4860]: I1211 08:27:26.169216 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a004cb7b-ee08-44c0-9dc7-d0b69509fe13-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a004cb7b-ee08-44c0-9dc7-d0b69509fe13\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:27:26 crc kubenswrapper[4860]: I1211 08:27:26.170239 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a004cb7b-ee08-44c0-9dc7-d0b69509fe13-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"a004cb7b-ee08-44c0-9dc7-d0b69509fe13\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:27:26 crc kubenswrapper[4860]: I1211 08:27:26.171193 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a004cb7b-ee08-44c0-9dc7-d0b69509fe13-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"a004cb7b-ee08-44c0-9dc7-d0b69509fe13\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:27:26 crc kubenswrapper[4860]: I1211 08:27:26.171875 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a004cb7b-ee08-44c0-9dc7-d0b69509fe13-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a004cb7b-ee08-44c0-9dc7-d0b69509fe13\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:27:26 crc kubenswrapper[4860]: I1211 08:27:26.174484 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a004cb7b-ee08-44c0-9dc7-d0b69509fe13-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"a004cb7b-ee08-44c0-9dc7-d0b69509fe13\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:27:26 crc kubenswrapper[4860]: I1211 08:27:26.175084 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a004cb7b-ee08-44c0-9dc7-d0b69509fe13-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"a004cb7b-ee08-44c0-9dc7-d0b69509fe13\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:27:26 crc kubenswrapper[4860]: I1211 08:27:26.176449 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a004cb7b-ee08-44c0-9dc7-d0b69509fe13-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"a004cb7b-ee08-44c0-9dc7-d0b69509fe13\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:27:26 crc kubenswrapper[4860]: I1211 08:27:26.179819 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a004cb7b-ee08-44c0-9dc7-d0b69509fe13-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"a004cb7b-ee08-44c0-9dc7-d0b69509fe13\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:27:26 crc kubenswrapper[4860]: I1211 08:27:26.196391 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hkscb\" (UniqueName: \"kubernetes.io/projected/a004cb7b-ee08-44c0-9dc7-d0b69509fe13-kube-api-access-hkscb\") pod \"rabbitmq-cell1-server-0\" (UID: \"a004cb7b-ee08-44c0-9dc7-d0b69509fe13\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:27:26 crc kubenswrapper[4860]: I1211 08:27:26.206370 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"a004cb7b-ee08-44c0-9dc7-d0b69509fe13\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:27:26 crc kubenswrapper[4860]: I1211 08:27:26.282581 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 11 08:27:26 crc kubenswrapper[4860]: I1211 08:27:26.381851 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:27:27 crc kubenswrapper[4860]: I1211 08:27:27.005885 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 11 08:27:27 crc kubenswrapper[4860]: I1211 08:27:27.043946 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 11 08:27:27 crc kubenswrapper[4860]: W1211 08:27:27.084802 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod97634ec0_7953_4e38_a07e_95967ce874d7.slice/crio-01823976a22b0507fcfd26eed82dc3121aabbf54791145971f8499276ddbea83 WatchSource:0}: Error finding container 01823976a22b0507fcfd26eed82dc3121aabbf54791145971f8499276ddbea83: Status 404 returned error can't find the container with id 01823976a22b0507fcfd26eed82dc3121aabbf54791145971f8499276ddbea83 Dec 11 08:27:27 crc kubenswrapper[4860]: I1211 08:27:27.282011 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Dec 11 08:27:27 crc kubenswrapper[4860]: I1211 08:27:27.285617 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 11 08:27:27 crc kubenswrapper[4860]: I1211 08:27:27.289400 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Dec 11 08:27:27 crc kubenswrapper[4860]: I1211 08:27:27.289595 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-c729z" Dec 11 08:27:27 crc kubenswrapper[4860]: I1211 08:27:27.290918 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Dec 11 08:27:27 crc kubenswrapper[4860]: I1211 08:27:27.292509 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Dec 11 08:27:27 crc kubenswrapper[4860]: I1211 08:27:27.296567 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Dec 11 08:27:27 crc kubenswrapper[4860]: I1211 08:27:27.309210 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 11 08:27:27 crc kubenswrapper[4860]: I1211 08:27:27.396113 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c0d5e719-9d28-4e31-9349-f882c4b72567-operator-scripts\") pod \"openstack-galera-0\" (UID: \"c0d5e719-9d28-4e31-9349-f882c4b72567\") " pod="openstack/openstack-galera-0" Dec 11 08:27:27 crc kubenswrapper[4860]: I1211 08:27:27.396200 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/c0d5e719-9d28-4e31-9349-f882c4b72567-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"c0d5e719-9d28-4e31-9349-f882c4b72567\") " pod="openstack/openstack-galera-0" Dec 11 08:27:27 crc kubenswrapper[4860]: I1211 08:27:27.396315 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-26hfr\" (UniqueName: \"kubernetes.io/projected/c0d5e719-9d28-4e31-9349-f882c4b72567-kube-api-access-26hfr\") pod \"openstack-galera-0\" (UID: \"c0d5e719-9d28-4e31-9349-f882c4b72567\") " pod="openstack/openstack-galera-0" Dec 11 08:27:27 crc kubenswrapper[4860]: I1211 08:27:27.396412 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/c0d5e719-9d28-4e31-9349-f882c4b72567-config-data-default\") pod \"openstack-galera-0\" (UID: \"c0d5e719-9d28-4e31-9349-f882c4b72567\") " pod="openstack/openstack-galera-0" Dec 11 08:27:27 crc kubenswrapper[4860]: I1211 08:27:27.396573 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-galera-0\" (UID: \"c0d5e719-9d28-4e31-9349-f882c4b72567\") " pod="openstack/openstack-galera-0" Dec 11 08:27:27 crc kubenswrapper[4860]: I1211 08:27:27.396603 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c0d5e719-9d28-4e31-9349-f882c4b72567-kolla-config\") pod \"openstack-galera-0\" (UID: \"c0d5e719-9d28-4e31-9349-f882c4b72567\") " pod="openstack/openstack-galera-0" Dec 11 08:27:27 crc kubenswrapper[4860]: I1211 08:27:27.396973 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/c0d5e719-9d28-4e31-9349-f882c4b72567-config-data-generated\") pod \"openstack-galera-0\" (UID: \"c0d5e719-9d28-4e31-9349-f882c4b72567\") " pod="openstack/openstack-galera-0" Dec 11 08:27:27 crc kubenswrapper[4860]: I1211 08:27:27.397002 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0d5e719-9d28-4e31-9349-f882c4b72567-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"c0d5e719-9d28-4e31-9349-f882c4b72567\") " pod="openstack/openstack-galera-0" Dec 11 08:27:27 crc kubenswrapper[4860]: I1211 08:27:27.500808 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/c0d5e719-9d28-4e31-9349-f882c4b72567-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"c0d5e719-9d28-4e31-9349-f882c4b72567\") " pod="openstack/openstack-galera-0" Dec 11 08:27:27 crc kubenswrapper[4860]: I1211 08:27:27.500862 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-26hfr\" (UniqueName: \"kubernetes.io/projected/c0d5e719-9d28-4e31-9349-f882c4b72567-kube-api-access-26hfr\") pod \"openstack-galera-0\" (UID: \"c0d5e719-9d28-4e31-9349-f882c4b72567\") " pod="openstack/openstack-galera-0" Dec 11 08:27:27 crc kubenswrapper[4860]: I1211 08:27:27.500893 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/c0d5e719-9d28-4e31-9349-f882c4b72567-config-data-default\") pod \"openstack-galera-0\" (UID: \"c0d5e719-9d28-4e31-9349-f882c4b72567\") " pod="openstack/openstack-galera-0" Dec 11 08:27:27 crc kubenswrapper[4860]: I1211 08:27:27.500943 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-galera-0\" (UID: \"c0d5e719-9d28-4e31-9349-f882c4b72567\") " pod="openstack/openstack-galera-0" Dec 11 08:27:27 crc kubenswrapper[4860]: I1211 08:27:27.500967 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c0d5e719-9d28-4e31-9349-f882c4b72567-kolla-config\") pod \"openstack-galera-0\" (UID: \"c0d5e719-9d28-4e31-9349-f882c4b72567\") " pod="openstack/openstack-galera-0" Dec 11 08:27:27 crc kubenswrapper[4860]: I1211 08:27:27.500990 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/c0d5e719-9d28-4e31-9349-f882c4b72567-config-data-generated\") pod \"openstack-galera-0\" (UID: \"c0d5e719-9d28-4e31-9349-f882c4b72567\") " pod="openstack/openstack-galera-0" Dec 11 08:27:27 crc kubenswrapper[4860]: I1211 08:27:27.501006 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0d5e719-9d28-4e31-9349-f882c4b72567-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"c0d5e719-9d28-4e31-9349-f882c4b72567\") " pod="openstack/openstack-galera-0" Dec 11 08:27:27 crc kubenswrapper[4860]: I1211 08:27:27.501045 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c0d5e719-9d28-4e31-9349-f882c4b72567-operator-scripts\") pod \"openstack-galera-0\" (UID: \"c0d5e719-9d28-4e31-9349-f882c4b72567\") " pod="openstack/openstack-galera-0" Dec 11 08:27:27 crc kubenswrapper[4860]: I1211 08:27:27.502192 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c0d5e719-9d28-4e31-9349-f882c4b72567-kolla-config\") pod \"openstack-galera-0\" (UID: \"c0d5e719-9d28-4e31-9349-f882c4b72567\") " pod="openstack/openstack-galera-0" Dec 11 08:27:27 crc kubenswrapper[4860]: I1211 08:27:27.502805 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c0d5e719-9d28-4e31-9349-f882c4b72567-operator-scripts\") pod \"openstack-galera-0\" (UID: \"c0d5e719-9d28-4e31-9349-f882c4b72567\") " pod="openstack/openstack-galera-0" Dec 11 08:27:27 crc kubenswrapper[4860]: I1211 08:27:27.503105 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/c0d5e719-9d28-4e31-9349-f882c4b72567-config-data-default\") pod \"openstack-galera-0\" (UID: \"c0d5e719-9d28-4e31-9349-f882c4b72567\") " pod="openstack/openstack-galera-0" Dec 11 08:27:27 crc kubenswrapper[4860]: I1211 08:27:27.503337 4860 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-galera-0\" (UID: \"c0d5e719-9d28-4e31-9349-f882c4b72567\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/openstack-galera-0" Dec 11 08:27:27 crc kubenswrapper[4860]: I1211 08:27:27.510458 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/c0d5e719-9d28-4e31-9349-f882c4b72567-config-data-generated\") pod \"openstack-galera-0\" (UID: \"c0d5e719-9d28-4e31-9349-f882c4b72567\") " pod="openstack/openstack-galera-0" Dec 11 08:27:27 crc kubenswrapper[4860]: I1211 08:27:27.515456 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/c0d5e719-9d28-4e31-9349-f882c4b72567-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"c0d5e719-9d28-4e31-9349-f882c4b72567\") " pod="openstack/openstack-galera-0" Dec 11 08:27:27 crc kubenswrapper[4860]: I1211 08:27:27.521094 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0d5e719-9d28-4e31-9349-f882c4b72567-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"c0d5e719-9d28-4e31-9349-f882c4b72567\") " pod="openstack/openstack-galera-0" Dec 11 08:27:27 crc kubenswrapper[4860]: I1211 08:27:27.525403 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-26hfr\" (UniqueName: \"kubernetes.io/projected/c0d5e719-9d28-4e31-9349-f882c4b72567-kube-api-access-26hfr\") pod \"openstack-galera-0\" (UID: \"c0d5e719-9d28-4e31-9349-f882c4b72567\") " pod="openstack/openstack-galera-0" Dec 11 08:27:27 crc kubenswrapper[4860]: I1211 08:27:27.568213 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-galera-0\" (UID: \"c0d5e719-9d28-4e31-9349-f882c4b72567\") " pod="openstack/openstack-galera-0" Dec 11 08:27:27 crc kubenswrapper[4860]: I1211 08:27:27.632106 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 11 08:27:27 crc kubenswrapper[4860]: I1211 08:27:27.977110 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"a004cb7b-ee08-44c0-9dc7-d0b69509fe13","Type":"ContainerStarted","Data":"ab72d932df3fb4339b63c207dbbb48deec5ee7b26febe2b3ac79346f2e9282fa"} Dec 11 08:27:27 crc kubenswrapper[4860]: I1211 08:27:27.980538 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"97634ec0-7953-4e38-a07e-95967ce874d7","Type":"ContainerStarted","Data":"01823976a22b0507fcfd26eed82dc3121aabbf54791145971f8499276ddbea83"} Dec 11 08:27:28 crc kubenswrapper[4860]: I1211 08:27:28.629537 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 11 08:27:28 crc kubenswrapper[4860]: I1211 08:27:28.647152 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 11 08:27:28 crc kubenswrapper[4860]: I1211 08:27:28.647297 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 11 08:27:28 crc kubenswrapper[4860]: I1211 08:27:28.655383 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-kcs9q" Dec 11 08:27:28 crc kubenswrapper[4860]: I1211 08:27:28.655556 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Dec 11 08:27:28 crc kubenswrapper[4860]: I1211 08:27:28.655713 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Dec 11 08:27:28 crc kubenswrapper[4860]: I1211 08:27:28.662835 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Dec 11 08:27:28 crc kubenswrapper[4860]: I1211 08:27:28.831603 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a0838f24-14b5-4018-ad26-8cd0394b4488-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"a0838f24-14b5-4018-ad26-8cd0394b4488\") " pod="openstack/openstack-cell1-galera-0" Dec 11 08:27:28 crc kubenswrapper[4860]: I1211 08:27:28.831699 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0838f24-14b5-4018-ad26-8cd0394b4488-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"a0838f24-14b5-4018-ad26-8cd0394b4488\") " pod="openstack/openstack-cell1-galera-0" Dec 11 08:27:28 crc kubenswrapper[4860]: I1211 08:27:28.831736 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/a0838f24-14b5-4018-ad26-8cd0394b4488-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"a0838f24-14b5-4018-ad26-8cd0394b4488\") " pod="openstack/openstack-cell1-galera-0" Dec 11 08:27:28 crc kubenswrapper[4860]: I1211 08:27:28.831765 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/a0838f24-14b5-4018-ad26-8cd0394b4488-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"a0838f24-14b5-4018-ad26-8cd0394b4488\") " pod="openstack/openstack-cell1-galera-0" Dec 11 08:27:28 crc kubenswrapper[4860]: I1211 08:27:28.831807 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/a0838f24-14b5-4018-ad26-8cd0394b4488-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"a0838f24-14b5-4018-ad26-8cd0394b4488\") " pod="openstack/openstack-cell1-galera-0" Dec 11 08:27:28 crc kubenswrapper[4860]: I1211 08:27:28.831846 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0838f24-14b5-4018-ad26-8cd0394b4488-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"a0838f24-14b5-4018-ad26-8cd0394b4488\") " pod="openstack/openstack-cell1-galera-0" Dec 11 08:27:28 crc kubenswrapper[4860]: I1211 08:27:28.831913 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l68mz\" (UniqueName: \"kubernetes.io/projected/a0838f24-14b5-4018-ad26-8cd0394b4488-kube-api-access-l68mz\") pod \"openstack-cell1-galera-0\" (UID: \"a0838f24-14b5-4018-ad26-8cd0394b4488\") " pod="openstack/openstack-cell1-galera-0" Dec 11 08:27:28 crc kubenswrapper[4860]: I1211 08:27:28.831935 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"openstack-cell1-galera-0\" (UID: \"a0838f24-14b5-4018-ad26-8cd0394b4488\") " pod="openstack/openstack-cell1-galera-0" Dec 11 08:27:28 crc kubenswrapper[4860]: I1211 08:27:28.933549 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l68mz\" (UniqueName: \"kubernetes.io/projected/a0838f24-14b5-4018-ad26-8cd0394b4488-kube-api-access-l68mz\") pod \"openstack-cell1-galera-0\" (UID: \"a0838f24-14b5-4018-ad26-8cd0394b4488\") " pod="openstack/openstack-cell1-galera-0" Dec 11 08:27:28 crc kubenswrapper[4860]: I1211 08:27:28.933619 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"openstack-cell1-galera-0\" (UID: \"a0838f24-14b5-4018-ad26-8cd0394b4488\") " pod="openstack/openstack-cell1-galera-0" Dec 11 08:27:28 crc kubenswrapper[4860]: I1211 08:27:28.933671 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a0838f24-14b5-4018-ad26-8cd0394b4488-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"a0838f24-14b5-4018-ad26-8cd0394b4488\") " pod="openstack/openstack-cell1-galera-0" Dec 11 08:27:28 crc kubenswrapper[4860]: I1211 08:27:28.933692 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0838f24-14b5-4018-ad26-8cd0394b4488-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"a0838f24-14b5-4018-ad26-8cd0394b4488\") " pod="openstack/openstack-cell1-galera-0" Dec 11 08:27:28 crc kubenswrapper[4860]: I1211 08:27:28.933716 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/a0838f24-14b5-4018-ad26-8cd0394b4488-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"a0838f24-14b5-4018-ad26-8cd0394b4488\") " pod="openstack/openstack-cell1-galera-0" Dec 11 08:27:28 crc kubenswrapper[4860]: I1211 08:27:28.933744 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/a0838f24-14b5-4018-ad26-8cd0394b4488-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"a0838f24-14b5-4018-ad26-8cd0394b4488\") " pod="openstack/openstack-cell1-galera-0" Dec 11 08:27:28 crc kubenswrapper[4860]: I1211 08:27:28.933773 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/a0838f24-14b5-4018-ad26-8cd0394b4488-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"a0838f24-14b5-4018-ad26-8cd0394b4488\") " pod="openstack/openstack-cell1-galera-0" Dec 11 08:27:28 crc kubenswrapper[4860]: I1211 08:27:28.933816 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0838f24-14b5-4018-ad26-8cd0394b4488-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"a0838f24-14b5-4018-ad26-8cd0394b4488\") " pod="openstack/openstack-cell1-galera-0" Dec 11 08:27:28 crc kubenswrapper[4860]: I1211 08:27:28.935395 4860 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"openstack-cell1-galera-0\" (UID: \"a0838f24-14b5-4018-ad26-8cd0394b4488\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/openstack-cell1-galera-0" Dec 11 08:27:28 crc kubenswrapper[4860]: I1211 08:27:28.937578 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/a0838f24-14b5-4018-ad26-8cd0394b4488-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"a0838f24-14b5-4018-ad26-8cd0394b4488\") " pod="openstack/openstack-cell1-galera-0" Dec 11 08:27:28 crc kubenswrapper[4860]: I1211 08:27:28.937598 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a0838f24-14b5-4018-ad26-8cd0394b4488-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"a0838f24-14b5-4018-ad26-8cd0394b4488\") " pod="openstack/openstack-cell1-galera-0" Dec 11 08:27:28 crc kubenswrapper[4860]: I1211 08:27:28.937961 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/a0838f24-14b5-4018-ad26-8cd0394b4488-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"a0838f24-14b5-4018-ad26-8cd0394b4488\") " pod="openstack/openstack-cell1-galera-0" Dec 11 08:27:28 crc kubenswrapper[4860]: I1211 08:27:28.938654 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/a0838f24-14b5-4018-ad26-8cd0394b4488-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"a0838f24-14b5-4018-ad26-8cd0394b4488\") " pod="openstack/openstack-cell1-galera-0" Dec 11 08:27:28 crc kubenswrapper[4860]: I1211 08:27:28.947290 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0838f24-14b5-4018-ad26-8cd0394b4488-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"a0838f24-14b5-4018-ad26-8cd0394b4488\") " pod="openstack/openstack-cell1-galera-0" Dec 11 08:27:28 crc kubenswrapper[4860]: I1211 08:27:28.969760 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Dec 11 08:27:28 crc kubenswrapper[4860]: I1211 08:27:28.972694 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 11 08:27:28 crc kubenswrapper[4860]: I1211 08:27:28.973920 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0838f24-14b5-4018-ad26-8cd0394b4488-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"a0838f24-14b5-4018-ad26-8cd0394b4488\") " pod="openstack/openstack-cell1-galera-0" Dec 11 08:27:28 crc kubenswrapper[4860]: I1211 08:27:28.978087 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Dec 11 08:27:28 crc kubenswrapper[4860]: I1211 08:27:28.983742 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Dec 11 08:27:28 crc kubenswrapper[4860]: I1211 08:27:28.992143 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-m7n7c" Dec 11 08:27:28 crc kubenswrapper[4860]: I1211 08:27:28.997257 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 11 08:27:29 crc kubenswrapper[4860]: I1211 08:27:29.024735 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l68mz\" (UniqueName: \"kubernetes.io/projected/a0838f24-14b5-4018-ad26-8cd0394b4488-kube-api-access-l68mz\") pod \"openstack-cell1-galera-0\" (UID: \"a0838f24-14b5-4018-ad26-8cd0394b4488\") " pod="openstack/openstack-cell1-galera-0" Dec 11 08:27:29 crc kubenswrapper[4860]: I1211 08:27:29.065103 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"openstack-cell1-galera-0\" (UID: \"a0838f24-14b5-4018-ad26-8cd0394b4488\") " pod="openstack/openstack-cell1-galera-0" Dec 11 08:27:29 crc kubenswrapper[4860]: I1211 08:27:29.140950 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1752934c-842c-45b6-88ec-0fc804374532-combined-ca-bundle\") pod \"memcached-0\" (UID: \"1752934c-842c-45b6-88ec-0fc804374532\") " pod="openstack/memcached-0" Dec 11 08:27:29 crc kubenswrapper[4860]: I1211 08:27:29.141016 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k967j\" (UniqueName: \"kubernetes.io/projected/1752934c-842c-45b6-88ec-0fc804374532-kube-api-access-k967j\") pod \"memcached-0\" (UID: \"1752934c-842c-45b6-88ec-0fc804374532\") " pod="openstack/memcached-0" Dec 11 08:27:29 crc kubenswrapper[4860]: I1211 08:27:29.141056 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1752934c-842c-45b6-88ec-0fc804374532-config-data\") pod \"memcached-0\" (UID: \"1752934c-842c-45b6-88ec-0fc804374532\") " pod="openstack/memcached-0" Dec 11 08:27:29 crc kubenswrapper[4860]: I1211 08:27:29.141072 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/1752934c-842c-45b6-88ec-0fc804374532-memcached-tls-certs\") pod \"memcached-0\" (UID: \"1752934c-842c-45b6-88ec-0fc804374532\") " pod="openstack/memcached-0" Dec 11 08:27:29 crc kubenswrapper[4860]: I1211 08:27:29.141106 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/1752934c-842c-45b6-88ec-0fc804374532-kolla-config\") pod \"memcached-0\" (UID: \"1752934c-842c-45b6-88ec-0fc804374532\") " pod="openstack/memcached-0" Dec 11 08:27:29 crc kubenswrapper[4860]: I1211 08:27:29.243878 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1752934c-842c-45b6-88ec-0fc804374532-combined-ca-bundle\") pod \"memcached-0\" (UID: \"1752934c-842c-45b6-88ec-0fc804374532\") " pod="openstack/memcached-0" Dec 11 08:27:29 crc kubenswrapper[4860]: I1211 08:27:29.243959 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k967j\" (UniqueName: \"kubernetes.io/projected/1752934c-842c-45b6-88ec-0fc804374532-kube-api-access-k967j\") pod \"memcached-0\" (UID: \"1752934c-842c-45b6-88ec-0fc804374532\") " pod="openstack/memcached-0" Dec 11 08:27:29 crc kubenswrapper[4860]: I1211 08:27:29.244008 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1752934c-842c-45b6-88ec-0fc804374532-config-data\") pod \"memcached-0\" (UID: \"1752934c-842c-45b6-88ec-0fc804374532\") " pod="openstack/memcached-0" Dec 11 08:27:29 crc kubenswrapper[4860]: I1211 08:27:29.244039 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/1752934c-842c-45b6-88ec-0fc804374532-memcached-tls-certs\") pod \"memcached-0\" (UID: \"1752934c-842c-45b6-88ec-0fc804374532\") " pod="openstack/memcached-0" Dec 11 08:27:29 crc kubenswrapper[4860]: I1211 08:27:29.244093 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/1752934c-842c-45b6-88ec-0fc804374532-kolla-config\") pod \"memcached-0\" (UID: \"1752934c-842c-45b6-88ec-0fc804374532\") " pod="openstack/memcached-0" Dec 11 08:27:29 crc kubenswrapper[4860]: I1211 08:27:29.245377 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/1752934c-842c-45b6-88ec-0fc804374532-kolla-config\") pod \"memcached-0\" (UID: \"1752934c-842c-45b6-88ec-0fc804374532\") " pod="openstack/memcached-0" Dec 11 08:27:29 crc kubenswrapper[4860]: I1211 08:27:29.245443 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1752934c-842c-45b6-88ec-0fc804374532-config-data\") pod \"memcached-0\" (UID: \"1752934c-842c-45b6-88ec-0fc804374532\") " pod="openstack/memcached-0" Dec 11 08:27:29 crc kubenswrapper[4860]: I1211 08:27:29.250595 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/1752934c-842c-45b6-88ec-0fc804374532-memcached-tls-certs\") pod \"memcached-0\" (UID: \"1752934c-842c-45b6-88ec-0fc804374532\") " pod="openstack/memcached-0" Dec 11 08:27:29 crc kubenswrapper[4860]: I1211 08:27:29.255533 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1752934c-842c-45b6-88ec-0fc804374532-combined-ca-bundle\") pod \"memcached-0\" (UID: \"1752934c-842c-45b6-88ec-0fc804374532\") " pod="openstack/memcached-0" Dec 11 08:27:29 crc kubenswrapper[4860]: I1211 08:27:29.264633 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k967j\" (UniqueName: \"kubernetes.io/projected/1752934c-842c-45b6-88ec-0fc804374532-kube-api-access-k967j\") pod \"memcached-0\" (UID: \"1752934c-842c-45b6-88ec-0fc804374532\") " pod="openstack/memcached-0" Dec 11 08:27:29 crc kubenswrapper[4860]: I1211 08:27:29.302244 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 11 08:27:29 crc kubenswrapper[4860]: I1211 08:27:29.438181 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 11 08:27:30 crc kubenswrapper[4860]: I1211 08:27:30.726674 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 11 08:27:30 crc kubenswrapper[4860]: I1211 08:27:30.729908 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 11 08:27:30 crc kubenswrapper[4860]: I1211 08:27:30.737447 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-8wb2b" Dec 11 08:27:30 crc kubenswrapper[4860]: I1211 08:27:30.747968 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 11 08:27:30 crc kubenswrapper[4860]: I1211 08:27:30.877501 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p469v\" (UniqueName: \"kubernetes.io/projected/bbb1e117-b966-48e4-bc09-08d66a1ce9eb-kube-api-access-p469v\") pod \"kube-state-metrics-0\" (UID: \"bbb1e117-b966-48e4-bc09-08d66a1ce9eb\") " pod="openstack/kube-state-metrics-0" Dec 11 08:27:30 crc kubenswrapper[4860]: I1211 08:27:30.979163 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p469v\" (UniqueName: \"kubernetes.io/projected/bbb1e117-b966-48e4-bc09-08d66a1ce9eb-kube-api-access-p469v\") pod \"kube-state-metrics-0\" (UID: \"bbb1e117-b966-48e4-bc09-08d66a1ce9eb\") " pod="openstack/kube-state-metrics-0" Dec 11 08:27:31 crc kubenswrapper[4860]: I1211 08:27:31.031626 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p469v\" (UniqueName: \"kubernetes.io/projected/bbb1e117-b966-48e4-bc09-08d66a1ce9eb-kube-api-access-p469v\") pod \"kube-state-metrics-0\" (UID: \"bbb1e117-b966-48e4-bc09-08d66a1ce9eb\") " pod="openstack/kube-state-metrics-0" Dec 11 08:27:31 crc kubenswrapper[4860]: I1211 08:27:31.070496 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 11 08:27:34 crc kubenswrapper[4860]: I1211 08:27:34.029187 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-dbdtz"] Dec 11 08:27:34 crc kubenswrapper[4860]: I1211 08:27:34.030554 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-dbdtz" Dec 11 08:27:34 crc kubenswrapper[4860]: I1211 08:27:34.034130 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Dec 11 08:27:34 crc kubenswrapper[4860]: I1211 08:27:34.034142 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Dec 11 08:27:34 crc kubenswrapper[4860]: I1211 08:27:34.034351 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-k87sp" Dec 11 08:27:34 crc kubenswrapper[4860]: I1211 08:27:34.056865 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-dbdtz"] Dec 11 08:27:34 crc kubenswrapper[4860]: I1211 08:27:34.079865 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-q86ch"] Dec 11 08:27:34 crc kubenswrapper[4860]: I1211 08:27:34.088620 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-q86ch" Dec 11 08:27:34 crc kubenswrapper[4860]: I1211 08:27:34.104715 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-q86ch"] Dec 11 08:27:34 crc kubenswrapper[4860]: I1211 08:27:34.153819 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/e834572f-9650-45d3-b978-c39141b44b74-var-run-ovn\") pod \"ovn-controller-dbdtz\" (UID: \"e834572f-9650-45d3-b978-c39141b44b74\") " pod="openstack/ovn-controller-dbdtz" Dec 11 08:27:34 crc kubenswrapper[4860]: I1211 08:27:34.153869 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e834572f-9650-45d3-b978-c39141b44b74-combined-ca-bundle\") pod \"ovn-controller-dbdtz\" (UID: \"e834572f-9650-45d3-b978-c39141b44b74\") " pod="openstack/ovn-controller-dbdtz" Dec 11 08:27:34 crc kubenswrapper[4860]: I1211 08:27:34.153892 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e834572f-9650-45d3-b978-c39141b44b74-scripts\") pod \"ovn-controller-dbdtz\" (UID: \"e834572f-9650-45d3-b978-c39141b44b74\") " pod="openstack/ovn-controller-dbdtz" Dec 11 08:27:34 crc kubenswrapper[4860]: I1211 08:27:34.153914 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/e73beb7d-c40c-491b-99f8-fabe037df9ba-var-lib\") pod \"ovn-controller-ovs-q86ch\" (UID: \"e73beb7d-c40c-491b-99f8-fabe037df9ba\") " pod="openstack/ovn-controller-ovs-q86ch" Dec 11 08:27:34 crc kubenswrapper[4860]: I1211 08:27:34.153947 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/e73beb7d-c40c-491b-99f8-fabe037df9ba-etc-ovs\") pod \"ovn-controller-ovs-q86ch\" (UID: \"e73beb7d-c40c-491b-99f8-fabe037df9ba\") " pod="openstack/ovn-controller-ovs-q86ch" Dec 11 08:27:34 crc kubenswrapper[4860]: I1211 08:27:34.153968 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e73beb7d-c40c-491b-99f8-fabe037df9ba-var-run\") pod \"ovn-controller-ovs-q86ch\" (UID: \"e73beb7d-c40c-491b-99f8-fabe037df9ba\") " pod="openstack/ovn-controller-ovs-q86ch" Dec 11 08:27:34 crc kubenswrapper[4860]: I1211 08:27:34.154005 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4v8dq\" (UniqueName: \"kubernetes.io/projected/e73beb7d-c40c-491b-99f8-fabe037df9ba-kube-api-access-4v8dq\") pod \"ovn-controller-ovs-q86ch\" (UID: \"e73beb7d-c40c-491b-99f8-fabe037df9ba\") " pod="openstack/ovn-controller-ovs-q86ch" Dec 11 08:27:34 crc kubenswrapper[4860]: I1211 08:27:34.154031 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e834572f-9650-45d3-b978-c39141b44b74-var-run\") pod \"ovn-controller-dbdtz\" (UID: \"e834572f-9650-45d3-b978-c39141b44b74\") " pod="openstack/ovn-controller-dbdtz" Dec 11 08:27:34 crc kubenswrapper[4860]: I1211 08:27:34.154050 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/e834572f-9650-45d3-b978-c39141b44b74-var-log-ovn\") pod \"ovn-controller-dbdtz\" (UID: \"e834572f-9650-45d3-b978-c39141b44b74\") " pod="openstack/ovn-controller-dbdtz" Dec 11 08:27:34 crc kubenswrapper[4860]: I1211 08:27:34.154074 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e73beb7d-c40c-491b-99f8-fabe037df9ba-scripts\") pod \"ovn-controller-ovs-q86ch\" (UID: \"e73beb7d-c40c-491b-99f8-fabe037df9ba\") " pod="openstack/ovn-controller-ovs-q86ch" Dec 11 08:27:34 crc kubenswrapper[4860]: I1211 08:27:34.154107 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8wzpb\" (UniqueName: \"kubernetes.io/projected/e834572f-9650-45d3-b978-c39141b44b74-kube-api-access-8wzpb\") pod \"ovn-controller-dbdtz\" (UID: \"e834572f-9650-45d3-b978-c39141b44b74\") " pod="openstack/ovn-controller-dbdtz" Dec 11 08:27:34 crc kubenswrapper[4860]: I1211 08:27:34.154130 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/e834572f-9650-45d3-b978-c39141b44b74-ovn-controller-tls-certs\") pod \"ovn-controller-dbdtz\" (UID: \"e834572f-9650-45d3-b978-c39141b44b74\") " pod="openstack/ovn-controller-dbdtz" Dec 11 08:27:34 crc kubenswrapper[4860]: I1211 08:27:34.154149 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/e73beb7d-c40c-491b-99f8-fabe037df9ba-var-log\") pod \"ovn-controller-ovs-q86ch\" (UID: \"e73beb7d-c40c-491b-99f8-fabe037df9ba\") " pod="openstack/ovn-controller-ovs-q86ch" Dec 11 08:27:34 crc kubenswrapper[4860]: I1211 08:27:34.255366 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/e834572f-9650-45d3-b978-c39141b44b74-var-run-ovn\") pod \"ovn-controller-dbdtz\" (UID: \"e834572f-9650-45d3-b978-c39141b44b74\") " pod="openstack/ovn-controller-dbdtz" Dec 11 08:27:34 crc kubenswrapper[4860]: I1211 08:27:34.255424 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e834572f-9650-45d3-b978-c39141b44b74-combined-ca-bundle\") pod \"ovn-controller-dbdtz\" (UID: \"e834572f-9650-45d3-b978-c39141b44b74\") " pod="openstack/ovn-controller-dbdtz" Dec 11 08:27:34 crc kubenswrapper[4860]: I1211 08:27:34.255457 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e834572f-9650-45d3-b978-c39141b44b74-scripts\") pod \"ovn-controller-dbdtz\" (UID: \"e834572f-9650-45d3-b978-c39141b44b74\") " pod="openstack/ovn-controller-dbdtz" Dec 11 08:27:34 crc kubenswrapper[4860]: I1211 08:27:34.255491 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/e73beb7d-c40c-491b-99f8-fabe037df9ba-var-lib\") pod \"ovn-controller-ovs-q86ch\" (UID: \"e73beb7d-c40c-491b-99f8-fabe037df9ba\") " pod="openstack/ovn-controller-ovs-q86ch" Dec 11 08:27:34 crc kubenswrapper[4860]: I1211 08:27:34.255527 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/e73beb7d-c40c-491b-99f8-fabe037df9ba-etc-ovs\") pod \"ovn-controller-ovs-q86ch\" (UID: \"e73beb7d-c40c-491b-99f8-fabe037df9ba\") " pod="openstack/ovn-controller-ovs-q86ch" Dec 11 08:27:34 crc kubenswrapper[4860]: I1211 08:27:34.255553 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e73beb7d-c40c-491b-99f8-fabe037df9ba-var-run\") pod \"ovn-controller-ovs-q86ch\" (UID: \"e73beb7d-c40c-491b-99f8-fabe037df9ba\") " pod="openstack/ovn-controller-ovs-q86ch" Dec 11 08:27:34 crc kubenswrapper[4860]: I1211 08:27:34.255592 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4v8dq\" (UniqueName: \"kubernetes.io/projected/e73beb7d-c40c-491b-99f8-fabe037df9ba-kube-api-access-4v8dq\") pod \"ovn-controller-ovs-q86ch\" (UID: \"e73beb7d-c40c-491b-99f8-fabe037df9ba\") " pod="openstack/ovn-controller-ovs-q86ch" Dec 11 08:27:34 crc kubenswrapper[4860]: I1211 08:27:34.255620 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e834572f-9650-45d3-b978-c39141b44b74-var-run\") pod \"ovn-controller-dbdtz\" (UID: \"e834572f-9650-45d3-b978-c39141b44b74\") " pod="openstack/ovn-controller-dbdtz" Dec 11 08:27:34 crc kubenswrapper[4860]: I1211 08:27:34.255636 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/e834572f-9650-45d3-b978-c39141b44b74-var-log-ovn\") pod \"ovn-controller-dbdtz\" (UID: \"e834572f-9650-45d3-b978-c39141b44b74\") " pod="openstack/ovn-controller-dbdtz" Dec 11 08:27:34 crc kubenswrapper[4860]: I1211 08:27:34.255747 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e73beb7d-c40c-491b-99f8-fabe037df9ba-scripts\") pod \"ovn-controller-ovs-q86ch\" (UID: \"e73beb7d-c40c-491b-99f8-fabe037df9ba\") " pod="openstack/ovn-controller-ovs-q86ch" Dec 11 08:27:34 crc kubenswrapper[4860]: I1211 08:27:34.255782 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8wzpb\" (UniqueName: \"kubernetes.io/projected/e834572f-9650-45d3-b978-c39141b44b74-kube-api-access-8wzpb\") pod \"ovn-controller-dbdtz\" (UID: \"e834572f-9650-45d3-b978-c39141b44b74\") " pod="openstack/ovn-controller-dbdtz" Dec 11 08:27:34 crc kubenswrapper[4860]: I1211 08:27:34.255807 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/e834572f-9650-45d3-b978-c39141b44b74-ovn-controller-tls-certs\") pod \"ovn-controller-dbdtz\" (UID: \"e834572f-9650-45d3-b978-c39141b44b74\") " pod="openstack/ovn-controller-dbdtz" Dec 11 08:27:34 crc kubenswrapper[4860]: I1211 08:27:34.255832 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/e73beb7d-c40c-491b-99f8-fabe037df9ba-var-log\") pod \"ovn-controller-ovs-q86ch\" (UID: \"e73beb7d-c40c-491b-99f8-fabe037df9ba\") " pod="openstack/ovn-controller-ovs-q86ch" Dec 11 08:27:34 crc kubenswrapper[4860]: I1211 08:27:34.256530 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/e73beb7d-c40c-491b-99f8-fabe037df9ba-var-log\") pod \"ovn-controller-ovs-q86ch\" (UID: \"e73beb7d-c40c-491b-99f8-fabe037df9ba\") " pod="openstack/ovn-controller-ovs-q86ch" Dec 11 08:27:34 crc kubenswrapper[4860]: I1211 08:27:34.256694 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/e834572f-9650-45d3-b978-c39141b44b74-var-run-ovn\") pod \"ovn-controller-dbdtz\" (UID: \"e834572f-9650-45d3-b978-c39141b44b74\") " pod="openstack/ovn-controller-dbdtz" Dec 11 08:27:34 crc kubenswrapper[4860]: I1211 08:27:34.258139 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/e73beb7d-c40c-491b-99f8-fabe037df9ba-etc-ovs\") pod \"ovn-controller-ovs-q86ch\" (UID: \"e73beb7d-c40c-491b-99f8-fabe037df9ba\") " pod="openstack/ovn-controller-ovs-q86ch" Dec 11 08:27:34 crc kubenswrapper[4860]: I1211 08:27:34.258237 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e73beb7d-c40c-491b-99f8-fabe037df9ba-var-run\") pod \"ovn-controller-ovs-q86ch\" (UID: \"e73beb7d-c40c-491b-99f8-fabe037df9ba\") " pod="openstack/ovn-controller-ovs-q86ch" Dec 11 08:27:34 crc kubenswrapper[4860]: I1211 08:27:34.258951 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e834572f-9650-45d3-b978-c39141b44b74-var-run\") pod \"ovn-controller-dbdtz\" (UID: \"e834572f-9650-45d3-b978-c39141b44b74\") " pod="openstack/ovn-controller-dbdtz" Dec 11 08:27:34 crc kubenswrapper[4860]: I1211 08:27:34.259265 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/e834572f-9650-45d3-b978-c39141b44b74-var-log-ovn\") pod \"ovn-controller-dbdtz\" (UID: \"e834572f-9650-45d3-b978-c39141b44b74\") " pod="openstack/ovn-controller-dbdtz" Dec 11 08:27:34 crc kubenswrapper[4860]: I1211 08:27:34.260213 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e834572f-9650-45d3-b978-c39141b44b74-scripts\") pod \"ovn-controller-dbdtz\" (UID: \"e834572f-9650-45d3-b978-c39141b44b74\") " pod="openstack/ovn-controller-dbdtz" Dec 11 08:27:34 crc kubenswrapper[4860]: I1211 08:27:34.260526 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e73beb7d-c40c-491b-99f8-fabe037df9ba-scripts\") pod \"ovn-controller-ovs-q86ch\" (UID: \"e73beb7d-c40c-491b-99f8-fabe037df9ba\") " pod="openstack/ovn-controller-ovs-q86ch" Dec 11 08:27:34 crc kubenswrapper[4860]: I1211 08:27:34.260724 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/e73beb7d-c40c-491b-99f8-fabe037df9ba-var-lib\") pod \"ovn-controller-ovs-q86ch\" (UID: \"e73beb7d-c40c-491b-99f8-fabe037df9ba\") " pod="openstack/ovn-controller-ovs-q86ch" Dec 11 08:27:34 crc kubenswrapper[4860]: I1211 08:27:34.277611 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/e834572f-9650-45d3-b978-c39141b44b74-ovn-controller-tls-certs\") pod \"ovn-controller-dbdtz\" (UID: \"e834572f-9650-45d3-b978-c39141b44b74\") " pod="openstack/ovn-controller-dbdtz" Dec 11 08:27:34 crc kubenswrapper[4860]: I1211 08:27:34.289986 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e834572f-9650-45d3-b978-c39141b44b74-combined-ca-bundle\") pod \"ovn-controller-dbdtz\" (UID: \"e834572f-9650-45d3-b978-c39141b44b74\") " pod="openstack/ovn-controller-dbdtz" Dec 11 08:27:34 crc kubenswrapper[4860]: I1211 08:27:34.294878 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8wzpb\" (UniqueName: \"kubernetes.io/projected/e834572f-9650-45d3-b978-c39141b44b74-kube-api-access-8wzpb\") pod \"ovn-controller-dbdtz\" (UID: \"e834572f-9650-45d3-b978-c39141b44b74\") " pod="openstack/ovn-controller-dbdtz" Dec 11 08:27:34 crc kubenswrapper[4860]: I1211 08:27:34.311628 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4v8dq\" (UniqueName: \"kubernetes.io/projected/e73beb7d-c40c-491b-99f8-fabe037df9ba-kube-api-access-4v8dq\") pod \"ovn-controller-ovs-q86ch\" (UID: \"e73beb7d-c40c-491b-99f8-fabe037df9ba\") " pod="openstack/ovn-controller-ovs-q86ch" Dec 11 08:27:34 crc kubenswrapper[4860]: I1211 08:27:34.372495 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-dbdtz" Dec 11 08:27:34 crc kubenswrapper[4860]: I1211 08:27:34.427797 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-q86ch" Dec 11 08:27:37 crc kubenswrapper[4860]: I1211 08:27:37.383432 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 11 08:27:37 crc kubenswrapper[4860]: I1211 08:27:37.386006 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 11 08:27:37 crc kubenswrapper[4860]: I1211 08:27:37.390403 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Dec 11 08:27:37 crc kubenswrapper[4860]: I1211 08:27:37.390737 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-x69sr" Dec 11 08:27:37 crc kubenswrapper[4860]: I1211 08:27:37.390876 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Dec 11 08:27:37 crc kubenswrapper[4860]: I1211 08:27:37.391520 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Dec 11 08:27:37 crc kubenswrapper[4860]: I1211 08:27:37.408679 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Dec 11 08:27:37 crc kubenswrapper[4860]: I1211 08:27:37.414388 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 11 08:27:37 crc kubenswrapper[4860]: I1211 08:27:37.552166 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9657082d-ff9c-4a54-8b4d-54e6ea411590-config\") pod \"ovsdbserver-sb-0\" (UID: \"9657082d-ff9c-4a54-8b4d-54e6ea411590\") " pod="openstack/ovsdbserver-sb-0" Dec 11 08:27:37 crc kubenswrapper[4860]: I1211 08:27:37.552254 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/9657082d-ff9c-4a54-8b4d-54e6ea411590-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"9657082d-ff9c-4a54-8b4d-54e6ea411590\") " pod="openstack/ovsdbserver-sb-0" Dec 11 08:27:37 crc kubenswrapper[4860]: I1211 08:27:37.552288 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9657082d-ff9c-4a54-8b4d-54e6ea411590-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"9657082d-ff9c-4a54-8b4d-54e6ea411590\") " pod="openstack/ovsdbserver-sb-0" Dec 11 08:27:37 crc kubenswrapper[4860]: I1211 08:27:37.552321 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9657082d-ff9c-4a54-8b4d-54e6ea411590-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"9657082d-ff9c-4a54-8b4d-54e6ea411590\") " pod="openstack/ovsdbserver-sb-0" Dec 11 08:27:37 crc kubenswrapper[4860]: I1211 08:27:37.552361 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/9657082d-ff9c-4a54-8b4d-54e6ea411590-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"9657082d-ff9c-4a54-8b4d-54e6ea411590\") " pod="openstack/ovsdbserver-sb-0" Dec 11 08:27:37 crc kubenswrapper[4860]: I1211 08:27:37.552381 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/9657082d-ff9c-4a54-8b4d-54e6ea411590-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"9657082d-ff9c-4a54-8b4d-54e6ea411590\") " pod="openstack/ovsdbserver-sb-0" Dec 11 08:27:37 crc kubenswrapper[4860]: I1211 08:27:37.552418 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-sb-0\" (UID: \"9657082d-ff9c-4a54-8b4d-54e6ea411590\") " pod="openstack/ovsdbserver-sb-0" Dec 11 08:27:37 crc kubenswrapper[4860]: I1211 08:27:37.552461 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nvrht\" (UniqueName: \"kubernetes.io/projected/9657082d-ff9c-4a54-8b4d-54e6ea411590-kube-api-access-nvrht\") pod \"ovsdbserver-sb-0\" (UID: \"9657082d-ff9c-4a54-8b4d-54e6ea411590\") " pod="openstack/ovsdbserver-sb-0" Dec 11 08:27:37 crc kubenswrapper[4860]: I1211 08:27:37.653973 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-sb-0\" (UID: \"9657082d-ff9c-4a54-8b4d-54e6ea411590\") " pod="openstack/ovsdbserver-sb-0" Dec 11 08:27:37 crc kubenswrapper[4860]: I1211 08:27:37.654075 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nvrht\" (UniqueName: \"kubernetes.io/projected/9657082d-ff9c-4a54-8b4d-54e6ea411590-kube-api-access-nvrht\") pod \"ovsdbserver-sb-0\" (UID: \"9657082d-ff9c-4a54-8b4d-54e6ea411590\") " pod="openstack/ovsdbserver-sb-0" Dec 11 08:27:37 crc kubenswrapper[4860]: I1211 08:27:37.654135 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9657082d-ff9c-4a54-8b4d-54e6ea411590-config\") pod \"ovsdbserver-sb-0\" (UID: \"9657082d-ff9c-4a54-8b4d-54e6ea411590\") " pod="openstack/ovsdbserver-sb-0" Dec 11 08:27:37 crc kubenswrapper[4860]: I1211 08:27:37.654195 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/9657082d-ff9c-4a54-8b4d-54e6ea411590-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"9657082d-ff9c-4a54-8b4d-54e6ea411590\") " pod="openstack/ovsdbserver-sb-0" Dec 11 08:27:37 crc kubenswrapper[4860]: I1211 08:27:37.654232 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9657082d-ff9c-4a54-8b4d-54e6ea411590-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"9657082d-ff9c-4a54-8b4d-54e6ea411590\") " pod="openstack/ovsdbserver-sb-0" Dec 11 08:27:37 crc kubenswrapper[4860]: I1211 08:27:37.654262 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9657082d-ff9c-4a54-8b4d-54e6ea411590-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"9657082d-ff9c-4a54-8b4d-54e6ea411590\") " pod="openstack/ovsdbserver-sb-0" Dec 11 08:27:37 crc kubenswrapper[4860]: I1211 08:27:37.654337 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/9657082d-ff9c-4a54-8b4d-54e6ea411590-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"9657082d-ff9c-4a54-8b4d-54e6ea411590\") " pod="openstack/ovsdbserver-sb-0" Dec 11 08:27:37 crc kubenswrapper[4860]: I1211 08:27:37.654359 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/9657082d-ff9c-4a54-8b4d-54e6ea411590-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"9657082d-ff9c-4a54-8b4d-54e6ea411590\") " pod="openstack/ovsdbserver-sb-0" Dec 11 08:27:37 crc kubenswrapper[4860]: I1211 08:27:37.654532 4860 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-sb-0\" (UID: \"9657082d-ff9c-4a54-8b4d-54e6ea411590\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/ovsdbserver-sb-0" Dec 11 08:27:37 crc kubenswrapper[4860]: I1211 08:27:37.656294 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/9657082d-ff9c-4a54-8b4d-54e6ea411590-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"9657082d-ff9c-4a54-8b4d-54e6ea411590\") " pod="openstack/ovsdbserver-sb-0" Dec 11 08:27:37 crc kubenswrapper[4860]: I1211 08:27:37.657244 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9657082d-ff9c-4a54-8b4d-54e6ea411590-config\") pod \"ovsdbserver-sb-0\" (UID: \"9657082d-ff9c-4a54-8b4d-54e6ea411590\") " pod="openstack/ovsdbserver-sb-0" Dec 11 08:27:37 crc kubenswrapper[4860]: I1211 08:27:37.657429 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9657082d-ff9c-4a54-8b4d-54e6ea411590-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"9657082d-ff9c-4a54-8b4d-54e6ea411590\") " pod="openstack/ovsdbserver-sb-0" Dec 11 08:27:37 crc kubenswrapper[4860]: I1211 08:27:37.672957 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/9657082d-ff9c-4a54-8b4d-54e6ea411590-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"9657082d-ff9c-4a54-8b4d-54e6ea411590\") " pod="openstack/ovsdbserver-sb-0" Dec 11 08:27:37 crc kubenswrapper[4860]: I1211 08:27:37.673233 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/9657082d-ff9c-4a54-8b4d-54e6ea411590-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"9657082d-ff9c-4a54-8b4d-54e6ea411590\") " pod="openstack/ovsdbserver-sb-0" Dec 11 08:27:37 crc kubenswrapper[4860]: I1211 08:27:37.678540 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9657082d-ff9c-4a54-8b4d-54e6ea411590-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"9657082d-ff9c-4a54-8b4d-54e6ea411590\") " pod="openstack/ovsdbserver-sb-0" Dec 11 08:27:37 crc kubenswrapper[4860]: I1211 08:27:37.679594 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nvrht\" (UniqueName: \"kubernetes.io/projected/9657082d-ff9c-4a54-8b4d-54e6ea411590-kube-api-access-nvrht\") pod \"ovsdbserver-sb-0\" (UID: \"9657082d-ff9c-4a54-8b4d-54e6ea411590\") " pod="openstack/ovsdbserver-sb-0" Dec 11 08:27:37 crc kubenswrapper[4860]: I1211 08:27:37.697978 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-sb-0\" (UID: \"9657082d-ff9c-4a54-8b4d-54e6ea411590\") " pod="openstack/ovsdbserver-sb-0" Dec 11 08:27:37 crc kubenswrapper[4860]: I1211 08:27:37.726240 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 11 08:27:38 crc kubenswrapper[4860]: I1211 08:27:38.108122 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 11 08:27:38 crc kubenswrapper[4860]: I1211 08:27:38.116498 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 11 08:27:38 crc kubenswrapper[4860]: I1211 08:27:38.117668 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 11 08:27:38 crc kubenswrapper[4860]: I1211 08:27:38.121024 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Dec 11 08:27:38 crc kubenswrapper[4860]: I1211 08:27:38.129074 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Dec 11 08:27:38 crc kubenswrapper[4860]: I1211 08:27:38.130248 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-rz8pt" Dec 11 08:27:38 crc kubenswrapper[4860]: I1211 08:27:38.130713 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Dec 11 08:27:38 crc kubenswrapper[4860]: I1211 08:27:38.303964 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c017f36f-48a3-4919-ac22-6a85a5c73530-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"c017f36f-48a3-4919-ac22-6a85a5c73530\") " pod="openstack/ovsdbserver-nb-0" Dec 11 08:27:38 crc kubenswrapper[4860]: I1211 08:27:38.304162 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c017f36f-48a3-4919-ac22-6a85a5c73530-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"c017f36f-48a3-4919-ac22-6a85a5c73530\") " pod="openstack/ovsdbserver-nb-0" Dec 11 08:27:38 crc kubenswrapper[4860]: I1211 08:27:38.304194 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jntjv\" (UniqueName: \"kubernetes.io/projected/c017f36f-48a3-4919-ac22-6a85a5c73530-kube-api-access-jntjv\") pod \"ovsdbserver-nb-0\" (UID: \"c017f36f-48a3-4919-ac22-6a85a5c73530\") " pod="openstack/ovsdbserver-nb-0" Dec 11 08:27:38 crc kubenswrapper[4860]: I1211 08:27:38.304229 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c017f36f-48a3-4919-ac22-6a85a5c73530-config\") pod \"ovsdbserver-nb-0\" (UID: \"c017f36f-48a3-4919-ac22-6a85a5c73530\") " pod="openstack/ovsdbserver-nb-0" Dec 11 08:27:38 crc kubenswrapper[4860]: I1211 08:27:38.304255 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c017f36f-48a3-4919-ac22-6a85a5c73530-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"c017f36f-48a3-4919-ac22-6a85a5c73530\") " pod="openstack/ovsdbserver-nb-0" Dec 11 08:27:38 crc kubenswrapper[4860]: I1211 08:27:38.304298 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"ovsdbserver-nb-0\" (UID: \"c017f36f-48a3-4919-ac22-6a85a5c73530\") " pod="openstack/ovsdbserver-nb-0" Dec 11 08:27:38 crc kubenswrapper[4860]: I1211 08:27:38.304319 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c017f36f-48a3-4919-ac22-6a85a5c73530-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"c017f36f-48a3-4919-ac22-6a85a5c73530\") " pod="openstack/ovsdbserver-nb-0" Dec 11 08:27:38 crc kubenswrapper[4860]: I1211 08:27:38.304362 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c017f36f-48a3-4919-ac22-6a85a5c73530-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"c017f36f-48a3-4919-ac22-6a85a5c73530\") " pod="openstack/ovsdbserver-nb-0" Dec 11 08:27:38 crc kubenswrapper[4860]: I1211 08:27:38.408713 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c017f36f-48a3-4919-ac22-6a85a5c73530-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"c017f36f-48a3-4919-ac22-6a85a5c73530\") " pod="openstack/ovsdbserver-nb-0" Dec 11 08:27:38 crc kubenswrapper[4860]: I1211 08:27:38.408979 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c017f36f-48a3-4919-ac22-6a85a5c73530-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"c017f36f-48a3-4919-ac22-6a85a5c73530\") " pod="openstack/ovsdbserver-nb-0" Dec 11 08:27:38 crc kubenswrapper[4860]: I1211 08:27:38.409017 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jntjv\" (UniqueName: \"kubernetes.io/projected/c017f36f-48a3-4919-ac22-6a85a5c73530-kube-api-access-jntjv\") pod \"ovsdbserver-nb-0\" (UID: \"c017f36f-48a3-4919-ac22-6a85a5c73530\") " pod="openstack/ovsdbserver-nb-0" Dec 11 08:27:38 crc kubenswrapper[4860]: I1211 08:27:38.409062 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c017f36f-48a3-4919-ac22-6a85a5c73530-config\") pod \"ovsdbserver-nb-0\" (UID: \"c017f36f-48a3-4919-ac22-6a85a5c73530\") " pod="openstack/ovsdbserver-nb-0" Dec 11 08:27:38 crc kubenswrapper[4860]: I1211 08:27:38.409099 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c017f36f-48a3-4919-ac22-6a85a5c73530-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"c017f36f-48a3-4919-ac22-6a85a5c73530\") " pod="openstack/ovsdbserver-nb-0" Dec 11 08:27:38 crc kubenswrapper[4860]: I1211 08:27:38.409162 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"ovsdbserver-nb-0\" (UID: \"c017f36f-48a3-4919-ac22-6a85a5c73530\") " pod="openstack/ovsdbserver-nb-0" Dec 11 08:27:38 crc kubenswrapper[4860]: I1211 08:27:38.409194 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c017f36f-48a3-4919-ac22-6a85a5c73530-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"c017f36f-48a3-4919-ac22-6a85a5c73530\") " pod="openstack/ovsdbserver-nb-0" Dec 11 08:27:38 crc kubenswrapper[4860]: I1211 08:27:38.409303 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c017f36f-48a3-4919-ac22-6a85a5c73530-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"c017f36f-48a3-4919-ac22-6a85a5c73530\") " pod="openstack/ovsdbserver-nb-0" Dec 11 08:27:38 crc kubenswrapper[4860]: I1211 08:27:38.409668 4860 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"ovsdbserver-nb-0\" (UID: \"c017f36f-48a3-4919-ac22-6a85a5c73530\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/ovsdbserver-nb-0" Dec 11 08:27:38 crc kubenswrapper[4860]: I1211 08:27:38.410155 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c017f36f-48a3-4919-ac22-6a85a5c73530-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"c017f36f-48a3-4919-ac22-6a85a5c73530\") " pod="openstack/ovsdbserver-nb-0" Dec 11 08:27:38 crc kubenswrapper[4860]: I1211 08:27:38.410960 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c017f36f-48a3-4919-ac22-6a85a5c73530-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"c017f36f-48a3-4919-ac22-6a85a5c73530\") " pod="openstack/ovsdbserver-nb-0" Dec 11 08:27:38 crc kubenswrapper[4860]: I1211 08:27:38.411422 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c017f36f-48a3-4919-ac22-6a85a5c73530-config\") pod \"ovsdbserver-nb-0\" (UID: \"c017f36f-48a3-4919-ac22-6a85a5c73530\") " pod="openstack/ovsdbserver-nb-0" Dec 11 08:27:38 crc kubenswrapper[4860]: I1211 08:27:38.413958 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c017f36f-48a3-4919-ac22-6a85a5c73530-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"c017f36f-48a3-4919-ac22-6a85a5c73530\") " pod="openstack/ovsdbserver-nb-0" Dec 11 08:27:38 crc kubenswrapper[4860]: I1211 08:27:38.415437 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c017f36f-48a3-4919-ac22-6a85a5c73530-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"c017f36f-48a3-4919-ac22-6a85a5c73530\") " pod="openstack/ovsdbserver-nb-0" Dec 11 08:27:38 crc kubenswrapper[4860]: I1211 08:27:38.426583 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c017f36f-48a3-4919-ac22-6a85a5c73530-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"c017f36f-48a3-4919-ac22-6a85a5c73530\") " pod="openstack/ovsdbserver-nb-0" Dec 11 08:27:38 crc kubenswrapper[4860]: I1211 08:27:38.431266 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jntjv\" (UniqueName: \"kubernetes.io/projected/c017f36f-48a3-4919-ac22-6a85a5c73530-kube-api-access-jntjv\") pod \"ovsdbserver-nb-0\" (UID: \"c017f36f-48a3-4919-ac22-6a85a5c73530\") " pod="openstack/ovsdbserver-nb-0" Dec 11 08:27:38 crc kubenswrapper[4860]: I1211 08:27:38.436496 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"ovsdbserver-nb-0\" (UID: \"c017f36f-48a3-4919-ac22-6a85a5c73530\") " pod="openstack/ovsdbserver-nb-0" Dec 11 08:27:38 crc kubenswrapper[4860]: I1211 08:27:38.444740 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 11 08:27:38 crc kubenswrapper[4860]: I1211 08:27:38.795827 4860 patch_prober.go:28] interesting pod/machine-config-daemon-99qgp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 08:27:38 crc kubenswrapper[4860]: I1211 08:27:38.795913 4860 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 08:27:43 crc kubenswrapper[4860]: E1211 08:27:43.161433 4860 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 11 08:27:43 crc kubenswrapper[4860]: E1211 08:27:43.162322 4860 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zfgcm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-w78r8_openstack(375a2a30-43d5-4557-be42-44a862c171c6): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 11 08:27:43 crc kubenswrapper[4860]: E1211 08:27:43.163621 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-w78r8" podUID="375a2a30-43d5-4557-be42-44a862c171c6" Dec 11 08:27:43 crc kubenswrapper[4860]: E1211 08:27:43.184377 4860 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 11 08:27:43 crc kubenswrapper[4860]: E1211 08:27:43.184606 4860 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4pfgv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-wt4jr_openstack(79fac2d1-5332-496d-86c9-b82df3c9209a): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 11 08:27:43 crc kubenswrapper[4860]: E1211 08:27:43.185925 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-wt4jr" podUID="79fac2d1-5332-496d-86c9-b82df3c9209a" Dec 11 08:27:43 crc kubenswrapper[4860]: E1211 08:27:43.205000 4860 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 11 08:27:43 crc kubenswrapper[4860]: E1211 08:27:43.205328 4860 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rgh8v,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-666b6646f7-vsf77_openstack(5ea9b400-63b3-480e-880e-978b406fce50): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 11 08:27:43 crc kubenswrapper[4860]: E1211 08:27:43.206613 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-666b6646f7-vsf77" podUID="5ea9b400-63b3-480e-880e-978b406fce50" Dec 11 08:27:43 crc kubenswrapper[4860]: E1211 08:27:43.220441 4860 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 11 08:27:43 crc kubenswrapper[4860]: E1211 08:27:43.220789 4860 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n659h4h664hbh658h587h67ch89h587h8fh679hc6hf9h55fh644h5d5h698h68dh5cdh5ffh669h54ch9h689hb8hd4h5bfhd8h5d7h5fh665h574q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-89cmp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-57d769cc4f-vwqtx_openstack(7e94e216-32ea-44d2-b7ae-f765ebbdd524): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 11 08:27:43 crc kubenswrapper[4860]: E1211 08:27:43.222832 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-57d769cc4f-vwqtx" podUID="7e94e216-32ea-44d2-b7ae-f765ebbdd524" Dec 11 08:27:43 crc kubenswrapper[4860]: E1211 08:27:43.247381 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-57d769cc4f-vwqtx" podUID="7e94e216-32ea-44d2-b7ae-f765ebbdd524" Dec 11 08:27:43 crc kubenswrapper[4860]: E1211 08:27:43.247822 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-666b6646f7-vsf77" podUID="5ea9b400-63b3-480e-880e-978b406fce50" Dec 11 08:27:44 crc kubenswrapper[4860]: I1211 08:27:44.888814 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-wt4jr" Dec 11 08:27:44 crc kubenswrapper[4860]: I1211 08:27:44.929032 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-w78r8" Dec 11 08:27:44 crc kubenswrapper[4860]: I1211 08:27:44.971852 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/79fac2d1-5332-496d-86c9-b82df3c9209a-config\") pod \"79fac2d1-5332-496d-86c9-b82df3c9209a\" (UID: \"79fac2d1-5332-496d-86c9-b82df3c9209a\") " Dec 11 08:27:44 crc kubenswrapper[4860]: I1211 08:27:44.971963 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/79fac2d1-5332-496d-86c9-b82df3c9209a-dns-svc\") pod \"79fac2d1-5332-496d-86c9-b82df3c9209a\" (UID: \"79fac2d1-5332-496d-86c9-b82df3c9209a\") " Dec 11 08:27:44 crc kubenswrapper[4860]: I1211 08:27:44.972010 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4pfgv\" (UniqueName: \"kubernetes.io/projected/79fac2d1-5332-496d-86c9-b82df3c9209a-kube-api-access-4pfgv\") pod \"79fac2d1-5332-496d-86c9-b82df3c9209a\" (UID: \"79fac2d1-5332-496d-86c9-b82df3c9209a\") " Dec 11 08:27:44 crc kubenswrapper[4860]: I1211 08:27:44.973050 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/79fac2d1-5332-496d-86c9-b82df3c9209a-config" (OuterVolumeSpecName: "config") pod "79fac2d1-5332-496d-86c9-b82df3c9209a" (UID: "79fac2d1-5332-496d-86c9-b82df3c9209a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:27:44 crc kubenswrapper[4860]: I1211 08:27:44.973066 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/79fac2d1-5332-496d-86c9-b82df3c9209a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "79fac2d1-5332-496d-86c9-b82df3c9209a" (UID: "79fac2d1-5332-496d-86c9-b82df3c9209a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:27:44 crc kubenswrapper[4860]: I1211 08:27:44.982995 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/79fac2d1-5332-496d-86c9-b82df3c9209a-kube-api-access-4pfgv" (OuterVolumeSpecName: "kube-api-access-4pfgv") pod "79fac2d1-5332-496d-86c9-b82df3c9209a" (UID: "79fac2d1-5332-496d-86c9-b82df3c9209a"). InnerVolumeSpecName "kube-api-access-4pfgv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:27:45 crc kubenswrapper[4860]: I1211 08:27:45.073999 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zfgcm\" (UniqueName: \"kubernetes.io/projected/375a2a30-43d5-4557-be42-44a862c171c6-kube-api-access-zfgcm\") pod \"375a2a30-43d5-4557-be42-44a862c171c6\" (UID: \"375a2a30-43d5-4557-be42-44a862c171c6\") " Dec 11 08:27:45 crc kubenswrapper[4860]: I1211 08:27:45.074080 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/375a2a30-43d5-4557-be42-44a862c171c6-config\") pod \"375a2a30-43d5-4557-be42-44a862c171c6\" (UID: \"375a2a30-43d5-4557-be42-44a862c171c6\") " Dec 11 08:27:45 crc kubenswrapper[4860]: I1211 08:27:45.074617 4860 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/79fac2d1-5332-496d-86c9-b82df3c9209a-config\") on node \"crc\" DevicePath \"\"" Dec 11 08:27:45 crc kubenswrapper[4860]: I1211 08:27:45.074634 4860 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/79fac2d1-5332-496d-86c9-b82df3c9209a-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 11 08:27:45 crc kubenswrapper[4860]: I1211 08:27:45.074669 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4pfgv\" (UniqueName: \"kubernetes.io/projected/79fac2d1-5332-496d-86c9-b82df3c9209a-kube-api-access-4pfgv\") on node \"crc\" DevicePath \"\"" Dec 11 08:27:45 crc kubenswrapper[4860]: I1211 08:27:45.075227 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/375a2a30-43d5-4557-be42-44a862c171c6-config" (OuterVolumeSpecName: "config") pod "375a2a30-43d5-4557-be42-44a862c171c6" (UID: "375a2a30-43d5-4557-be42-44a862c171c6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:27:45 crc kubenswrapper[4860]: I1211 08:27:45.079363 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/375a2a30-43d5-4557-be42-44a862c171c6-kube-api-access-zfgcm" (OuterVolumeSpecName: "kube-api-access-zfgcm") pod "375a2a30-43d5-4557-be42-44a862c171c6" (UID: "375a2a30-43d5-4557-be42-44a862c171c6"). InnerVolumeSpecName "kube-api-access-zfgcm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:27:45 crc kubenswrapper[4860]: I1211 08:27:45.142613 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 11 08:27:45 crc kubenswrapper[4860]: I1211 08:27:45.176260 4860 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/375a2a30-43d5-4557-be42-44a862c171c6-config\") on node \"crc\" DevicePath \"\"" Dec 11 08:27:45 crc kubenswrapper[4860]: I1211 08:27:45.176302 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zfgcm\" (UniqueName: \"kubernetes.io/projected/375a2a30-43d5-4557-be42-44a862c171c6-kube-api-access-zfgcm\") on node \"crc\" DevicePath \"\"" Dec 11 08:27:45 crc kubenswrapper[4860]: I1211 08:27:45.246476 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 11 08:27:45 crc kubenswrapper[4860]: I1211 08:27:45.264499 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"c0d5e719-9d28-4e31-9349-f882c4b72567","Type":"ContainerStarted","Data":"fdfaa55791f1a45eca6413294bac55be91570ce8b6225b358161f8397e2f7c81"} Dec 11 08:27:45 crc kubenswrapper[4860]: I1211 08:27:45.266130 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"1752934c-842c-45b6-88ec-0fc804374532","Type":"ContainerStarted","Data":"b75fb6201499a4c2496e145968a783ec31cd08dba3d54c7ecfce4e7b45a5e074"} Dec 11 08:27:45 crc kubenswrapper[4860]: I1211 08:27:45.267084 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-w78r8" event={"ID":"375a2a30-43d5-4557-be42-44a862c171c6","Type":"ContainerDied","Data":"f7db3763bed5d831657c1f12305c76e0924691e1ecc7060cc05b12b83ce81d38"} Dec 11 08:27:45 crc kubenswrapper[4860]: I1211 08:27:45.267255 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-w78r8" Dec 11 08:27:45 crc kubenswrapper[4860]: I1211 08:27:45.268407 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-wt4jr" event={"ID":"79fac2d1-5332-496d-86c9-b82df3c9209a","Type":"ContainerDied","Data":"5a1c218decb2a91b10a6ae30a4c33885dfd804d66fd99ecc270e7728c34d5bf3"} Dec 11 08:27:45 crc kubenswrapper[4860]: I1211 08:27:45.268488 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-wt4jr" Dec 11 08:27:45 crc kubenswrapper[4860]: I1211 08:27:45.352686 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-w78r8"] Dec 11 08:27:45 crc kubenswrapper[4860]: I1211 08:27:45.372029 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-w78r8"] Dec 11 08:27:45 crc kubenswrapper[4860]: I1211 08:27:45.392152 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-dbdtz"] Dec 11 08:27:45 crc kubenswrapper[4860]: I1211 08:27:45.400886 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 11 08:27:45 crc kubenswrapper[4860]: I1211 08:27:45.407375 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-wt4jr"] Dec 11 08:27:45 crc kubenswrapper[4860]: I1211 08:27:45.415578 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-wt4jr"] Dec 11 08:27:45 crc kubenswrapper[4860]: I1211 08:27:45.456118 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-q86ch"] Dec 11 08:27:45 crc kubenswrapper[4860]: I1211 08:27:45.505937 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 11 08:27:45 crc kubenswrapper[4860]: I1211 08:27:45.593108 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="375a2a30-43d5-4557-be42-44a862c171c6" path="/var/lib/kubelet/pods/375a2a30-43d5-4557-be42-44a862c171c6/volumes" Dec 11 08:27:45 crc kubenswrapper[4860]: I1211 08:27:45.594292 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="79fac2d1-5332-496d-86c9-b82df3c9209a" path="/var/lib/kubelet/pods/79fac2d1-5332-496d-86c9-b82df3c9209a/volumes" Dec 11 08:27:45 crc kubenswrapper[4860]: I1211 08:27:45.715854 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 11 08:27:45 crc kubenswrapper[4860]: W1211 08:27:45.786114 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbbb1e117_b966_48e4_bc09_08d66a1ce9eb.slice/crio-5fafe2235e6146e5bdcc9cb3c753ad5723d2ba870a789157311460891393f8d7 WatchSource:0}: Error finding container 5fafe2235e6146e5bdcc9cb3c753ad5723d2ba870a789157311460891393f8d7: Status 404 returned error can't find the container with id 5fafe2235e6146e5bdcc9cb3c753ad5723d2ba870a789157311460891393f8d7 Dec 11 08:27:45 crc kubenswrapper[4860]: W1211 08:27:45.791154 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode73beb7d_c40c_491b_99f8_fabe037df9ba.slice/crio-929591ff16c83a565442fafac456f70a9fec844eaaa0495f7a275794871e1140 WatchSource:0}: Error finding container 929591ff16c83a565442fafac456f70a9fec844eaaa0495f7a275794871e1140: Status 404 returned error can't find the container with id 929591ff16c83a565442fafac456f70a9fec844eaaa0495f7a275794871e1140 Dec 11 08:27:45 crc kubenswrapper[4860]: W1211 08:27:45.887546 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda0838f24_14b5_4018_ad26_8cd0394b4488.slice/crio-e9c82794516e63736e4d2a9b7de0ecd8f5a25b76bda95793ddae5a840f7a9bf4 WatchSource:0}: Error finding container e9c82794516e63736e4d2a9b7de0ecd8f5a25b76bda95793ddae5a840f7a9bf4: Status 404 returned error can't find the container with id e9c82794516e63736e4d2a9b7de0ecd8f5a25b76bda95793ddae5a840f7a9bf4 Dec 11 08:27:46 crc kubenswrapper[4860]: I1211 08:27:46.278832 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-dbdtz" event={"ID":"e834572f-9650-45d3-b978-c39141b44b74","Type":"ContainerStarted","Data":"0b1bd9af03a4d6e06217de9e214387ac58c655bb44678c95b64fc196f931e5aa"} Dec 11 08:27:46 crc kubenswrapper[4860]: I1211 08:27:46.280068 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"bbb1e117-b966-48e4-bc09-08d66a1ce9eb","Type":"ContainerStarted","Data":"5fafe2235e6146e5bdcc9cb3c753ad5723d2ba870a789157311460891393f8d7"} Dec 11 08:27:46 crc kubenswrapper[4860]: I1211 08:27:46.281736 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"a0838f24-14b5-4018-ad26-8cd0394b4488","Type":"ContainerStarted","Data":"e9c82794516e63736e4d2a9b7de0ecd8f5a25b76bda95793ddae5a840f7a9bf4"} Dec 11 08:27:46 crc kubenswrapper[4860]: I1211 08:27:46.283234 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-q86ch" event={"ID":"e73beb7d-c40c-491b-99f8-fabe037df9ba","Type":"ContainerStarted","Data":"929591ff16c83a565442fafac456f70a9fec844eaaa0495f7a275794871e1140"} Dec 11 08:27:46 crc kubenswrapper[4860]: I1211 08:27:46.284856 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"9657082d-ff9c-4a54-8b4d-54e6ea411590","Type":"ContainerStarted","Data":"e1cff939030130bbdc44fca737dea416016e19d46edbe24fa0e4d14a352d4c7a"} Dec 11 08:27:46 crc kubenswrapper[4860]: I1211 08:27:46.601122 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 11 08:27:47 crc kubenswrapper[4860]: I1211 08:27:47.294879 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"a004cb7b-ee08-44c0-9dc7-d0b69509fe13","Type":"ContainerStarted","Data":"7b2484370445a6fcc6a996d0a46f40caeeb99158af08c51c6b35e54bd1ef3322"} Dec 11 08:27:47 crc kubenswrapper[4860]: I1211 08:27:47.297840 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"97634ec0-7953-4e38-a07e-95967ce874d7","Type":"ContainerStarted","Data":"c1aae3ef18a4011e77f7a6a7c3f8803d1d98f58a1d231ac7824be320f058f72a"} Dec 11 08:27:48 crc kubenswrapper[4860]: I1211 08:27:48.307302 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"c017f36f-48a3-4919-ac22-6a85a5c73530","Type":"ContainerStarted","Data":"215d2127477340862b25c9a3a9863b1ea829d526a78f5a7371758d12bed6efa5"} Dec 11 08:27:55 crc kubenswrapper[4860]: I1211 08:27:55.372535 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-dbdtz" event={"ID":"e834572f-9650-45d3-b978-c39141b44b74","Type":"ContainerStarted","Data":"5556f9d5a90602233fcdb109403ab0283154ef9c5969875b023ea58a02c84c54"} Dec 11 08:27:55 crc kubenswrapper[4860]: I1211 08:27:55.373418 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-dbdtz" Dec 11 08:27:55 crc kubenswrapper[4860]: I1211 08:27:55.378431 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"1752934c-842c-45b6-88ec-0fc804374532","Type":"ContainerStarted","Data":"ba9edb3622c3c148fe2c591efc0402b3363bf7f45c9f2c7695bf41b6e2dfc7d9"} Dec 11 08:27:55 crc kubenswrapper[4860]: I1211 08:27:55.378593 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Dec 11 08:27:55 crc kubenswrapper[4860]: I1211 08:27:55.382019 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"a0838f24-14b5-4018-ad26-8cd0394b4488","Type":"ContainerStarted","Data":"013c0e97ccb30e79e4437ef9a250965c46c6dbedf99a16f644fb289f2672cec2"} Dec 11 08:27:55 crc kubenswrapper[4860]: I1211 08:27:55.385214 4860 generic.go:334] "Generic (PLEG): container finished" podID="e73beb7d-c40c-491b-99f8-fabe037df9ba" containerID="7ff35ef9b1854e0e87565c5421c9bea31a315683ff2751289ea46a4a34cb40f0" exitCode=0 Dec 11 08:27:55 crc kubenswrapper[4860]: I1211 08:27:55.385298 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-q86ch" event={"ID":"e73beb7d-c40c-491b-99f8-fabe037df9ba","Type":"ContainerDied","Data":"7ff35ef9b1854e0e87565c5421c9bea31a315683ff2751289ea46a4a34cb40f0"} Dec 11 08:27:55 crc kubenswrapper[4860]: I1211 08:27:55.387163 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"9657082d-ff9c-4a54-8b4d-54e6ea411590","Type":"ContainerStarted","Data":"fe0d1a7f9c15b23fce360c6d4c8597a46f1e51deb6226b4a87e2bbf690323329"} Dec 11 08:27:55 crc kubenswrapper[4860]: I1211 08:27:55.388997 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"c017f36f-48a3-4919-ac22-6a85a5c73530","Type":"ContainerStarted","Data":"1286451efb3fa75aed3246f4a6a53c88b42d6b98c397fe1dc4e55a9c3faac0a0"} Dec 11 08:27:55 crc kubenswrapper[4860]: I1211 08:27:55.391790 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"c0d5e719-9d28-4e31-9349-f882c4b72567","Type":"ContainerStarted","Data":"9154ab6eed6c785b46d64ee5d4e7e0362b17862a4607fcb3227b3b9e23010359"} Dec 11 08:27:55 crc kubenswrapper[4860]: I1211 08:27:55.398580 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"bbb1e117-b966-48e4-bc09-08d66a1ce9eb","Type":"ContainerStarted","Data":"23e919a40caf8668edf5943b9e36eecf6963a5c9dddb7d9d93a14c7cffe56e62"} Dec 11 08:27:55 crc kubenswrapper[4860]: I1211 08:27:55.398766 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 11 08:27:55 crc kubenswrapper[4860]: I1211 08:27:55.401351 4860 generic.go:334] "Generic (PLEG): container finished" podID="7e94e216-32ea-44d2-b7ae-f765ebbdd524" containerID="cfb46f31c65be76e20aa6ca0bf2dbdefae62704d4a1000eebc9eab3c25bd9e45" exitCode=0 Dec 11 08:27:55 crc kubenswrapper[4860]: I1211 08:27:55.401420 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-vwqtx" event={"ID":"7e94e216-32ea-44d2-b7ae-f765ebbdd524","Type":"ContainerDied","Data":"cfb46f31c65be76e20aa6ca0bf2dbdefae62704d4a1000eebc9eab3c25bd9e45"} Dec 11 08:27:55 crc kubenswrapper[4860]: I1211 08:27:55.410852 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-dbdtz" podStartSLOduration=14.057865437 podStartE2EDuration="21.410821339s" podCreationTimestamp="2025-12-11 08:27:34 +0000 UTC" firstStartedPulling="2025-12-11 08:27:45.635809027 +0000 UTC m=+998.364328082" lastFinishedPulling="2025-12-11 08:27:52.988764909 +0000 UTC m=+1005.717283984" observedRunningTime="2025-12-11 08:27:55.394032662 +0000 UTC m=+1008.122551737" watchObservedRunningTime="2025-12-11 08:27:55.410821339 +0000 UTC m=+1008.139340404" Dec 11 08:27:55 crc kubenswrapper[4860]: I1211 08:27:55.418778 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=20.349432073 podStartE2EDuration="27.418759438s" podCreationTimestamp="2025-12-11 08:27:28 +0000 UTC" firstStartedPulling="2025-12-11 08:27:45.151745834 +0000 UTC m=+997.880264889" lastFinishedPulling="2025-12-11 08:27:52.221073199 +0000 UTC m=+1004.949592254" observedRunningTime="2025-12-11 08:27:55.415603679 +0000 UTC m=+1008.144122734" watchObservedRunningTime="2025-12-11 08:27:55.418759438 +0000 UTC m=+1008.147278503" Dec 11 08:27:55 crc kubenswrapper[4860]: I1211 08:27:55.523395 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=17.451816028 podStartE2EDuration="25.52337236s" podCreationTimestamp="2025-12-11 08:27:30 +0000 UTC" firstStartedPulling="2025-12-11 08:27:45.788375172 +0000 UTC m=+998.516894227" lastFinishedPulling="2025-12-11 08:27:53.859931504 +0000 UTC m=+1006.588450559" observedRunningTime="2025-12-11 08:27:55.517423363 +0000 UTC m=+1008.245942418" watchObservedRunningTime="2025-12-11 08:27:55.52337236 +0000 UTC m=+1008.251891415" Dec 11 08:27:56 crc kubenswrapper[4860]: I1211 08:27:56.413083 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-vwqtx" event={"ID":"7e94e216-32ea-44d2-b7ae-f765ebbdd524","Type":"ContainerStarted","Data":"86b7165f60940daeb1d7bbc82363223b064fa6f14dcbead9c91d2aa5b859ca34"} Dec 11 08:27:56 crc kubenswrapper[4860]: I1211 08:27:56.413607 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57d769cc4f-vwqtx" Dec 11 08:27:56 crc kubenswrapper[4860]: I1211 08:27:56.416857 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-q86ch" event={"ID":"e73beb7d-c40c-491b-99f8-fabe037df9ba","Type":"ContainerStarted","Data":"ceae2d24a90932ae9e2ad0ccbee18093e9c9e267b7512a4910f5c6ce11d524f0"} Dec 11 08:27:56 crc kubenswrapper[4860]: I1211 08:27:56.416900 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-q86ch" event={"ID":"e73beb7d-c40c-491b-99f8-fabe037df9ba","Type":"ContainerStarted","Data":"3d43e5edc96a50de7a2e6c7b318c02bb9593efd107ff076e717c344a0f0c9d98"} Dec 11 08:27:56 crc kubenswrapper[4860]: I1211 08:27:56.417346 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-q86ch" Dec 11 08:27:56 crc kubenswrapper[4860]: I1211 08:27:56.417399 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-q86ch" Dec 11 08:27:56 crc kubenswrapper[4860]: I1211 08:27:56.436808 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57d769cc4f-vwqtx" podStartSLOduration=3.447929038 podStartE2EDuration="32.43678936s" podCreationTimestamp="2025-12-11 08:27:24 +0000 UTC" firstStartedPulling="2025-12-11 08:27:25.69195225 +0000 UTC m=+978.420471305" lastFinishedPulling="2025-12-11 08:27:54.680812572 +0000 UTC m=+1007.409331627" observedRunningTime="2025-12-11 08:27:56.430089869 +0000 UTC m=+1009.158608924" watchObservedRunningTime="2025-12-11 08:27:56.43678936 +0000 UTC m=+1009.165308415" Dec 11 08:27:56 crc kubenswrapper[4860]: I1211 08:27:56.451434 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-q86ch" podStartSLOduration=15.613561351 podStartE2EDuration="22.451416888s" podCreationTimestamp="2025-12-11 08:27:34 +0000 UTC" firstStartedPulling="2025-12-11 08:27:45.793850593 +0000 UTC m=+998.522369648" lastFinishedPulling="2025-12-11 08:27:52.63170613 +0000 UTC m=+1005.360225185" observedRunningTime="2025-12-11 08:27:56.446938788 +0000 UTC m=+1009.175457843" watchObservedRunningTime="2025-12-11 08:27:56.451416888 +0000 UTC m=+1009.179935943" Dec 11 08:27:57 crc kubenswrapper[4860]: I1211 08:27:57.352301 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-lgzmd"] Dec 11 08:27:57 crc kubenswrapper[4860]: I1211 08:27:57.364629 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-lgzmd" Dec 11 08:27:57 crc kubenswrapper[4860]: I1211 08:27:57.366906 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Dec 11 08:27:57 crc kubenswrapper[4860]: I1211 08:27:57.368892 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-lgzmd"] Dec 11 08:27:57 crc kubenswrapper[4860]: I1211 08:27:57.480013 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/06395cef-744f-402e-844e-275a5025b81f-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-lgzmd\" (UID: \"06395cef-744f-402e-844e-275a5025b81f\") " pod="openstack/ovn-controller-metrics-lgzmd" Dec 11 08:27:57 crc kubenswrapper[4860]: I1211 08:27:57.480090 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/06395cef-744f-402e-844e-275a5025b81f-ovn-rundir\") pod \"ovn-controller-metrics-lgzmd\" (UID: \"06395cef-744f-402e-844e-275a5025b81f\") " pod="openstack/ovn-controller-metrics-lgzmd" Dec 11 08:27:57 crc kubenswrapper[4860]: I1211 08:27:57.480138 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cp2cm\" (UniqueName: \"kubernetes.io/projected/06395cef-744f-402e-844e-275a5025b81f-kube-api-access-cp2cm\") pod \"ovn-controller-metrics-lgzmd\" (UID: \"06395cef-744f-402e-844e-275a5025b81f\") " pod="openstack/ovn-controller-metrics-lgzmd" Dec 11 08:27:57 crc kubenswrapper[4860]: I1211 08:27:57.480184 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06395cef-744f-402e-844e-275a5025b81f-combined-ca-bundle\") pod \"ovn-controller-metrics-lgzmd\" (UID: \"06395cef-744f-402e-844e-275a5025b81f\") " pod="openstack/ovn-controller-metrics-lgzmd" Dec 11 08:27:57 crc kubenswrapper[4860]: I1211 08:27:57.480293 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06395cef-744f-402e-844e-275a5025b81f-config\") pod \"ovn-controller-metrics-lgzmd\" (UID: \"06395cef-744f-402e-844e-275a5025b81f\") " pod="openstack/ovn-controller-metrics-lgzmd" Dec 11 08:27:57 crc kubenswrapper[4860]: I1211 08:27:57.480314 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/06395cef-744f-402e-844e-275a5025b81f-ovs-rundir\") pod \"ovn-controller-metrics-lgzmd\" (UID: \"06395cef-744f-402e-844e-275a5025b81f\") " pod="openstack/ovn-controller-metrics-lgzmd" Dec 11 08:27:57 crc kubenswrapper[4860]: I1211 08:27:57.529201 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-vwqtx"] Dec 11 08:27:57 crc kubenswrapper[4860]: I1211 08:27:57.570685 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6bc7876d45-fzprz"] Dec 11 08:27:57 crc kubenswrapper[4860]: I1211 08:27:57.572122 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bc7876d45-fzprz" Dec 11 08:27:57 crc kubenswrapper[4860]: I1211 08:27:57.576228 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Dec 11 08:27:57 crc kubenswrapper[4860]: I1211 08:27:57.581382 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06395cef-744f-402e-844e-275a5025b81f-combined-ca-bundle\") pod \"ovn-controller-metrics-lgzmd\" (UID: \"06395cef-744f-402e-844e-275a5025b81f\") " pod="openstack/ovn-controller-metrics-lgzmd" Dec 11 08:27:57 crc kubenswrapper[4860]: I1211 08:27:57.581540 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/06395cef-744f-402e-844e-275a5025b81f-ovs-rundir\") pod \"ovn-controller-metrics-lgzmd\" (UID: \"06395cef-744f-402e-844e-275a5025b81f\") " pod="openstack/ovn-controller-metrics-lgzmd" Dec 11 08:27:57 crc kubenswrapper[4860]: I1211 08:27:57.581574 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06395cef-744f-402e-844e-275a5025b81f-config\") pod \"ovn-controller-metrics-lgzmd\" (UID: \"06395cef-744f-402e-844e-275a5025b81f\") " pod="openstack/ovn-controller-metrics-lgzmd" Dec 11 08:27:57 crc kubenswrapper[4860]: I1211 08:27:57.581626 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/06395cef-744f-402e-844e-275a5025b81f-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-lgzmd\" (UID: \"06395cef-744f-402e-844e-275a5025b81f\") " pod="openstack/ovn-controller-metrics-lgzmd" Dec 11 08:27:57 crc kubenswrapper[4860]: I1211 08:27:57.581690 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/06395cef-744f-402e-844e-275a5025b81f-ovn-rundir\") pod \"ovn-controller-metrics-lgzmd\" (UID: \"06395cef-744f-402e-844e-275a5025b81f\") " pod="openstack/ovn-controller-metrics-lgzmd" Dec 11 08:27:57 crc kubenswrapper[4860]: I1211 08:27:57.581732 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cp2cm\" (UniqueName: \"kubernetes.io/projected/06395cef-744f-402e-844e-275a5025b81f-kube-api-access-cp2cm\") pod \"ovn-controller-metrics-lgzmd\" (UID: \"06395cef-744f-402e-844e-275a5025b81f\") " pod="openstack/ovn-controller-metrics-lgzmd" Dec 11 08:27:57 crc kubenswrapper[4860]: I1211 08:27:57.583804 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06395cef-744f-402e-844e-275a5025b81f-config\") pod \"ovn-controller-metrics-lgzmd\" (UID: \"06395cef-744f-402e-844e-275a5025b81f\") " pod="openstack/ovn-controller-metrics-lgzmd" Dec 11 08:27:57 crc kubenswrapper[4860]: I1211 08:27:57.584100 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/06395cef-744f-402e-844e-275a5025b81f-ovs-rundir\") pod \"ovn-controller-metrics-lgzmd\" (UID: \"06395cef-744f-402e-844e-275a5025b81f\") " pod="openstack/ovn-controller-metrics-lgzmd" Dec 11 08:27:57 crc kubenswrapper[4860]: I1211 08:27:57.584209 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/06395cef-744f-402e-844e-275a5025b81f-ovn-rundir\") pod \"ovn-controller-metrics-lgzmd\" (UID: \"06395cef-744f-402e-844e-275a5025b81f\") " pod="openstack/ovn-controller-metrics-lgzmd" Dec 11 08:27:57 crc kubenswrapper[4860]: I1211 08:27:57.596130 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06395cef-744f-402e-844e-275a5025b81f-combined-ca-bundle\") pod \"ovn-controller-metrics-lgzmd\" (UID: \"06395cef-744f-402e-844e-275a5025b81f\") " pod="openstack/ovn-controller-metrics-lgzmd" Dec 11 08:27:57 crc kubenswrapper[4860]: I1211 08:27:57.596305 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/06395cef-744f-402e-844e-275a5025b81f-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-lgzmd\" (UID: \"06395cef-744f-402e-844e-275a5025b81f\") " pod="openstack/ovn-controller-metrics-lgzmd" Dec 11 08:27:57 crc kubenswrapper[4860]: I1211 08:27:57.599802 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6bc7876d45-fzprz"] Dec 11 08:27:57 crc kubenswrapper[4860]: I1211 08:27:57.624151 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cp2cm\" (UniqueName: \"kubernetes.io/projected/06395cef-744f-402e-844e-275a5025b81f-kube-api-access-cp2cm\") pod \"ovn-controller-metrics-lgzmd\" (UID: \"06395cef-744f-402e-844e-275a5025b81f\") " pod="openstack/ovn-controller-metrics-lgzmd" Dec 11 08:27:57 crc kubenswrapper[4860]: I1211 08:27:57.684037 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2319ad62-b109-4f60-880f-98c8da946fa6-config\") pod \"dnsmasq-dns-6bc7876d45-fzprz\" (UID: \"2319ad62-b109-4f60-880f-98c8da946fa6\") " pod="openstack/dnsmasq-dns-6bc7876d45-fzprz" Dec 11 08:27:57 crc kubenswrapper[4860]: I1211 08:27:57.684145 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lpg5d\" (UniqueName: \"kubernetes.io/projected/2319ad62-b109-4f60-880f-98c8da946fa6-kube-api-access-lpg5d\") pod \"dnsmasq-dns-6bc7876d45-fzprz\" (UID: \"2319ad62-b109-4f60-880f-98c8da946fa6\") " pod="openstack/dnsmasq-dns-6bc7876d45-fzprz" Dec 11 08:27:57 crc kubenswrapper[4860]: I1211 08:27:57.684168 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2319ad62-b109-4f60-880f-98c8da946fa6-dns-svc\") pod \"dnsmasq-dns-6bc7876d45-fzprz\" (UID: \"2319ad62-b109-4f60-880f-98c8da946fa6\") " pod="openstack/dnsmasq-dns-6bc7876d45-fzprz" Dec 11 08:27:57 crc kubenswrapper[4860]: I1211 08:27:57.684237 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2319ad62-b109-4f60-880f-98c8da946fa6-ovsdbserver-sb\") pod \"dnsmasq-dns-6bc7876d45-fzprz\" (UID: \"2319ad62-b109-4f60-880f-98c8da946fa6\") " pod="openstack/dnsmasq-dns-6bc7876d45-fzprz" Dec 11 08:27:57 crc kubenswrapper[4860]: I1211 08:27:57.692937 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-lgzmd" Dec 11 08:27:57 crc kubenswrapper[4860]: I1211 08:27:57.743119 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-vsf77"] Dec 11 08:27:57 crc kubenswrapper[4860]: I1211 08:27:57.758076 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8554648995-p5r62"] Dec 11 08:27:57 crc kubenswrapper[4860]: I1211 08:27:57.760054 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-p5r62" Dec 11 08:27:57 crc kubenswrapper[4860]: I1211 08:27:57.763712 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Dec 11 08:27:57 crc kubenswrapper[4860]: I1211 08:27:57.769413 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8554648995-p5r62"] Dec 11 08:27:57 crc kubenswrapper[4860]: I1211 08:27:57.790476 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2319ad62-b109-4f60-880f-98c8da946fa6-config\") pod \"dnsmasq-dns-6bc7876d45-fzprz\" (UID: \"2319ad62-b109-4f60-880f-98c8da946fa6\") " pod="openstack/dnsmasq-dns-6bc7876d45-fzprz" Dec 11 08:27:57 crc kubenswrapper[4860]: I1211 08:27:57.791696 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c4edc2e0-ea25-46f4-a127-b25e87244b28-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-p5r62\" (UID: \"c4edc2e0-ea25-46f4-a127-b25e87244b28\") " pod="openstack/dnsmasq-dns-8554648995-p5r62" Dec 11 08:27:57 crc kubenswrapper[4860]: I1211 08:27:57.791827 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lpg5d\" (UniqueName: \"kubernetes.io/projected/2319ad62-b109-4f60-880f-98c8da946fa6-kube-api-access-lpg5d\") pod \"dnsmasq-dns-6bc7876d45-fzprz\" (UID: \"2319ad62-b109-4f60-880f-98c8da946fa6\") " pod="openstack/dnsmasq-dns-6bc7876d45-fzprz" Dec 11 08:27:57 crc kubenswrapper[4860]: I1211 08:27:57.791889 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2319ad62-b109-4f60-880f-98c8da946fa6-dns-svc\") pod \"dnsmasq-dns-6bc7876d45-fzprz\" (UID: \"2319ad62-b109-4f60-880f-98c8da946fa6\") " pod="openstack/dnsmasq-dns-6bc7876d45-fzprz" Dec 11 08:27:57 crc kubenswrapper[4860]: I1211 08:27:57.791615 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2319ad62-b109-4f60-880f-98c8da946fa6-config\") pod \"dnsmasq-dns-6bc7876d45-fzprz\" (UID: \"2319ad62-b109-4f60-880f-98c8da946fa6\") " pod="openstack/dnsmasq-dns-6bc7876d45-fzprz" Dec 11 08:27:57 crc kubenswrapper[4860]: I1211 08:27:57.791956 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c4edc2e0-ea25-46f4-a127-b25e87244b28-dns-svc\") pod \"dnsmasq-dns-8554648995-p5r62\" (UID: \"c4edc2e0-ea25-46f4-a127-b25e87244b28\") " pod="openstack/dnsmasq-dns-8554648995-p5r62" Dec 11 08:27:57 crc kubenswrapper[4860]: I1211 08:27:57.792086 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c4edc2e0-ea25-46f4-a127-b25e87244b28-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-p5r62\" (UID: \"c4edc2e0-ea25-46f4-a127-b25e87244b28\") " pod="openstack/dnsmasq-dns-8554648995-p5r62" Dec 11 08:27:57 crc kubenswrapper[4860]: I1211 08:27:57.792771 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2319ad62-b109-4f60-880f-98c8da946fa6-dns-svc\") pod \"dnsmasq-dns-6bc7876d45-fzprz\" (UID: \"2319ad62-b109-4f60-880f-98c8da946fa6\") " pod="openstack/dnsmasq-dns-6bc7876d45-fzprz" Dec 11 08:27:57 crc kubenswrapper[4860]: I1211 08:27:57.792929 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jxdhf\" (UniqueName: \"kubernetes.io/projected/c4edc2e0-ea25-46f4-a127-b25e87244b28-kube-api-access-jxdhf\") pod \"dnsmasq-dns-8554648995-p5r62\" (UID: \"c4edc2e0-ea25-46f4-a127-b25e87244b28\") " pod="openstack/dnsmasq-dns-8554648995-p5r62" Dec 11 08:27:57 crc kubenswrapper[4860]: I1211 08:27:57.792982 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4edc2e0-ea25-46f4-a127-b25e87244b28-config\") pod \"dnsmasq-dns-8554648995-p5r62\" (UID: \"c4edc2e0-ea25-46f4-a127-b25e87244b28\") " pod="openstack/dnsmasq-dns-8554648995-p5r62" Dec 11 08:27:57 crc kubenswrapper[4860]: I1211 08:27:57.793027 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2319ad62-b109-4f60-880f-98c8da946fa6-ovsdbserver-sb\") pod \"dnsmasq-dns-6bc7876d45-fzprz\" (UID: \"2319ad62-b109-4f60-880f-98c8da946fa6\") " pod="openstack/dnsmasq-dns-6bc7876d45-fzprz" Dec 11 08:27:57 crc kubenswrapper[4860]: I1211 08:27:57.793893 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2319ad62-b109-4f60-880f-98c8da946fa6-ovsdbserver-sb\") pod \"dnsmasq-dns-6bc7876d45-fzprz\" (UID: \"2319ad62-b109-4f60-880f-98c8da946fa6\") " pod="openstack/dnsmasq-dns-6bc7876d45-fzprz" Dec 11 08:27:57 crc kubenswrapper[4860]: I1211 08:27:57.823738 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lpg5d\" (UniqueName: \"kubernetes.io/projected/2319ad62-b109-4f60-880f-98c8da946fa6-kube-api-access-lpg5d\") pod \"dnsmasq-dns-6bc7876d45-fzprz\" (UID: \"2319ad62-b109-4f60-880f-98c8da946fa6\") " pod="openstack/dnsmasq-dns-6bc7876d45-fzprz" Dec 11 08:27:57 crc kubenswrapper[4860]: I1211 08:27:57.895124 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c4edc2e0-ea25-46f4-a127-b25e87244b28-dns-svc\") pod \"dnsmasq-dns-8554648995-p5r62\" (UID: \"c4edc2e0-ea25-46f4-a127-b25e87244b28\") " pod="openstack/dnsmasq-dns-8554648995-p5r62" Dec 11 08:27:57 crc kubenswrapper[4860]: I1211 08:27:57.895380 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c4edc2e0-ea25-46f4-a127-b25e87244b28-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-p5r62\" (UID: \"c4edc2e0-ea25-46f4-a127-b25e87244b28\") " pod="openstack/dnsmasq-dns-8554648995-p5r62" Dec 11 08:27:57 crc kubenswrapper[4860]: I1211 08:27:57.895415 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jxdhf\" (UniqueName: \"kubernetes.io/projected/c4edc2e0-ea25-46f4-a127-b25e87244b28-kube-api-access-jxdhf\") pod \"dnsmasq-dns-8554648995-p5r62\" (UID: \"c4edc2e0-ea25-46f4-a127-b25e87244b28\") " pod="openstack/dnsmasq-dns-8554648995-p5r62" Dec 11 08:27:57 crc kubenswrapper[4860]: I1211 08:27:57.895439 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4edc2e0-ea25-46f4-a127-b25e87244b28-config\") pod \"dnsmasq-dns-8554648995-p5r62\" (UID: \"c4edc2e0-ea25-46f4-a127-b25e87244b28\") " pod="openstack/dnsmasq-dns-8554648995-p5r62" Dec 11 08:27:57 crc kubenswrapper[4860]: I1211 08:27:57.895506 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c4edc2e0-ea25-46f4-a127-b25e87244b28-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-p5r62\" (UID: \"c4edc2e0-ea25-46f4-a127-b25e87244b28\") " pod="openstack/dnsmasq-dns-8554648995-p5r62" Dec 11 08:27:57 crc kubenswrapper[4860]: I1211 08:27:57.896113 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c4edc2e0-ea25-46f4-a127-b25e87244b28-dns-svc\") pod \"dnsmasq-dns-8554648995-p5r62\" (UID: \"c4edc2e0-ea25-46f4-a127-b25e87244b28\") " pod="openstack/dnsmasq-dns-8554648995-p5r62" Dec 11 08:27:57 crc kubenswrapper[4860]: I1211 08:27:57.896200 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4edc2e0-ea25-46f4-a127-b25e87244b28-config\") pod \"dnsmasq-dns-8554648995-p5r62\" (UID: \"c4edc2e0-ea25-46f4-a127-b25e87244b28\") " pod="openstack/dnsmasq-dns-8554648995-p5r62" Dec 11 08:27:57 crc kubenswrapper[4860]: I1211 08:27:57.896448 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c4edc2e0-ea25-46f4-a127-b25e87244b28-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-p5r62\" (UID: \"c4edc2e0-ea25-46f4-a127-b25e87244b28\") " pod="openstack/dnsmasq-dns-8554648995-p5r62" Dec 11 08:27:57 crc kubenswrapper[4860]: I1211 08:27:57.896855 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c4edc2e0-ea25-46f4-a127-b25e87244b28-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-p5r62\" (UID: \"c4edc2e0-ea25-46f4-a127-b25e87244b28\") " pod="openstack/dnsmasq-dns-8554648995-p5r62" Dec 11 08:27:57 crc kubenswrapper[4860]: I1211 08:27:57.914520 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jxdhf\" (UniqueName: \"kubernetes.io/projected/c4edc2e0-ea25-46f4-a127-b25e87244b28-kube-api-access-jxdhf\") pod \"dnsmasq-dns-8554648995-p5r62\" (UID: \"c4edc2e0-ea25-46f4-a127-b25e87244b28\") " pod="openstack/dnsmasq-dns-8554648995-p5r62" Dec 11 08:27:57 crc kubenswrapper[4860]: I1211 08:27:57.973033 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bc7876d45-fzprz" Dec 11 08:27:58 crc kubenswrapper[4860]: I1211 08:27:58.096285 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-p5r62" Dec 11 08:27:58 crc kubenswrapper[4860]: I1211 08:27:58.186573 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-vsf77" Dec 11 08:27:58 crc kubenswrapper[4860]: I1211 08:27:58.299999 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5ea9b400-63b3-480e-880e-978b406fce50-config\") pod \"5ea9b400-63b3-480e-880e-978b406fce50\" (UID: \"5ea9b400-63b3-480e-880e-978b406fce50\") " Dec 11 08:27:58 crc kubenswrapper[4860]: I1211 08:27:58.300122 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rgh8v\" (UniqueName: \"kubernetes.io/projected/5ea9b400-63b3-480e-880e-978b406fce50-kube-api-access-rgh8v\") pod \"5ea9b400-63b3-480e-880e-978b406fce50\" (UID: \"5ea9b400-63b3-480e-880e-978b406fce50\") " Dec 11 08:27:58 crc kubenswrapper[4860]: I1211 08:27:58.300187 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5ea9b400-63b3-480e-880e-978b406fce50-dns-svc\") pod \"5ea9b400-63b3-480e-880e-978b406fce50\" (UID: \"5ea9b400-63b3-480e-880e-978b406fce50\") " Dec 11 08:27:58 crc kubenswrapper[4860]: I1211 08:27:58.300611 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5ea9b400-63b3-480e-880e-978b406fce50-config" (OuterVolumeSpecName: "config") pod "5ea9b400-63b3-480e-880e-978b406fce50" (UID: "5ea9b400-63b3-480e-880e-978b406fce50"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:27:58 crc kubenswrapper[4860]: I1211 08:27:58.301001 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5ea9b400-63b3-480e-880e-978b406fce50-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5ea9b400-63b3-480e-880e-978b406fce50" (UID: "5ea9b400-63b3-480e-880e-978b406fce50"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:27:58 crc kubenswrapper[4860]: I1211 08:27:58.304637 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ea9b400-63b3-480e-880e-978b406fce50-kube-api-access-rgh8v" (OuterVolumeSpecName: "kube-api-access-rgh8v") pod "5ea9b400-63b3-480e-880e-978b406fce50" (UID: "5ea9b400-63b3-480e-880e-978b406fce50"). InnerVolumeSpecName "kube-api-access-rgh8v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:27:58 crc kubenswrapper[4860]: I1211 08:27:58.402675 4860 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5ea9b400-63b3-480e-880e-978b406fce50-config\") on node \"crc\" DevicePath \"\"" Dec 11 08:27:58 crc kubenswrapper[4860]: I1211 08:27:58.402757 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rgh8v\" (UniqueName: \"kubernetes.io/projected/5ea9b400-63b3-480e-880e-978b406fce50-kube-api-access-rgh8v\") on node \"crc\" DevicePath \"\"" Dec 11 08:27:58 crc kubenswrapper[4860]: I1211 08:27:58.402775 4860 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5ea9b400-63b3-480e-880e-978b406fce50-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 11 08:27:58 crc kubenswrapper[4860]: I1211 08:27:58.443710 4860 generic.go:334] "Generic (PLEG): container finished" podID="c0d5e719-9d28-4e31-9349-f882c4b72567" containerID="9154ab6eed6c785b46d64ee5d4e7e0362b17862a4607fcb3227b3b9e23010359" exitCode=0 Dec 11 08:27:58 crc kubenswrapper[4860]: I1211 08:27:58.443817 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"c0d5e719-9d28-4e31-9349-f882c4b72567","Type":"ContainerDied","Data":"9154ab6eed6c785b46d64ee5d4e7e0362b17862a4607fcb3227b3b9e23010359"} Dec 11 08:27:58 crc kubenswrapper[4860]: I1211 08:27:58.446942 4860 generic.go:334] "Generic (PLEG): container finished" podID="a0838f24-14b5-4018-ad26-8cd0394b4488" containerID="013c0e97ccb30e79e4437ef9a250965c46c6dbedf99a16f644fb289f2672cec2" exitCode=0 Dec 11 08:27:58 crc kubenswrapper[4860]: I1211 08:27:58.447062 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"a0838f24-14b5-4018-ad26-8cd0394b4488","Type":"ContainerDied","Data":"013c0e97ccb30e79e4437ef9a250965c46c6dbedf99a16f644fb289f2672cec2"} Dec 11 08:27:58 crc kubenswrapper[4860]: I1211 08:27:58.448721 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-vsf77" event={"ID":"5ea9b400-63b3-480e-880e-978b406fce50","Type":"ContainerDied","Data":"4d3997c82d9e0064cf79f1a5f121041549d0feddf6308665af53a842c9d5a66c"} Dec 11 08:27:58 crc kubenswrapper[4860]: I1211 08:27:58.448869 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57d769cc4f-vwqtx" podUID="7e94e216-32ea-44d2-b7ae-f765ebbdd524" containerName="dnsmasq-dns" containerID="cri-o://86b7165f60940daeb1d7bbc82363223b064fa6f14dcbead9c91d2aa5b859ca34" gracePeriod=10 Dec 11 08:27:58 crc kubenswrapper[4860]: I1211 08:27:58.448959 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-vsf77" Dec 11 08:27:58 crc kubenswrapper[4860]: I1211 08:27:58.536692 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-vsf77"] Dec 11 08:27:58 crc kubenswrapper[4860]: I1211 08:27:58.542998 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-vsf77"] Dec 11 08:27:59 crc kubenswrapper[4860]: I1211 08:27:59.072678 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-vwqtx" Dec 11 08:27:59 crc kubenswrapper[4860]: I1211 08:27:59.122903 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e94e216-32ea-44d2-b7ae-f765ebbdd524-config\") pod \"7e94e216-32ea-44d2-b7ae-f765ebbdd524\" (UID: \"7e94e216-32ea-44d2-b7ae-f765ebbdd524\") " Dec 11 08:27:59 crc kubenswrapper[4860]: I1211 08:27:59.123783 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-89cmp\" (UniqueName: \"kubernetes.io/projected/7e94e216-32ea-44d2-b7ae-f765ebbdd524-kube-api-access-89cmp\") pod \"7e94e216-32ea-44d2-b7ae-f765ebbdd524\" (UID: \"7e94e216-32ea-44d2-b7ae-f765ebbdd524\") " Dec 11 08:27:59 crc kubenswrapper[4860]: I1211 08:27:59.123874 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7e94e216-32ea-44d2-b7ae-f765ebbdd524-dns-svc\") pod \"7e94e216-32ea-44d2-b7ae-f765ebbdd524\" (UID: \"7e94e216-32ea-44d2-b7ae-f765ebbdd524\") " Dec 11 08:27:59 crc kubenswrapper[4860]: I1211 08:27:59.129001 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e94e216-32ea-44d2-b7ae-f765ebbdd524-kube-api-access-89cmp" (OuterVolumeSpecName: "kube-api-access-89cmp") pod "7e94e216-32ea-44d2-b7ae-f765ebbdd524" (UID: "7e94e216-32ea-44d2-b7ae-f765ebbdd524"). InnerVolumeSpecName "kube-api-access-89cmp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:27:59 crc kubenswrapper[4860]: I1211 08:27:59.161953 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7e94e216-32ea-44d2-b7ae-f765ebbdd524-config" (OuterVolumeSpecName: "config") pod "7e94e216-32ea-44d2-b7ae-f765ebbdd524" (UID: "7e94e216-32ea-44d2-b7ae-f765ebbdd524"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:27:59 crc kubenswrapper[4860]: I1211 08:27:59.169052 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7e94e216-32ea-44d2-b7ae-f765ebbdd524-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7e94e216-32ea-44d2-b7ae-f765ebbdd524" (UID: "7e94e216-32ea-44d2-b7ae-f765ebbdd524"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:27:59 crc kubenswrapper[4860]: I1211 08:27:59.227386 4860 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e94e216-32ea-44d2-b7ae-f765ebbdd524-config\") on node \"crc\" DevicePath \"\"" Dec 11 08:27:59 crc kubenswrapper[4860]: I1211 08:27:59.227440 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-89cmp\" (UniqueName: \"kubernetes.io/projected/7e94e216-32ea-44d2-b7ae-f765ebbdd524-kube-api-access-89cmp\") on node \"crc\" DevicePath \"\"" Dec 11 08:27:59 crc kubenswrapper[4860]: I1211 08:27:59.227853 4860 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7e94e216-32ea-44d2-b7ae-f765ebbdd524-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 11 08:27:59 crc kubenswrapper[4860]: I1211 08:27:59.228315 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6bc7876d45-fzprz"] Dec 11 08:27:59 crc kubenswrapper[4860]: W1211 08:27:59.234188 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2319ad62_b109_4f60_880f_98c8da946fa6.slice/crio-305264eae8509d819b23ac35b0724986927586f0e7c048c9d2c2f709b8a266ac WatchSource:0}: Error finding container 305264eae8509d819b23ac35b0724986927586f0e7c048c9d2c2f709b8a266ac: Status 404 returned error can't find the container with id 305264eae8509d819b23ac35b0724986927586f0e7c048c9d2c2f709b8a266ac Dec 11 08:27:59 crc kubenswrapper[4860]: I1211 08:27:59.331261 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-lgzmd"] Dec 11 08:27:59 crc kubenswrapper[4860]: W1211 08:27:59.332986 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc4edc2e0_ea25_46f4_a127_b25e87244b28.slice/crio-59c9f2586b5d93fe3aca0b531a4082e900da0e764d39b54ff47b521939064be7 WatchSource:0}: Error finding container 59c9f2586b5d93fe3aca0b531a4082e900da0e764d39b54ff47b521939064be7: Status 404 returned error can't find the container with id 59c9f2586b5d93fe3aca0b531a4082e900da0e764d39b54ff47b521939064be7 Dec 11 08:27:59 crc kubenswrapper[4860]: W1211 08:27:59.334676 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod06395cef_744f_402e_844e_275a5025b81f.slice/crio-a7b806628dc48bb3d1c8470d63caacc4e0546b480eac9bec8bd14d9f3a40d04a WatchSource:0}: Error finding container a7b806628dc48bb3d1c8470d63caacc4e0546b480eac9bec8bd14d9f3a40d04a: Status 404 returned error can't find the container with id a7b806628dc48bb3d1c8470d63caacc4e0546b480eac9bec8bd14d9f3a40d04a Dec 11 08:27:59 crc kubenswrapper[4860]: I1211 08:27:59.341147 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8554648995-p5r62"] Dec 11 08:27:59 crc kubenswrapper[4860]: I1211 08:27:59.444123 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Dec 11 08:27:59 crc kubenswrapper[4860]: I1211 08:27:59.464400 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-p5r62" event={"ID":"c4edc2e0-ea25-46f4-a127-b25e87244b28","Type":"ContainerStarted","Data":"59c9f2586b5d93fe3aca0b531a4082e900da0e764d39b54ff47b521939064be7"} Dec 11 08:27:59 crc kubenswrapper[4860]: I1211 08:27:59.466923 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bc7876d45-fzprz" event={"ID":"2319ad62-b109-4f60-880f-98c8da946fa6","Type":"ContainerStarted","Data":"305264eae8509d819b23ac35b0724986927586f0e7c048c9d2c2f709b8a266ac"} Dec 11 08:27:59 crc kubenswrapper[4860]: I1211 08:27:59.468549 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-lgzmd" event={"ID":"06395cef-744f-402e-844e-275a5025b81f","Type":"ContainerStarted","Data":"a7b806628dc48bb3d1c8470d63caacc4e0546b480eac9bec8bd14d9f3a40d04a"} Dec 11 08:27:59 crc kubenswrapper[4860]: I1211 08:27:59.472854 4860 generic.go:334] "Generic (PLEG): container finished" podID="7e94e216-32ea-44d2-b7ae-f765ebbdd524" containerID="86b7165f60940daeb1d7bbc82363223b064fa6f14dcbead9c91d2aa5b859ca34" exitCode=0 Dec 11 08:27:59 crc kubenswrapper[4860]: I1211 08:27:59.472937 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-vwqtx" event={"ID":"7e94e216-32ea-44d2-b7ae-f765ebbdd524","Type":"ContainerDied","Data":"86b7165f60940daeb1d7bbc82363223b064fa6f14dcbead9c91d2aa5b859ca34"} Dec 11 08:27:59 crc kubenswrapper[4860]: I1211 08:27:59.472976 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-vwqtx" event={"ID":"7e94e216-32ea-44d2-b7ae-f765ebbdd524","Type":"ContainerDied","Data":"31913ca3198c2b25c60b6218937620ced66fe11ea1da0f74d81684db7eb8693a"} Dec 11 08:27:59 crc kubenswrapper[4860]: I1211 08:27:59.473029 4860 scope.go:117] "RemoveContainer" containerID="86b7165f60940daeb1d7bbc82363223b064fa6f14dcbead9c91d2aa5b859ca34" Dec 11 08:27:59 crc kubenswrapper[4860]: I1211 08:27:59.473250 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-vwqtx" Dec 11 08:27:59 crc kubenswrapper[4860]: I1211 08:27:59.506628 4860 scope.go:117] "RemoveContainer" containerID="cfb46f31c65be76e20aa6ca0bf2dbdefae62704d4a1000eebc9eab3c25bd9e45" Dec 11 08:27:59 crc kubenswrapper[4860]: I1211 08:27:59.528173 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-vwqtx"] Dec 11 08:27:59 crc kubenswrapper[4860]: I1211 08:27:59.539330 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-vwqtx"] Dec 11 08:27:59 crc kubenswrapper[4860]: I1211 08:27:59.561389 4860 scope.go:117] "RemoveContainer" containerID="86b7165f60940daeb1d7bbc82363223b064fa6f14dcbead9c91d2aa5b859ca34" Dec 11 08:27:59 crc kubenswrapper[4860]: E1211 08:27:59.562089 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"86b7165f60940daeb1d7bbc82363223b064fa6f14dcbead9c91d2aa5b859ca34\": container with ID starting with 86b7165f60940daeb1d7bbc82363223b064fa6f14dcbead9c91d2aa5b859ca34 not found: ID does not exist" containerID="86b7165f60940daeb1d7bbc82363223b064fa6f14dcbead9c91d2aa5b859ca34" Dec 11 08:27:59 crc kubenswrapper[4860]: I1211 08:27:59.562144 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"86b7165f60940daeb1d7bbc82363223b064fa6f14dcbead9c91d2aa5b859ca34"} err="failed to get container status \"86b7165f60940daeb1d7bbc82363223b064fa6f14dcbead9c91d2aa5b859ca34\": rpc error: code = NotFound desc = could not find container \"86b7165f60940daeb1d7bbc82363223b064fa6f14dcbead9c91d2aa5b859ca34\": container with ID starting with 86b7165f60940daeb1d7bbc82363223b064fa6f14dcbead9c91d2aa5b859ca34 not found: ID does not exist" Dec 11 08:27:59 crc kubenswrapper[4860]: I1211 08:27:59.562181 4860 scope.go:117] "RemoveContainer" containerID="cfb46f31c65be76e20aa6ca0bf2dbdefae62704d4a1000eebc9eab3c25bd9e45" Dec 11 08:27:59 crc kubenswrapper[4860]: E1211 08:27:59.562517 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cfb46f31c65be76e20aa6ca0bf2dbdefae62704d4a1000eebc9eab3c25bd9e45\": container with ID starting with cfb46f31c65be76e20aa6ca0bf2dbdefae62704d4a1000eebc9eab3c25bd9e45 not found: ID does not exist" containerID="cfb46f31c65be76e20aa6ca0bf2dbdefae62704d4a1000eebc9eab3c25bd9e45" Dec 11 08:27:59 crc kubenswrapper[4860]: I1211 08:27:59.562552 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cfb46f31c65be76e20aa6ca0bf2dbdefae62704d4a1000eebc9eab3c25bd9e45"} err="failed to get container status \"cfb46f31c65be76e20aa6ca0bf2dbdefae62704d4a1000eebc9eab3c25bd9e45\": rpc error: code = NotFound desc = could not find container \"cfb46f31c65be76e20aa6ca0bf2dbdefae62704d4a1000eebc9eab3c25bd9e45\": container with ID starting with cfb46f31c65be76e20aa6ca0bf2dbdefae62704d4a1000eebc9eab3c25bd9e45 not found: ID does not exist" Dec 11 08:27:59 crc kubenswrapper[4860]: I1211 08:27:59.590056 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ea9b400-63b3-480e-880e-978b406fce50" path="/var/lib/kubelet/pods/5ea9b400-63b3-480e-880e-978b406fce50/volumes" Dec 11 08:27:59 crc kubenswrapper[4860]: I1211 08:27:59.590548 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7e94e216-32ea-44d2-b7ae-f765ebbdd524" path="/var/lib/kubelet/pods/7e94e216-32ea-44d2-b7ae-f765ebbdd524/volumes" Dec 11 08:28:00 crc kubenswrapper[4860]: I1211 08:28:00.487578 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"a0838f24-14b5-4018-ad26-8cd0394b4488","Type":"ContainerStarted","Data":"1346d466896947ac4750a868d9d0603004efb6550b9fefae115a0c33b493b224"} Dec 11 08:28:01 crc kubenswrapper[4860]: I1211 08:28:01.076139 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 11 08:28:01 crc kubenswrapper[4860]: I1211 08:28:01.257860 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bc7876d45-fzprz"] Dec 11 08:28:01 crc kubenswrapper[4860]: I1211 08:28:01.328614 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-l7vfs"] Dec 11 08:28:01 crc kubenswrapper[4860]: E1211 08:28:01.328989 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e94e216-32ea-44d2-b7ae-f765ebbdd524" containerName="dnsmasq-dns" Dec 11 08:28:01 crc kubenswrapper[4860]: I1211 08:28:01.329008 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e94e216-32ea-44d2-b7ae-f765ebbdd524" containerName="dnsmasq-dns" Dec 11 08:28:01 crc kubenswrapper[4860]: E1211 08:28:01.329039 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e94e216-32ea-44d2-b7ae-f765ebbdd524" containerName="init" Dec 11 08:28:01 crc kubenswrapper[4860]: I1211 08:28:01.329047 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e94e216-32ea-44d2-b7ae-f765ebbdd524" containerName="init" Dec 11 08:28:01 crc kubenswrapper[4860]: I1211 08:28:01.329195 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e94e216-32ea-44d2-b7ae-f765ebbdd524" containerName="dnsmasq-dns" Dec 11 08:28:01 crc kubenswrapper[4860]: I1211 08:28:01.330119 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-l7vfs" Dec 11 08:28:01 crc kubenswrapper[4860]: I1211 08:28:01.356024 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-l7vfs"] Dec 11 08:28:01 crc kubenswrapper[4860]: I1211 08:28:01.375541 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/77f658e2-b468-45dd-8204-691cdfe9bf61-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-l7vfs\" (UID: \"77f658e2-b468-45dd-8204-691cdfe9bf61\") " pod="openstack/dnsmasq-dns-b8fbc5445-l7vfs" Dec 11 08:28:01 crc kubenswrapper[4860]: I1211 08:28:01.375596 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/77f658e2-b468-45dd-8204-691cdfe9bf61-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-l7vfs\" (UID: \"77f658e2-b468-45dd-8204-691cdfe9bf61\") " pod="openstack/dnsmasq-dns-b8fbc5445-l7vfs" Dec 11 08:28:01 crc kubenswrapper[4860]: I1211 08:28:01.375658 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/77f658e2-b468-45dd-8204-691cdfe9bf61-config\") pod \"dnsmasq-dns-b8fbc5445-l7vfs\" (UID: \"77f658e2-b468-45dd-8204-691cdfe9bf61\") " pod="openstack/dnsmasq-dns-b8fbc5445-l7vfs" Dec 11 08:28:01 crc kubenswrapper[4860]: I1211 08:28:01.375960 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dxtnk\" (UniqueName: \"kubernetes.io/projected/77f658e2-b468-45dd-8204-691cdfe9bf61-kube-api-access-dxtnk\") pod \"dnsmasq-dns-b8fbc5445-l7vfs\" (UID: \"77f658e2-b468-45dd-8204-691cdfe9bf61\") " pod="openstack/dnsmasq-dns-b8fbc5445-l7vfs" Dec 11 08:28:01 crc kubenswrapper[4860]: I1211 08:28:01.376047 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/77f658e2-b468-45dd-8204-691cdfe9bf61-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-l7vfs\" (UID: \"77f658e2-b468-45dd-8204-691cdfe9bf61\") " pod="openstack/dnsmasq-dns-b8fbc5445-l7vfs" Dec 11 08:28:01 crc kubenswrapper[4860]: I1211 08:28:01.477674 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dxtnk\" (UniqueName: \"kubernetes.io/projected/77f658e2-b468-45dd-8204-691cdfe9bf61-kube-api-access-dxtnk\") pod \"dnsmasq-dns-b8fbc5445-l7vfs\" (UID: \"77f658e2-b468-45dd-8204-691cdfe9bf61\") " pod="openstack/dnsmasq-dns-b8fbc5445-l7vfs" Dec 11 08:28:01 crc kubenswrapper[4860]: I1211 08:28:01.477753 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/77f658e2-b468-45dd-8204-691cdfe9bf61-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-l7vfs\" (UID: \"77f658e2-b468-45dd-8204-691cdfe9bf61\") " pod="openstack/dnsmasq-dns-b8fbc5445-l7vfs" Dec 11 08:28:01 crc kubenswrapper[4860]: I1211 08:28:01.478796 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/77f658e2-b468-45dd-8204-691cdfe9bf61-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-l7vfs\" (UID: \"77f658e2-b468-45dd-8204-691cdfe9bf61\") " pod="openstack/dnsmasq-dns-b8fbc5445-l7vfs" Dec 11 08:28:01 crc kubenswrapper[4860]: I1211 08:28:01.478875 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/77f658e2-b468-45dd-8204-691cdfe9bf61-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-l7vfs\" (UID: \"77f658e2-b468-45dd-8204-691cdfe9bf61\") " pod="openstack/dnsmasq-dns-b8fbc5445-l7vfs" Dec 11 08:28:01 crc kubenswrapper[4860]: I1211 08:28:01.479420 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/77f658e2-b468-45dd-8204-691cdfe9bf61-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-l7vfs\" (UID: \"77f658e2-b468-45dd-8204-691cdfe9bf61\") " pod="openstack/dnsmasq-dns-b8fbc5445-l7vfs" Dec 11 08:28:01 crc kubenswrapper[4860]: I1211 08:28:01.479493 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/77f658e2-b468-45dd-8204-691cdfe9bf61-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-l7vfs\" (UID: \"77f658e2-b468-45dd-8204-691cdfe9bf61\") " pod="openstack/dnsmasq-dns-b8fbc5445-l7vfs" Dec 11 08:28:01 crc kubenswrapper[4860]: I1211 08:28:01.480070 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/77f658e2-b468-45dd-8204-691cdfe9bf61-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-l7vfs\" (UID: \"77f658e2-b468-45dd-8204-691cdfe9bf61\") " pod="openstack/dnsmasq-dns-b8fbc5445-l7vfs" Dec 11 08:28:01 crc kubenswrapper[4860]: I1211 08:28:01.480137 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/77f658e2-b468-45dd-8204-691cdfe9bf61-config\") pod \"dnsmasq-dns-b8fbc5445-l7vfs\" (UID: \"77f658e2-b468-45dd-8204-691cdfe9bf61\") " pod="openstack/dnsmasq-dns-b8fbc5445-l7vfs" Dec 11 08:28:01 crc kubenswrapper[4860]: I1211 08:28:01.480695 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/77f658e2-b468-45dd-8204-691cdfe9bf61-config\") pod \"dnsmasq-dns-b8fbc5445-l7vfs\" (UID: \"77f658e2-b468-45dd-8204-691cdfe9bf61\") " pod="openstack/dnsmasq-dns-b8fbc5445-l7vfs" Dec 11 08:28:01 crc kubenswrapper[4860]: I1211 08:28:01.513633 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dxtnk\" (UniqueName: \"kubernetes.io/projected/77f658e2-b468-45dd-8204-691cdfe9bf61-kube-api-access-dxtnk\") pod \"dnsmasq-dns-b8fbc5445-l7vfs\" (UID: \"77f658e2-b468-45dd-8204-691cdfe9bf61\") " pod="openstack/dnsmasq-dns-b8fbc5445-l7vfs" Dec 11 08:28:01 crc kubenswrapper[4860]: I1211 08:28:01.648692 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-l7vfs" Dec 11 08:28:02 crc kubenswrapper[4860]: I1211 08:28:02.166267 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-l7vfs"] Dec 11 08:28:02 crc kubenswrapper[4860]: I1211 08:28:02.348387 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Dec 11 08:28:02 crc kubenswrapper[4860]: I1211 08:28:02.367532 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 11 08:28:02 crc kubenswrapper[4860]: I1211 08:28:02.370045 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-jnmdx" Dec 11 08:28:02 crc kubenswrapper[4860]: I1211 08:28:02.370568 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Dec 11 08:28:02 crc kubenswrapper[4860]: I1211 08:28:02.370869 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Dec 11 08:28:02 crc kubenswrapper[4860]: I1211 08:28:02.371097 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Dec 11 08:28:02 crc kubenswrapper[4860]: I1211 08:28:02.379912 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Dec 11 08:28:02 crc kubenswrapper[4860]: I1211 08:28:02.397354 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a672e32c-a924-4080-a7d1-bf7012b6725b-etc-swift\") pod \"swift-storage-0\" (UID: \"a672e32c-a924-4080-a7d1-bf7012b6725b\") " pod="openstack/swift-storage-0" Dec 11 08:28:02 crc kubenswrapper[4860]: I1211 08:28:02.397447 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"swift-storage-0\" (UID: \"a672e32c-a924-4080-a7d1-bf7012b6725b\") " pod="openstack/swift-storage-0" Dec 11 08:28:02 crc kubenswrapper[4860]: I1211 08:28:02.397838 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/a672e32c-a924-4080-a7d1-bf7012b6725b-cache\") pod \"swift-storage-0\" (UID: \"a672e32c-a924-4080-a7d1-bf7012b6725b\") " pod="openstack/swift-storage-0" Dec 11 08:28:02 crc kubenswrapper[4860]: I1211 08:28:02.397990 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/a672e32c-a924-4080-a7d1-bf7012b6725b-lock\") pod \"swift-storage-0\" (UID: \"a672e32c-a924-4080-a7d1-bf7012b6725b\") " pod="openstack/swift-storage-0" Dec 11 08:28:02 crc kubenswrapper[4860]: I1211 08:28:02.398043 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qsmps\" (UniqueName: \"kubernetes.io/projected/a672e32c-a924-4080-a7d1-bf7012b6725b-kube-api-access-qsmps\") pod \"swift-storage-0\" (UID: \"a672e32c-a924-4080-a7d1-bf7012b6725b\") " pod="openstack/swift-storage-0" Dec 11 08:28:02 crc kubenswrapper[4860]: I1211 08:28:02.500556 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"swift-storage-0\" (UID: \"a672e32c-a924-4080-a7d1-bf7012b6725b\") " pod="openstack/swift-storage-0" Dec 11 08:28:02 crc kubenswrapper[4860]: I1211 08:28:02.500684 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/a672e32c-a924-4080-a7d1-bf7012b6725b-cache\") pod \"swift-storage-0\" (UID: \"a672e32c-a924-4080-a7d1-bf7012b6725b\") " pod="openstack/swift-storage-0" Dec 11 08:28:02 crc kubenswrapper[4860]: I1211 08:28:02.500715 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/a672e32c-a924-4080-a7d1-bf7012b6725b-lock\") pod \"swift-storage-0\" (UID: \"a672e32c-a924-4080-a7d1-bf7012b6725b\") " pod="openstack/swift-storage-0" Dec 11 08:28:02 crc kubenswrapper[4860]: I1211 08:28:02.500737 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qsmps\" (UniqueName: \"kubernetes.io/projected/a672e32c-a924-4080-a7d1-bf7012b6725b-kube-api-access-qsmps\") pod \"swift-storage-0\" (UID: \"a672e32c-a924-4080-a7d1-bf7012b6725b\") " pod="openstack/swift-storage-0" Dec 11 08:28:02 crc kubenswrapper[4860]: I1211 08:28:02.500788 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a672e32c-a924-4080-a7d1-bf7012b6725b-etc-swift\") pod \"swift-storage-0\" (UID: \"a672e32c-a924-4080-a7d1-bf7012b6725b\") " pod="openstack/swift-storage-0" Dec 11 08:28:02 crc kubenswrapper[4860]: E1211 08:28:02.501019 4860 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 11 08:28:02 crc kubenswrapper[4860]: E1211 08:28:02.501042 4860 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 11 08:28:02 crc kubenswrapper[4860]: E1211 08:28:02.501124 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/a672e32c-a924-4080-a7d1-bf7012b6725b-etc-swift podName:a672e32c-a924-4080-a7d1-bf7012b6725b nodeName:}" failed. No retries permitted until 2025-12-11 08:28:03.001093172 +0000 UTC m=+1015.729612227 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/a672e32c-a924-4080-a7d1-bf7012b6725b-etc-swift") pod "swift-storage-0" (UID: "a672e32c-a924-4080-a7d1-bf7012b6725b") : configmap "swift-ring-files" not found Dec 11 08:28:02 crc kubenswrapper[4860]: I1211 08:28:02.501212 4860 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"swift-storage-0\" (UID: \"a672e32c-a924-4080-a7d1-bf7012b6725b\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/swift-storage-0" Dec 11 08:28:02 crc kubenswrapper[4860]: I1211 08:28:02.501436 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/a672e32c-a924-4080-a7d1-bf7012b6725b-lock\") pod \"swift-storage-0\" (UID: \"a672e32c-a924-4080-a7d1-bf7012b6725b\") " pod="openstack/swift-storage-0" Dec 11 08:28:02 crc kubenswrapper[4860]: I1211 08:28:02.502060 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/a672e32c-a924-4080-a7d1-bf7012b6725b-cache\") pod \"swift-storage-0\" (UID: \"a672e32c-a924-4080-a7d1-bf7012b6725b\") " pod="openstack/swift-storage-0" Dec 11 08:28:02 crc kubenswrapper[4860]: I1211 08:28:02.509318 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-l7vfs" event={"ID":"77f658e2-b468-45dd-8204-691cdfe9bf61","Type":"ContainerStarted","Data":"4696d8e509df5e9af91eb56115f4eaf5bb108a19f8ef126b9099787d80ca2d77"} Dec 11 08:28:02 crc kubenswrapper[4860]: I1211 08:28:02.511616 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"c0d5e719-9d28-4e31-9349-f882c4b72567","Type":"ContainerStarted","Data":"81911770e8c8e91309f7bc6761ff87d04b6573cc278f322e8c5fef89700e6235"} Dec 11 08:28:02 crc kubenswrapper[4860]: I1211 08:28:02.523904 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qsmps\" (UniqueName: \"kubernetes.io/projected/a672e32c-a924-4080-a7d1-bf7012b6725b-kube-api-access-qsmps\") pod \"swift-storage-0\" (UID: \"a672e32c-a924-4080-a7d1-bf7012b6725b\") " pod="openstack/swift-storage-0" Dec 11 08:28:02 crc kubenswrapper[4860]: I1211 08:28:02.525845 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"swift-storage-0\" (UID: \"a672e32c-a924-4080-a7d1-bf7012b6725b\") " pod="openstack/swift-storage-0" Dec 11 08:28:03 crc kubenswrapper[4860]: I1211 08:28:03.007513 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a672e32c-a924-4080-a7d1-bf7012b6725b-etc-swift\") pod \"swift-storage-0\" (UID: \"a672e32c-a924-4080-a7d1-bf7012b6725b\") " pod="openstack/swift-storage-0" Dec 11 08:28:03 crc kubenswrapper[4860]: E1211 08:28:03.007880 4860 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 11 08:28:03 crc kubenswrapper[4860]: E1211 08:28:03.008055 4860 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 11 08:28:03 crc kubenswrapper[4860]: E1211 08:28:03.008191 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/a672e32c-a924-4080-a7d1-bf7012b6725b-etc-swift podName:a672e32c-a924-4080-a7d1-bf7012b6725b nodeName:}" failed. No retries permitted until 2025-12-11 08:28:04.008140716 +0000 UTC m=+1016.736659781 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/a672e32c-a924-4080-a7d1-bf7012b6725b-etc-swift") pod "swift-storage-0" (UID: "a672e32c-a924-4080-a7d1-bf7012b6725b") : configmap "swift-ring-files" not found Dec 11 08:28:04 crc kubenswrapper[4860]: I1211 08:28:04.023993 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a672e32c-a924-4080-a7d1-bf7012b6725b-etc-swift\") pod \"swift-storage-0\" (UID: \"a672e32c-a924-4080-a7d1-bf7012b6725b\") " pod="openstack/swift-storage-0" Dec 11 08:28:04 crc kubenswrapper[4860]: E1211 08:28:04.024320 4860 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 11 08:28:04 crc kubenswrapper[4860]: E1211 08:28:04.024676 4860 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 11 08:28:04 crc kubenswrapper[4860]: E1211 08:28:04.024755 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/a672e32c-a924-4080-a7d1-bf7012b6725b-etc-swift podName:a672e32c-a924-4080-a7d1-bf7012b6725b nodeName:}" failed. No retries permitted until 2025-12-11 08:28:06.024736952 +0000 UTC m=+1018.753256007 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/a672e32c-a924-4080-a7d1-bf7012b6725b-etc-swift") pod "swift-storage-0" (UID: "a672e32c-a924-4080-a7d1-bf7012b6725b") : configmap "swift-ring-files" not found Dec 11 08:28:06 crc kubenswrapper[4860]: I1211 08:28:06.058908 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a672e32c-a924-4080-a7d1-bf7012b6725b-etc-swift\") pod \"swift-storage-0\" (UID: \"a672e32c-a924-4080-a7d1-bf7012b6725b\") " pod="openstack/swift-storage-0" Dec 11 08:28:06 crc kubenswrapper[4860]: E1211 08:28:06.059211 4860 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 11 08:28:06 crc kubenswrapper[4860]: E1211 08:28:06.059433 4860 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 11 08:28:06 crc kubenswrapper[4860]: E1211 08:28:06.059496 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/a672e32c-a924-4080-a7d1-bf7012b6725b-etc-swift podName:a672e32c-a924-4080-a7d1-bf7012b6725b nodeName:}" failed. No retries permitted until 2025-12-11 08:28:10.059478844 +0000 UTC m=+1022.787997899 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/a672e32c-a924-4080-a7d1-bf7012b6725b-etc-swift") pod "swift-storage-0" (UID: "a672e32c-a924-4080-a7d1-bf7012b6725b") : configmap "swift-ring-files" not found Dec 11 08:28:06 crc kubenswrapper[4860]: I1211 08:28:06.259264 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-bphmx"] Dec 11 08:28:06 crc kubenswrapper[4860]: I1211 08:28:06.260374 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-bphmx" Dec 11 08:28:06 crc kubenswrapper[4860]: I1211 08:28:06.264625 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 11 08:28:06 crc kubenswrapper[4860]: I1211 08:28:06.264719 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Dec 11 08:28:06 crc kubenswrapper[4860]: I1211 08:28:06.264724 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Dec 11 08:28:06 crc kubenswrapper[4860]: I1211 08:28:06.293841 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-bphmx"] Dec 11 08:28:06 crc kubenswrapper[4860]: I1211 08:28:06.363487 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79bcab38-b87b-44ff-ba2b-6027fd20719d-combined-ca-bundle\") pod \"swift-ring-rebalance-bphmx\" (UID: \"79bcab38-b87b-44ff-ba2b-6027fd20719d\") " pod="openstack/swift-ring-rebalance-bphmx" Dec 11 08:28:06 crc kubenswrapper[4860]: I1211 08:28:06.363558 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/79bcab38-b87b-44ff-ba2b-6027fd20719d-ring-data-devices\") pod \"swift-ring-rebalance-bphmx\" (UID: \"79bcab38-b87b-44ff-ba2b-6027fd20719d\") " pod="openstack/swift-ring-rebalance-bphmx" Dec 11 08:28:06 crc kubenswrapper[4860]: I1211 08:28:06.363708 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/79bcab38-b87b-44ff-ba2b-6027fd20719d-etc-swift\") pod \"swift-ring-rebalance-bphmx\" (UID: \"79bcab38-b87b-44ff-ba2b-6027fd20719d\") " pod="openstack/swift-ring-rebalance-bphmx" Dec 11 08:28:06 crc kubenswrapper[4860]: I1211 08:28:06.363736 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/79bcab38-b87b-44ff-ba2b-6027fd20719d-dispersionconf\") pod \"swift-ring-rebalance-bphmx\" (UID: \"79bcab38-b87b-44ff-ba2b-6027fd20719d\") " pod="openstack/swift-ring-rebalance-bphmx" Dec 11 08:28:06 crc kubenswrapper[4860]: I1211 08:28:06.363890 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/79bcab38-b87b-44ff-ba2b-6027fd20719d-swiftconf\") pod \"swift-ring-rebalance-bphmx\" (UID: \"79bcab38-b87b-44ff-ba2b-6027fd20719d\") " pod="openstack/swift-ring-rebalance-bphmx" Dec 11 08:28:06 crc kubenswrapper[4860]: I1211 08:28:06.363957 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/79bcab38-b87b-44ff-ba2b-6027fd20719d-scripts\") pod \"swift-ring-rebalance-bphmx\" (UID: \"79bcab38-b87b-44ff-ba2b-6027fd20719d\") " pod="openstack/swift-ring-rebalance-bphmx" Dec 11 08:28:06 crc kubenswrapper[4860]: I1211 08:28:06.364137 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gjh4n\" (UniqueName: \"kubernetes.io/projected/79bcab38-b87b-44ff-ba2b-6027fd20719d-kube-api-access-gjh4n\") pod \"swift-ring-rebalance-bphmx\" (UID: \"79bcab38-b87b-44ff-ba2b-6027fd20719d\") " pod="openstack/swift-ring-rebalance-bphmx" Dec 11 08:28:06 crc kubenswrapper[4860]: I1211 08:28:06.466057 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gjh4n\" (UniqueName: \"kubernetes.io/projected/79bcab38-b87b-44ff-ba2b-6027fd20719d-kube-api-access-gjh4n\") pod \"swift-ring-rebalance-bphmx\" (UID: \"79bcab38-b87b-44ff-ba2b-6027fd20719d\") " pod="openstack/swift-ring-rebalance-bphmx" Dec 11 08:28:06 crc kubenswrapper[4860]: I1211 08:28:06.466134 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79bcab38-b87b-44ff-ba2b-6027fd20719d-combined-ca-bundle\") pod \"swift-ring-rebalance-bphmx\" (UID: \"79bcab38-b87b-44ff-ba2b-6027fd20719d\") " pod="openstack/swift-ring-rebalance-bphmx" Dec 11 08:28:06 crc kubenswrapper[4860]: I1211 08:28:06.466183 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/79bcab38-b87b-44ff-ba2b-6027fd20719d-ring-data-devices\") pod \"swift-ring-rebalance-bphmx\" (UID: \"79bcab38-b87b-44ff-ba2b-6027fd20719d\") " pod="openstack/swift-ring-rebalance-bphmx" Dec 11 08:28:06 crc kubenswrapper[4860]: I1211 08:28:06.466284 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/79bcab38-b87b-44ff-ba2b-6027fd20719d-etc-swift\") pod \"swift-ring-rebalance-bphmx\" (UID: \"79bcab38-b87b-44ff-ba2b-6027fd20719d\") " pod="openstack/swift-ring-rebalance-bphmx" Dec 11 08:28:06 crc kubenswrapper[4860]: I1211 08:28:06.466311 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/79bcab38-b87b-44ff-ba2b-6027fd20719d-dispersionconf\") pod \"swift-ring-rebalance-bphmx\" (UID: \"79bcab38-b87b-44ff-ba2b-6027fd20719d\") " pod="openstack/swift-ring-rebalance-bphmx" Dec 11 08:28:06 crc kubenswrapper[4860]: I1211 08:28:06.466349 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/79bcab38-b87b-44ff-ba2b-6027fd20719d-swiftconf\") pod \"swift-ring-rebalance-bphmx\" (UID: \"79bcab38-b87b-44ff-ba2b-6027fd20719d\") " pod="openstack/swift-ring-rebalance-bphmx" Dec 11 08:28:06 crc kubenswrapper[4860]: I1211 08:28:06.466373 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/79bcab38-b87b-44ff-ba2b-6027fd20719d-scripts\") pod \"swift-ring-rebalance-bphmx\" (UID: \"79bcab38-b87b-44ff-ba2b-6027fd20719d\") " pod="openstack/swift-ring-rebalance-bphmx" Dec 11 08:28:06 crc kubenswrapper[4860]: I1211 08:28:06.466910 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/79bcab38-b87b-44ff-ba2b-6027fd20719d-etc-swift\") pod \"swift-ring-rebalance-bphmx\" (UID: \"79bcab38-b87b-44ff-ba2b-6027fd20719d\") " pod="openstack/swift-ring-rebalance-bphmx" Dec 11 08:28:06 crc kubenswrapper[4860]: I1211 08:28:06.467386 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/79bcab38-b87b-44ff-ba2b-6027fd20719d-scripts\") pod \"swift-ring-rebalance-bphmx\" (UID: \"79bcab38-b87b-44ff-ba2b-6027fd20719d\") " pod="openstack/swift-ring-rebalance-bphmx" Dec 11 08:28:06 crc kubenswrapper[4860]: I1211 08:28:06.468577 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/79bcab38-b87b-44ff-ba2b-6027fd20719d-ring-data-devices\") pod \"swift-ring-rebalance-bphmx\" (UID: \"79bcab38-b87b-44ff-ba2b-6027fd20719d\") " pod="openstack/swift-ring-rebalance-bphmx" Dec 11 08:28:06 crc kubenswrapper[4860]: I1211 08:28:06.473039 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/79bcab38-b87b-44ff-ba2b-6027fd20719d-dispersionconf\") pod \"swift-ring-rebalance-bphmx\" (UID: \"79bcab38-b87b-44ff-ba2b-6027fd20719d\") " pod="openstack/swift-ring-rebalance-bphmx" Dec 11 08:28:06 crc kubenswrapper[4860]: I1211 08:28:06.473430 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/79bcab38-b87b-44ff-ba2b-6027fd20719d-swiftconf\") pod \"swift-ring-rebalance-bphmx\" (UID: \"79bcab38-b87b-44ff-ba2b-6027fd20719d\") " pod="openstack/swift-ring-rebalance-bphmx" Dec 11 08:28:06 crc kubenswrapper[4860]: I1211 08:28:06.474052 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79bcab38-b87b-44ff-ba2b-6027fd20719d-combined-ca-bundle\") pod \"swift-ring-rebalance-bphmx\" (UID: \"79bcab38-b87b-44ff-ba2b-6027fd20719d\") " pod="openstack/swift-ring-rebalance-bphmx" Dec 11 08:28:06 crc kubenswrapper[4860]: I1211 08:28:06.485603 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gjh4n\" (UniqueName: \"kubernetes.io/projected/79bcab38-b87b-44ff-ba2b-6027fd20719d-kube-api-access-gjh4n\") pod \"swift-ring-rebalance-bphmx\" (UID: \"79bcab38-b87b-44ff-ba2b-6027fd20719d\") " pod="openstack/swift-ring-rebalance-bphmx" Dec 11 08:28:06 crc kubenswrapper[4860]: I1211 08:28:06.544487 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"c017f36f-48a3-4919-ac22-6a85a5c73530","Type":"ContainerStarted","Data":"4b36ced357fd343fbdcb589b954126968bfee0f50789d6176180ed107d2eddd2"} Dec 11 08:28:06 crc kubenswrapper[4860]: I1211 08:28:06.546760 4860 generic.go:334] "Generic (PLEG): container finished" podID="c4edc2e0-ea25-46f4-a127-b25e87244b28" containerID="5b92e8aeef3624eb41bb55b1fee8cad1e5714ab6bc8966344d337f3f80dc4cac" exitCode=0 Dec 11 08:28:06 crc kubenswrapper[4860]: I1211 08:28:06.546825 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-p5r62" event={"ID":"c4edc2e0-ea25-46f4-a127-b25e87244b28","Type":"ContainerDied","Data":"5b92e8aeef3624eb41bb55b1fee8cad1e5714ab6bc8966344d337f3f80dc4cac"} Dec 11 08:28:06 crc kubenswrapper[4860]: I1211 08:28:06.549019 4860 generic.go:334] "Generic (PLEG): container finished" podID="2319ad62-b109-4f60-880f-98c8da946fa6" containerID="aa2a2931d901fefa0c797b0acff2c9702219130a440e51cc0d1dd55092897a5e" exitCode=0 Dec 11 08:28:06 crc kubenswrapper[4860]: I1211 08:28:06.549135 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bc7876d45-fzprz" event={"ID":"2319ad62-b109-4f60-880f-98c8da946fa6","Type":"ContainerDied","Data":"aa2a2931d901fefa0c797b0acff2c9702219130a440e51cc0d1dd55092897a5e"} Dec 11 08:28:06 crc kubenswrapper[4860]: I1211 08:28:06.558804 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-lgzmd" event={"ID":"06395cef-744f-402e-844e-275a5025b81f","Type":"ContainerStarted","Data":"76e141446f818395627837eaa8471c2f8d5541f3579dfb4cfb9c44277aa38df3"} Dec 11 08:28:06 crc kubenswrapper[4860]: I1211 08:28:06.563462 4860 generic.go:334] "Generic (PLEG): container finished" podID="77f658e2-b468-45dd-8204-691cdfe9bf61" containerID="49c890dc4873c5f72a9b92e00eabd501e06c2c566ccf2eeab6288d19dddd536b" exitCode=0 Dec 11 08:28:06 crc kubenswrapper[4860]: I1211 08:28:06.563553 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-l7vfs" event={"ID":"77f658e2-b468-45dd-8204-691cdfe9bf61","Type":"ContainerDied","Data":"49c890dc4873c5f72a9b92e00eabd501e06c2c566ccf2eeab6288d19dddd536b"} Dec 11 08:28:06 crc kubenswrapper[4860]: I1211 08:28:06.573607 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"9657082d-ff9c-4a54-8b4d-54e6ea411590","Type":"ContainerStarted","Data":"aeaff7cacb7de3977c32dd7fbcda4229df3f4b1134a4493c87f3a195de1747dd"} Dec 11 08:28:06 crc kubenswrapper[4860]: I1211 08:28:06.577695 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=18.384876094 podStartE2EDuration="29.577671477s" podCreationTimestamp="2025-12-11 08:27:37 +0000 UTC" firstStartedPulling="2025-12-11 08:27:47.567191386 +0000 UTC m=+1000.295710441" lastFinishedPulling="2025-12-11 08:27:58.759986769 +0000 UTC m=+1011.488505824" observedRunningTime="2025-12-11 08:28:06.565241988 +0000 UTC m=+1019.293761043" watchObservedRunningTime="2025-12-11 08:28:06.577671477 +0000 UTC m=+1019.306190532" Dec 11 08:28:06 crc kubenswrapper[4860]: I1211 08:28:06.581106 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-bphmx" Dec 11 08:28:06 crc kubenswrapper[4860]: I1211 08:28:06.633584 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-lgzmd" podStartSLOduration=9.6335549 podStartE2EDuration="9.6335549s" podCreationTimestamp="2025-12-11 08:27:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:28:06.625392184 +0000 UTC m=+1019.353911259" watchObservedRunningTime="2025-12-11 08:28:06.6335549 +0000 UTC m=+1019.362073955" Dec 11 08:28:06 crc kubenswrapper[4860]: I1211 08:28:06.784786 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=33.409108638 podStartE2EDuration="40.784755923s" podCreationTimestamp="2025-12-11 08:27:26 +0000 UTC" firstStartedPulling="2025-12-11 08:27:45.256189339 +0000 UTC m=+997.984708394" lastFinishedPulling="2025-12-11 08:27:52.631836624 +0000 UTC m=+1005.360355679" observedRunningTime="2025-12-11 08:28:06.702935957 +0000 UTC m=+1019.431455102" watchObservedRunningTime="2025-12-11 08:28:06.784755923 +0000 UTC m=+1019.513274968" Dec 11 08:28:06 crc kubenswrapper[4860]: I1211 08:28:06.807086 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=32.795960912 podStartE2EDuration="39.807060012s" podCreationTimestamp="2025-12-11 08:27:27 +0000 UTC" firstStartedPulling="2025-12-11 08:27:45.891370082 +0000 UTC m=+998.619889137" lastFinishedPulling="2025-12-11 08:27:52.902469182 +0000 UTC m=+1005.630988237" observedRunningTime="2025-12-11 08:28:06.752333996 +0000 UTC m=+1019.480853051" watchObservedRunningTime="2025-12-11 08:28:06.807060012 +0000 UTC m=+1019.535579067" Dec 11 08:28:06 crc kubenswrapper[4860]: I1211 08:28:06.816128 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=18.004148258 podStartE2EDuration="30.816108126s" podCreationTimestamp="2025-12-11 08:27:36 +0000 UTC" firstStartedPulling="2025-12-11 08:27:45.901448869 +0000 UTC m=+998.629967934" lastFinishedPulling="2025-12-11 08:27:58.713408747 +0000 UTC m=+1011.441927802" observedRunningTime="2025-12-11 08:28:06.784089583 +0000 UTC m=+1019.512608638" watchObservedRunningTime="2025-12-11 08:28:06.816108126 +0000 UTC m=+1019.544627181" Dec 11 08:28:06 crc kubenswrapper[4860]: I1211 08:28:06.923828 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bc7876d45-fzprz" Dec 11 08:28:06 crc kubenswrapper[4860]: E1211 08:28:06.943299 4860 log.go:32] "CreateContainer in sandbox from runtime service failed" err=< Dec 11 08:28:06 crc kubenswrapper[4860]: rpc error: code = Unknown desc = container create failed: mount `/var/lib/kubelet/pods/c4edc2e0-ea25-46f4-a127-b25e87244b28/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Dec 11 08:28:06 crc kubenswrapper[4860]: > podSandboxID="59c9f2586b5d93fe3aca0b531a4082e900da0e764d39b54ff47b521939064be7" Dec 11 08:28:06 crc kubenswrapper[4860]: E1211 08:28:06.943718 4860 kuberuntime_manager.go:1274] "Unhandled Error" err=< Dec 11 08:28:06 crc kubenswrapper[4860]: container &Container{Name:dnsmasq-dns,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n654h99h64ch5dbh6dh555h587h64bh5cfh647h5fdh57ch679h9h597h5f5hbch59bh54fh575h566h667h586h5f5h65ch5bch57h68h65ch58bh694h5cfq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-nb,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/ovsdbserver-nb,SubPath:ovsdbserver-nb,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-sb,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/ovsdbserver-sb,SubPath:ovsdbserver-sb,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-jxdhf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-8554648995-p5r62_openstack(c4edc2e0-ea25-46f4-a127-b25e87244b28): CreateContainerError: container create failed: mount `/var/lib/kubelet/pods/c4edc2e0-ea25-46f4-a127-b25e87244b28/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Dec 11 08:28:06 crc kubenswrapper[4860]: > logger="UnhandledError" Dec 11 08:28:06 crc kubenswrapper[4860]: E1211 08:28:06.945273 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"dnsmasq-dns\" with CreateContainerError: \"container create failed: mount `/var/lib/kubelet/pods/c4edc2e0-ea25-46f4-a127-b25e87244b28/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory\\n\"" pod="openstack/dnsmasq-dns-8554648995-p5r62" podUID="c4edc2e0-ea25-46f4-a127-b25e87244b28" Dec 11 08:28:06 crc kubenswrapper[4860]: I1211 08:28:06.980338 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lpg5d\" (UniqueName: \"kubernetes.io/projected/2319ad62-b109-4f60-880f-98c8da946fa6-kube-api-access-lpg5d\") pod \"2319ad62-b109-4f60-880f-98c8da946fa6\" (UID: \"2319ad62-b109-4f60-880f-98c8da946fa6\") " Dec 11 08:28:06 crc kubenswrapper[4860]: I1211 08:28:06.980430 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2319ad62-b109-4f60-880f-98c8da946fa6-config\") pod \"2319ad62-b109-4f60-880f-98c8da946fa6\" (UID: \"2319ad62-b109-4f60-880f-98c8da946fa6\") " Dec 11 08:28:06 crc kubenswrapper[4860]: I1211 08:28:06.980520 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2319ad62-b109-4f60-880f-98c8da946fa6-dns-svc\") pod \"2319ad62-b109-4f60-880f-98c8da946fa6\" (UID: \"2319ad62-b109-4f60-880f-98c8da946fa6\") " Dec 11 08:28:06 crc kubenswrapper[4860]: I1211 08:28:06.980587 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2319ad62-b109-4f60-880f-98c8da946fa6-ovsdbserver-sb\") pod \"2319ad62-b109-4f60-880f-98c8da946fa6\" (UID: \"2319ad62-b109-4f60-880f-98c8da946fa6\") " Dec 11 08:28:06 crc kubenswrapper[4860]: I1211 08:28:06.986193 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2319ad62-b109-4f60-880f-98c8da946fa6-kube-api-access-lpg5d" (OuterVolumeSpecName: "kube-api-access-lpg5d") pod "2319ad62-b109-4f60-880f-98c8da946fa6" (UID: "2319ad62-b109-4f60-880f-98c8da946fa6"). InnerVolumeSpecName "kube-api-access-lpg5d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:28:06 crc kubenswrapper[4860]: I1211 08:28:06.998923 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-bphmx"] Dec 11 08:28:07 crc kubenswrapper[4860]: I1211 08:28:07.002740 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2319ad62-b109-4f60-880f-98c8da946fa6-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "2319ad62-b109-4f60-880f-98c8da946fa6" (UID: "2319ad62-b109-4f60-880f-98c8da946fa6"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:28:07 crc kubenswrapper[4860]: I1211 08:28:07.003404 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2319ad62-b109-4f60-880f-98c8da946fa6-config" (OuterVolumeSpecName: "config") pod "2319ad62-b109-4f60-880f-98c8da946fa6" (UID: "2319ad62-b109-4f60-880f-98c8da946fa6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:28:07 crc kubenswrapper[4860]: W1211 08:28:07.008025 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod79bcab38_b87b_44ff_ba2b_6027fd20719d.slice/crio-060cb1eff63ca3df9aae672b635a61c937f3ef8b11a707f07a7f3715f31ed27f WatchSource:0}: Error finding container 060cb1eff63ca3df9aae672b635a61c937f3ef8b11a707f07a7f3715f31ed27f: Status 404 returned error can't find the container with id 060cb1eff63ca3df9aae672b635a61c937f3ef8b11a707f07a7f3715f31ed27f Dec 11 08:28:07 crc kubenswrapper[4860]: I1211 08:28:07.010806 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2319ad62-b109-4f60-880f-98c8da946fa6-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "2319ad62-b109-4f60-880f-98c8da946fa6" (UID: "2319ad62-b109-4f60-880f-98c8da946fa6"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:28:07 crc kubenswrapper[4860]: I1211 08:28:07.083421 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lpg5d\" (UniqueName: \"kubernetes.io/projected/2319ad62-b109-4f60-880f-98c8da946fa6-kube-api-access-lpg5d\") on node \"crc\" DevicePath \"\"" Dec 11 08:28:07 crc kubenswrapper[4860]: I1211 08:28:07.083465 4860 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2319ad62-b109-4f60-880f-98c8da946fa6-config\") on node \"crc\" DevicePath \"\"" Dec 11 08:28:07 crc kubenswrapper[4860]: I1211 08:28:07.083493 4860 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2319ad62-b109-4f60-880f-98c8da946fa6-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 11 08:28:07 crc kubenswrapper[4860]: I1211 08:28:07.083506 4860 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2319ad62-b109-4f60-880f-98c8da946fa6-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 11 08:28:07 crc kubenswrapper[4860]: I1211 08:28:07.582013 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bc7876d45-fzprz" Dec 11 08:28:07 crc kubenswrapper[4860]: I1211 08:28:07.592289 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bc7876d45-fzprz" event={"ID":"2319ad62-b109-4f60-880f-98c8da946fa6","Type":"ContainerDied","Data":"305264eae8509d819b23ac35b0724986927586f0e7c048c9d2c2f709b8a266ac"} Dec 11 08:28:07 crc kubenswrapper[4860]: I1211 08:28:07.592583 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-bphmx" event={"ID":"79bcab38-b87b-44ff-ba2b-6027fd20719d","Type":"ContainerStarted","Data":"060cb1eff63ca3df9aae672b635a61c937f3ef8b11a707f07a7f3715f31ed27f"} Dec 11 08:28:07 crc kubenswrapper[4860]: I1211 08:28:07.592699 4860 scope.go:117] "RemoveContainer" containerID="aa2a2931d901fefa0c797b0acff2c9702219130a440e51cc0d1dd55092897a5e" Dec 11 08:28:07 crc kubenswrapper[4860]: I1211 08:28:07.632725 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Dec 11 08:28:07 crc kubenswrapper[4860]: I1211 08:28:07.633097 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Dec 11 08:28:07 crc kubenswrapper[4860]: I1211 08:28:07.692022 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bc7876d45-fzprz"] Dec 11 08:28:07 crc kubenswrapper[4860]: I1211 08:28:07.699296 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6bc7876d45-fzprz"] Dec 11 08:28:07 crc kubenswrapper[4860]: I1211 08:28:07.727337 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Dec 11 08:28:07 crc kubenswrapper[4860]: I1211 08:28:07.727400 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Dec 11 08:28:07 crc kubenswrapper[4860]: I1211 08:28:07.767290 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Dec 11 08:28:08 crc kubenswrapper[4860]: I1211 08:28:08.445878 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Dec 11 08:28:08 crc kubenswrapper[4860]: I1211 08:28:08.445939 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Dec 11 08:28:08 crc kubenswrapper[4860]: I1211 08:28:08.483222 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Dec 11 08:28:08 crc kubenswrapper[4860]: I1211 08:28:08.594470 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-l7vfs" event={"ID":"77f658e2-b468-45dd-8204-691cdfe9bf61","Type":"ContainerStarted","Data":"aca4cdc244eaa6aa25f8078b98879c5acf09174e9e804d8ee9f36118d2bd6cdb"} Dec 11 08:28:08 crc kubenswrapper[4860]: I1211 08:28:08.594817 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-b8fbc5445-l7vfs" Dec 11 08:28:08 crc kubenswrapper[4860]: I1211 08:28:08.601451 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-p5r62" event={"ID":"c4edc2e0-ea25-46f4-a127-b25e87244b28","Type":"ContainerStarted","Data":"8226dc997d8b1dae24bbc9cb06a1b475e115c465c6cc1569ec2cb131f69fb247"} Dec 11 08:28:08 crc kubenswrapper[4860]: I1211 08:28:08.622943 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-b8fbc5445-l7vfs" podStartSLOduration=7.622918959 podStartE2EDuration="7.622918959s" podCreationTimestamp="2025-12-11 08:28:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:28:08.615889568 +0000 UTC m=+1021.344408633" watchObservedRunningTime="2025-12-11 08:28:08.622918959 +0000 UTC m=+1021.351438014" Dec 11 08:28:08 crc kubenswrapper[4860]: I1211 08:28:08.634125 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-8554648995-p5r62" podStartSLOduration=11.63410914 podStartE2EDuration="11.63410914s" podCreationTimestamp="2025-12-11 08:27:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:28:08.631218229 +0000 UTC m=+1021.359737284" watchObservedRunningTime="2025-12-11 08:28:08.63410914 +0000 UTC m=+1021.362628185" Dec 11 08:28:08 crc kubenswrapper[4860]: I1211 08:28:08.648664 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Dec 11 08:28:08 crc kubenswrapper[4860]: I1211 08:28:08.671306 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Dec 11 08:28:08 crc kubenswrapper[4860]: I1211 08:28:08.795776 4860 patch_prober.go:28] interesting pod/machine-config-daemon-99qgp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 08:28:08 crc kubenswrapper[4860]: I1211 08:28:08.795876 4860 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 08:28:08 crc kubenswrapper[4860]: I1211 08:28:08.944887 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Dec 11 08:28:08 crc kubenswrapper[4860]: E1211 08:28:08.945306 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2319ad62-b109-4f60-880f-98c8da946fa6" containerName="init" Dec 11 08:28:08 crc kubenswrapper[4860]: I1211 08:28:08.945322 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="2319ad62-b109-4f60-880f-98c8da946fa6" containerName="init" Dec 11 08:28:08 crc kubenswrapper[4860]: I1211 08:28:08.945527 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="2319ad62-b109-4f60-880f-98c8da946fa6" containerName="init" Dec 11 08:28:08 crc kubenswrapper[4860]: I1211 08:28:08.946464 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 11 08:28:08 crc kubenswrapper[4860]: I1211 08:28:08.953103 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Dec 11 08:28:08 crc kubenswrapper[4860]: I1211 08:28:08.955630 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-jdzwr" Dec 11 08:28:08 crc kubenswrapper[4860]: I1211 08:28:08.955927 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Dec 11 08:28:08 crc kubenswrapper[4860]: I1211 08:28:08.956716 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Dec 11 08:28:08 crc kubenswrapper[4860]: I1211 08:28:08.970566 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 11 08:28:09 crc kubenswrapper[4860]: I1211 08:28:09.137755 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad53aa7d-2d02-4201-a90b-05bc8cca2f68-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"ad53aa7d-2d02-4201-a90b-05bc8cca2f68\") " pod="openstack/ovn-northd-0" Dec 11 08:28:09 crc kubenswrapper[4860]: I1211 08:28:09.137970 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sksls\" (UniqueName: \"kubernetes.io/projected/ad53aa7d-2d02-4201-a90b-05bc8cca2f68-kube-api-access-sksls\") pod \"ovn-northd-0\" (UID: \"ad53aa7d-2d02-4201-a90b-05bc8cca2f68\") " pod="openstack/ovn-northd-0" Dec 11 08:28:09 crc kubenswrapper[4860]: I1211 08:28:09.138005 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ad53aa7d-2d02-4201-a90b-05bc8cca2f68-config\") pod \"ovn-northd-0\" (UID: \"ad53aa7d-2d02-4201-a90b-05bc8cca2f68\") " pod="openstack/ovn-northd-0" Dec 11 08:28:09 crc kubenswrapper[4860]: I1211 08:28:09.138085 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/ad53aa7d-2d02-4201-a90b-05bc8cca2f68-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"ad53aa7d-2d02-4201-a90b-05bc8cca2f68\") " pod="openstack/ovn-northd-0" Dec 11 08:28:09 crc kubenswrapper[4860]: I1211 08:28:09.138162 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ad53aa7d-2d02-4201-a90b-05bc8cca2f68-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"ad53aa7d-2d02-4201-a90b-05bc8cca2f68\") " pod="openstack/ovn-northd-0" Dec 11 08:28:09 crc kubenswrapper[4860]: I1211 08:28:09.138213 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ad53aa7d-2d02-4201-a90b-05bc8cca2f68-scripts\") pod \"ovn-northd-0\" (UID: \"ad53aa7d-2d02-4201-a90b-05bc8cca2f68\") " pod="openstack/ovn-northd-0" Dec 11 08:28:09 crc kubenswrapper[4860]: I1211 08:28:09.138247 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ad53aa7d-2d02-4201-a90b-05bc8cca2f68-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"ad53aa7d-2d02-4201-a90b-05bc8cca2f68\") " pod="openstack/ovn-northd-0" Dec 11 08:28:09 crc kubenswrapper[4860]: I1211 08:28:09.240288 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sksls\" (UniqueName: \"kubernetes.io/projected/ad53aa7d-2d02-4201-a90b-05bc8cca2f68-kube-api-access-sksls\") pod \"ovn-northd-0\" (UID: \"ad53aa7d-2d02-4201-a90b-05bc8cca2f68\") " pod="openstack/ovn-northd-0" Dec 11 08:28:09 crc kubenswrapper[4860]: I1211 08:28:09.240346 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ad53aa7d-2d02-4201-a90b-05bc8cca2f68-config\") pod \"ovn-northd-0\" (UID: \"ad53aa7d-2d02-4201-a90b-05bc8cca2f68\") " pod="openstack/ovn-northd-0" Dec 11 08:28:09 crc kubenswrapper[4860]: I1211 08:28:09.240376 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/ad53aa7d-2d02-4201-a90b-05bc8cca2f68-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"ad53aa7d-2d02-4201-a90b-05bc8cca2f68\") " pod="openstack/ovn-northd-0" Dec 11 08:28:09 crc kubenswrapper[4860]: I1211 08:28:09.240404 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ad53aa7d-2d02-4201-a90b-05bc8cca2f68-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"ad53aa7d-2d02-4201-a90b-05bc8cca2f68\") " pod="openstack/ovn-northd-0" Dec 11 08:28:09 crc kubenswrapper[4860]: I1211 08:28:09.240439 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ad53aa7d-2d02-4201-a90b-05bc8cca2f68-scripts\") pod \"ovn-northd-0\" (UID: \"ad53aa7d-2d02-4201-a90b-05bc8cca2f68\") " pod="openstack/ovn-northd-0" Dec 11 08:28:09 crc kubenswrapper[4860]: I1211 08:28:09.240463 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ad53aa7d-2d02-4201-a90b-05bc8cca2f68-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"ad53aa7d-2d02-4201-a90b-05bc8cca2f68\") " pod="openstack/ovn-northd-0" Dec 11 08:28:09 crc kubenswrapper[4860]: I1211 08:28:09.240550 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad53aa7d-2d02-4201-a90b-05bc8cca2f68-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"ad53aa7d-2d02-4201-a90b-05bc8cca2f68\") " pod="openstack/ovn-northd-0" Dec 11 08:28:09 crc kubenswrapper[4860]: I1211 08:28:09.241369 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ad53aa7d-2d02-4201-a90b-05bc8cca2f68-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"ad53aa7d-2d02-4201-a90b-05bc8cca2f68\") " pod="openstack/ovn-northd-0" Dec 11 08:28:09 crc kubenswrapper[4860]: I1211 08:28:09.241729 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ad53aa7d-2d02-4201-a90b-05bc8cca2f68-scripts\") pod \"ovn-northd-0\" (UID: \"ad53aa7d-2d02-4201-a90b-05bc8cca2f68\") " pod="openstack/ovn-northd-0" Dec 11 08:28:09 crc kubenswrapper[4860]: I1211 08:28:09.241807 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ad53aa7d-2d02-4201-a90b-05bc8cca2f68-config\") pod \"ovn-northd-0\" (UID: \"ad53aa7d-2d02-4201-a90b-05bc8cca2f68\") " pod="openstack/ovn-northd-0" Dec 11 08:28:09 crc kubenswrapper[4860]: I1211 08:28:09.251615 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad53aa7d-2d02-4201-a90b-05bc8cca2f68-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"ad53aa7d-2d02-4201-a90b-05bc8cca2f68\") " pod="openstack/ovn-northd-0" Dec 11 08:28:09 crc kubenswrapper[4860]: I1211 08:28:09.270573 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ad53aa7d-2d02-4201-a90b-05bc8cca2f68-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"ad53aa7d-2d02-4201-a90b-05bc8cca2f68\") " pod="openstack/ovn-northd-0" Dec 11 08:28:09 crc kubenswrapper[4860]: I1211 08:28:09.274484 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/ad53aa7d-2d02-4201-a90b-05bc8cca2f68-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"ad53aa7d-2d02-4201-a90b-05bc8cca2f68\") " pod="openstack/ovn-northd-0" Dec 11 08:28:09 crc kubenswrapper[4860]: I1211 08:28:09.284817 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sksls\" (UniqueName: \"kubernetes.io/projected/ad53aa7d-2d02-4201-a90b-05bc8cca2f68-kube-api-access-sksls\") pod \"ovn-northd-0\" (UID: \"ad53aa7d-2d02-4201-a90b-05bc8cca2f68\") " pod="openstack/ovn-northd-0" Dec 11 08:28:09 crc kubenswrapper[4860]: I1211 08:28:09.308001 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Dec 11 08:28:09 crc kubenswrapper[4860]: I1211 08:28:09.309956 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Dec 11 08:28:09 crc kubenswrapper[4860]: I1211 08:28:09.577204 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 11 08:28:09 crc kubenswrapper[4860]: I1211 08:28:09.589592 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2319ad62-b109-4f60-880f-98c8da946fa6" path="/var/lib/kubelet/pods/2319ad62-b109-4f60-880f-98c8da946fa6/volumes" Dec 11 08:28:10 crc kubenswrapper[4860]: I1211 08:28:10.160654 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a672e32c-a924-4080-a7d1-bf7012b6725b-etc-swift\") pod \"swift-storage-0\" (UID: \"a672e32c-a924-4080-a7d1-bf7012b6725b\") " pod="openstack/swift-storage-0" Dec 11 08:28:10 crc kubenswrapper[4860]: E1211 08:28:10.161053 4860 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 11 08:28:10 crc kubenswrapper[4860]: E1211 08:28:10.161069 4860 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 11 08:28:10 crc kubenswrapper[4860]: E1211 08:28:10.161123 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/a672e32c-a924-4080-a7d1-bf7012b6725b-etc-swift podName:a672e32c-a924-4080-a7d1-bf7012b6725b nodeName:}" failed. No retries permitted until 2025-12-11 08:28:18.161106105 +0000 UTC m=+1030.889625150 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/a672e32c-a924-4080-a7d1-bf7012b6725b-etc-swift") pod "swift-storage-0" (UID: "a672e32c-a924-4080-a7d1-bf7012b6725b") : configmap "swift-ring-files" not found Dec 11 08:28:11 crc kubenswrapper[4860]: W1211 08:28:11.227127 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podad53aa7d_2d02_4201_a90b_05bc8cca2f68.slice/crio-2aeae4658d143141e39fa1de02f6160164708f597a2440b904f435003dc2f944 WatchSource:0}: Error finding container 2aeae4658d143141e39fa1de02f6160164708f597a2440b904f435003dc2f944: Status 404 returned error can't find the container with id 2aeae4658d143141e39fa1de02f6160164708f597a2440b904f435003dc2f944 Dec 11 08:28:11 crc kubenswrapper[4860]: I1211 08:28:11.227883 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 11 08:28:11 crc kubenswrapper[4860]: I1211 08:28:11.510265 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Dec 11 08:28:11 crc kubenswrapper[4860]: I1211 08:28:11.602996 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Dec 11 08:28:11 crc kubenswrapper[4860]: I1211 08:28:11.629963 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-bphmx" event={"ID":"79bcab38-b87b-44ff-ba2b-6027fd20719d","Type":"ContainerStarted","Data":"0846a272687f0e3b7ae72decde0f1570151a373cb45061265768e42f24780dcf"} Dec 11 08:28:11 crc kubenswrapper[4860]: I1211 08:28:11.635258 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"ad53aa7d-2d02-4201-a90b-05bc8cca2f68","Type":"ContainerStarted","Data":"2aeae4658d143141e39fa1de02f6160164708f597a2440b904f435003dc2f944"} Dec 11 08:28:11 crc kubenswrapper[4860]: I1211 08:28:11.669088 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-bphmx" podStartSLOduration=1.8672514169999999 podStartE2EDuration="5.669069745s" podCreationTimestamp="2025-12-11 08:28:06 +0000 UTC" firstStartedPulling="2025-12-11 08:28:07.010087831 +0000 UTC m=+1019.738606886" lastFinishedPulling="2025-12-11 08:28:10.811906159 +0000 UTC m=+1023.540425214" observedRunningTime="2025-12-11 08:28:11.664965345 +0000 UTC m=+1024.393484420" watchObservedRunningTime="2025-12-11 08:28:11.669069745 +0000 UTC m=+1024.397588800" Dec 11 08:28:13 crc kubenswrapper[4860]: I1211 08:28:13.097170 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-8554648995-p5r62" Dec 11 08:28:13 crc kubenswrapper[4860]: I1211 08:28:13.099442 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-8554648995-p5r62" Dec 11 08:28:13 crc kubenswrapper[4860]: I1211 08:28:13.653363 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"ad53aa7d-2d02-4201-a90b-05bc8cca2f68","Type":"ContainerStarted","Data":"c1ae7fd5f9b71b21447dc6efab693cba77eea1ae599afc6964a22201d795d0a3"} Dec 11 08:28:13 crc kubenswrapper[4860]: I1211 08:28:13.653876 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Dec 11 08:28:13 crc kubenswrapper[4860]: I1211 08:28:13.653897 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"ad53aa7d-2d02-4201-a90b-05bc8cca2f68","Type":"ContainerStarted","Data":"82ecba1596ef92054fc5c876d3193ecf58e6f0848940bd755ad7ad1e9341cd8c"} Dec 11 08:28:13 crc kubenswrapper[4860]: I1211 08:28:13.679896 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=3.96006314 podStartE2EDuration="5.679837203s" podCreationTimestamp="2025-12-11 08:28:08 +0000 UTC" firstStartedPulling="2025-12-11 08:28:11.229988021 +0000 UTC m=+1023.958507076" lastFinishedPulling="2025-12-11 08:28:12.949762084 +0000 UTC m=+1025.678281139" observedRunningTime="2025-12-11 08:28:13.672240905 +0000 UTC m=+1026.400759990" watchObservedRunningTime="2025-12-11 08:28:13.679837203 +0000 UTC m=+1026.408356288" Dec 11 08:28:13 crc kubenswrapper[4860]: I1211 08:28:13.744458 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Dec 11 08:28:13 crc kubenswrapper[4860]: I1211 08:28:13.814675 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Dec 11 08:28:16 crc kubenswrapper[4860]: I1211 08:28:16.649799 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-b8fbc5445-l7vfs" Dec 11 08:28:16 crc kubenswrapper[4860]: I1211 08:28:16.741249 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8554648995-p5r62"] Dec 11 08:28:16 crc kubenswrapper[4860]: I1211 08:28:16.741605 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-8554648995-p5r62" podUID="c4edc2e0-ea25-46f4-a127-b25e87244b28" containerName="dnsmasq-dns" containerID="cri-o://8226dc997d8b1dae24bbc9cb06a1b475e115c465c6cc1569ec2cb131f69fb247" gracePeriod=10 Dec 11 08:28:17 crc kubenswrapper[4860]: I1211 08:28:17.181074 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-p5r62" Dec 11 08:28:17 crc kubenswrapper[4860]: I1211 08:28:17.208292 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jxdhf\" (UniqueName: \"kubernetes.io/projected/c4edc2e0-ea25-46f4-a127-b25e87244b28-kube-api-access-jxdhf\") pod \"c4edc2e0-ea25-46f4-a127-b25e87244b28\" (UID: \"c4edc2e0-ea25-46f4-a127-b25e87244b28\") " Dec 11 08:28:17 crc kubenswrapper[4860]: I1211 08:28:17.208356 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c4edc2e0-ea25-46f4-a127-b25e87244b28-ovsdbserver-sb\") pod \"c4edc2e0-ea25-46f4-a127-b25e87244b28\" (UID: \"c4edc2e0-ea25-46f4-a127-b25e87244b28\") " Dec 11 08:28:17 crc kubenswrapper[4860]: I1211 08:28:17.208541 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c4edc2e0-ea25-46f4-a127-b25e87244b28-dns-svc\") pod \"c4edc2e0-ea25-46f4-a127-b25e87244b28\" (UID: \"c4edc2e0-ea25-46f4-a127-b25e87244b28\") " Dec 11 08:28:17 crc kubenswrapper[4860]: I1211 08:28:17.208625 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c4edc2e0-ea25-46f4-a127-b25e87244b28-ovsdbserver-nb\") pod \"c4edc2e0-ea25-46f4-a127-b25e87244b28\" (UID: \"c4edc2e0-ea25-46f4-a127-b25e87244b28\") " Dec 11 08:28:17 crc kubenswrapper[4860]: I1211 08:28:17.208689 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4edc2e0-ea25-46f4-a127-b25e87244b28-config\") pod \"c4edc2e0-ea25-46f4-a127-b25e87244b28\" (UID: \"c4edc2e0-ea25-46f4-a127-b25e87244b28\") " Dec 11 08:28:17 crc kubenswrapper[4860]: I1211 08:28:17.227368 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c4edc2e0-ea25-46f4-a127-b25e87244b28-kube-api-access-jxdhf" (OuterVolumeSpecName: "kube-api-access-jxdhf") pod "c4edc2e0-ea25-46f4-a127-b25e87244b28" (UID: "c4edc2e0-ea25-46f4-a127-b25e87244b28"). InnerVolumeSpecName "kube-api-access-jxdhf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:28:17 crc kubenswrapper[4860]: I1211 08:28:17.266404 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c4edc2e0-ea25-46f4-a127-b25e87244b28-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "c4edc2e0-ea25-46f4-a127-b25e87244b28" (UID: "c4edc2e0-ea25-46f4-a127-b25e87244b28"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:28:17 crc kubenswrapper[4860]: I1211 08:28:17.270049 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c4edc2e0-ea25-46f4-a127-b25e87244b28-config" (OuterVolumeSpecName: "config") pod "c4edc2e0-ea25-46f4-a127-b25e87244b28" (UID: "c4edc2e0-ea25-46f4-a127-b25e87244b28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:28:17 crc kubenswrapper[4860]: I1211 08:28:17.280487 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c4edc2e0-ea25-46f4-a127-b25e87244b28-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c4edc2e0-ea25-46f4-a127-b25e87244b28" (UID: "c4edc2e0-ea25-46f4-a127-b25e87244b28"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:28:17 crc kubenswrapper[4860]: I1211 08:28:17.285634 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c4edc2e0-ea25-46f4-a127-b25e87244b28-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "c4edc2e0-ea25-46f4-a127-b25e87244b28" (UID: "c4edc2e0-ea25-46f4-a127-b25e87244b28"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:28:17 crc kubenswrapper[4860]: I1211 08:28:17.311417 4860 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c4edc2e0-ea25-46f4-a127-b25e87244b28-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 11 08:28:17 crc kubenswrapper[4860]: I1211 08:28:17.311453 4860 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c4edc2e0-ea25-46f4-a127-b25e87244b28-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 11 08:28:17 crc kubenswrapper[4860]: I1211 08:28:17.311467 4860 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4edc2e0-ea25-46f4-a127-b25e87244b28-config\") on node \"crc\" DevicePath \"\"" Dec 11 08:28:17 crc kubenswrapper[4860]: I1211 08:28:17.311481 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jxdhf\" (UniqueName: \"kubernetes.io/projected/c4edc2e0-ea25-46f4-a127-b25e87244b28-kube-api-access-jxdhf\") on node \"crc\" DevicePath \"\"" Dec 11 08:28:17 crc kubenswrapper[4860]: I1211 08:28:17.311492 4860 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c4edc2e0-ea25-46f4-a127-b25e87244b28-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 11 08:28:17 crc kubenswrapper[4860]: I1211 08:28:17.699878 4860 generic.go:334] "Generic (PLEG): container finished" podID="c4edc2e0-ea25-46f4-a127-b25e87244b28" containerID="8226dc997d8b1dae24bbc9cb06a1b475e115c465c6cc1569ec2cb131f69fb247" exitCode=0 Dec 11 08:28:17 crc kubenswrapper[4860]: I1211 08:28:17.699927 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-p5r62" event={"ID":"c4edc2e0-ea25-46f4-a127-b25e87244b28","Type":"ContainerDied","Data":"8226dc997d8b1dae24bbc9cb06a1b475e115c465c6cc1569ec2cb131f69fb247"} Dec 11 08:28:17 crc kubenswrapper[4860]: I1211 08:28:17.699958 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-p5r62" event={"ID":"c4edc2e0-ea25-46f4-a127-b25e87244b28","Type":"ContainerDied","Data":"59c9f2586b5d93fe3aca0b531a4082e900da0e764d39b54ff47b521939064be7"} Dec 11 08:28:17 crc kubenswrapper[4860]: I1211 08:28:17.699978 4860 scope.go:117] "RemoveContainer" containerID="8226dc997d8b1dae24bbc9cb06a1b475e115c465c6cc1569ec2cb131f69fb247" Dec 11 08:28:17 crc kubenswrapper[4860]: I1211 08:28:17.700107 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-p5r62" Dec 11 08:28:17 crc kubenswrapper[4860]: I1211 08:28:17.728363 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8554648995-p5r62"] Dec 11 08:28:17 crc kubenswrapper[4860]: I1211 08:28:17.738061 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-8554648995-p5r62"] Dec 11 08:28:17 crc kubenswrapper[4860]: I1211 08:28:17.744764 4860 scope.go:117] "RemoveContainer" containerID="5b92e8aeef3624eb41bb55b1fee8cad1e5714ab6bc8966344d337f3f80dc4cac" Dec 11 08:28:17 crc kubenswrapper[4860]: I1211 08:28:17.767258 4860 scope.go:117] "RemoveContainer" containerID="8226dc997d8b1dae24bbc9cb06a1b475e115c465c6cc1569ec2cb131f69fb247" Dec 11 08:28:17 crc kubenswrapper[4860]: E1211 08:28:17.768081 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8226dc997d8b1dae24bbc9cb06a1b475e115c465c6cc1569ec2cb131f69fb247\": container with ID starting with 8226dc997d8b1dae24bbc9cb06a1b475e115c465c6cc1569ec2cb131f69fb247 not found: ID does not exist" containerID="8226dc997d8b1dae24bbc9cb06a1b475e115c465c6cc1569ec2cb131f69fb247" Dec 11 08:28:17 crc kubenswrapper[4860]: I1211 08:28:17.768126 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8226dc997d8b1dae24bbc9cb06a1b475e115c465c6cc1569ec2cb131f69fb247"} err="failed to get container status \"8226dc997d8b1dae24bbc9cb06a1b475e115c465c6cc1569ec2cb131f69fb247\": rpc error: code = NotFound desc = could not find container \"8226dc997d8b1dae24bbc9cb06a1b475e115c465c6cc1569ec2cb131f69fb247\": container with ID starting with 8226dc997d8b1dae24bbc9cb06a1b475e115c465c6cc1569ec2cb131f69fb247 not found: ID does not exist" Dec 11 08:28:17 crc kubenswrapper[4860]: I1211 08:28:17.768156 4860 scope.go:117] "RemoveContainer" containerID="5b92e8aeef3624eb41bb55b1fee8cad1e5714ab6bc8966344d337f3f80dc4cac" Dec 11 08:28:17 crc kubenswrapper[4860]: E1211 08:28:17.768498 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5b92e8aeef3624eb41bb55b1fee8cad1e5714ab6bc8966344d337f3f80dc4cac\": container with ID starting with 5b92e8aeef3624eb41bb55b1fee8cad1e5714ab6bc8966344d337f3f80dc4cac not found: ID does not exist" containerID="5b92e8aeef3624eb41bb55b1fee8cad1e5714ab6bc8966344d337f3f80dc4cac" Dec 11 08:28:17 crc kubenswrapper[4860]: I1211 08:28:17.768523 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5b92e8aeef3624eb41bb55b1fee8cad1e5714ab6bc8966344d337f3f80dc4cac"} err="failed to get container status \"5b92e8aeef3624eb41bb55b1fee8cad1e5714ab6bc8966344d337f3f80dc4cac\": rpc error: code = NotFound desc = could not find container \"5b92e8aeef3624eb41bb55b1fee8cad1e5714ab6bc8966344d337f3f80dc4cac\": container with ID starting with 5b92e8aeef3624eb41bb55b1fee8cad1e5714ab6bc8966344d337f3f80dc4cac not found: ID does not exist" Dec 11 08:28:18 crc kubenswrapper[4860]: I1211 08:28:18.236617 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a672e32c-a924-4080-a7d1-bf7012b6725b-etc-swift\") pod \"swift-storage-0\" (UID: \"a672e32c-a924-4080-a7d1-bf7012b6725b\") " pod="openstack/swift-storage-0" Dec 11 08:28:18 crc kubenswrapper[4860]: I1211 08:28:18.242894 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a672e32c-a924-4080-a7d1-bf7012b6725b-etc-swift\") pod \"swift-storage-0\" (UID: \"a672e32c-a924-4080-a7d1-bf7012b6725b\") " pod="openstack/swift-storage-0" Dec 11 08:28:18 crc kubenswrapper[4860]: I1211 08:28:18.297039 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 11 08:28:18 crc kubenswrapper[4860]: I1211 08:28:18.710535 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-bphmx" event={"ID":"79bcab38-b87b-44ff-ba2b-6027fd20719d","Type":"ContainerDied","Data":"0846a272687f0e3b7ae72decde0f1570151a373cb45061265768e42f24780dcf"} Dec 11 08:28:18 crc kubenswrapper[4860]: I1211 08:28:18.710537 4860 generic.go:334] "Generic (PLEG): container finished" podID="79bcab38-b87b-44ff-ba2b-6027fd20719d" containerID="0846a272687f0e3b7ae72decde0f1570151a373cb45061265768e42f24780dcf" exitCode=0 Dec 11 08:28:18 crc kubenswrapper[4860]: I1211 08:28:18.733112 4860 generic.go:334] "Generic (PLEG): container finished" podID="a004cb7b-ee08-44c0-9dc7-d0b69509fe13" containerID="7b2484370445a6fcc6a996d0a46f40caeeb99158af08c51c6b35e54bd1ef3322" exitCode=0 Dec 11 08:28:18 crc kubenswrapper[4860]: I1211 08:28:18.733231 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"a004cb7b-ee08-44c0-9dc7-d0b69509fe13","Type":"ContainerDied","Data":"7b2484370445a6fcc6a996d0a46f40caeeb99158af08c51c6b35e54bd1ef3322"} Dec 11 08:28:18 crc kubenswrapper[4860]: I1211 08:28:18.737117 4860 generic.go:334] "Generic (PLEG): container finished" podID="97634ec0-7953-4e38-a07e-95967ce874d7" containerID="c1aae3ef18a4011e77f7a6a7c3f8803d1d98f58a1d231ac7824be320f058f72a" exitCode=0 Dec 11 08:28:18 crc kubenswrapper[4860]: I1211 08:28:18.737162 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"97634ec0-7953-4e38-a07e-95967ce874d7","Type":"ContainerDied","Data":"c1aae3ef18a4011e77f7a6a7c3f8803d1d98f58a1d231ac7824be320f058f72a"} Dec 11 08:28:18 crc kubenswrapper[4860]: I1211 08:28:18.841357 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Dec 11 08:28:18 crc kubenswrapper[4860]: I1211 08:28:18.988902 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-66fgz"] Dec 11 08:28:18 crc kubenswrapper[4860]: E1211 08:28:18.989505 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4edc2e0-ea25-46f4-a127-b25e87244b28" containerName="init" Dec 11 08:28:18 crc kubenswrapper[4860]: I1211 08:28:18.989524 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4edc2e0-ea25-46f4-a127-b25e87244b28" containerName="init" Dec 11 08:28:18 crc kubenswrapper[4860]: E1211 08:28:18.989579 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4edc2e0-ea25-46f4-a127-b25e87244b28" containerName="dnsmasq-dns" Dec 11 08:28:18 crc kubenswrapper[4860]: I1211 08:28:18.989587 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4edc2e0-ea25-46f4-a127-b25e87244b28" containerName="dnsmasq-dns" Dec 11 08:28:18 crc kubenswrapper[4860]: I1211 08:28:18.989826 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4edc2e0-ea25-46f4-a127-b25e87244b28" containerName="dnsmasq-dns" Dec 11 08:28:18 crc kubenswrapper[4860]: I1211 08:28:18.990614 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-66fgz" Dec 11 08:28:19 crc kubenswrapper[4860]: I1211 08:28:19.004712 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-d5e8-account-create-update-76hlk"] Dec 11 08:28:19 crc kubenswrapper[4860]: I1211 08:28:19.010485 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-d5e8-account-create-update-76hlk" Dec 11 08:28:19 crc kubenswrapper[4860]: I1211 08:28:19.014203 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Dec 11 08:28:19 crc kubenswrapper[4860]: I1211 08:28:19.018326 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-66fgz"] Dec 11 08:28:19 crc kubenswrapper[4860]: I1211 08:28:19.033920 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-d5e8-account-create-update-76hlk"] Dec 11 08:28:19 crc kubenswrapper[4860]: I1211 08:28:19.156695 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/79aae41e-c682-42c8-8bbc-5d060804d535-operator-scripts\") pod \"keystone-db-create-66fgz\" (UID: \"79aae41e-c682-42c8-8bbc-5d060804d535\") " pod="openstack/keystone-db-create-66fgz" Dec 11 08:28:19 crc kubenswrapper[4860]: I1211 08:28:19.157731 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/25bead29-f8d6-4539-97f8-9c3be58fbf64-operator-scripts\") pod \"keystone-d5e8-account-create-update-76hlk\" (UID: \"25bead29-f8d6-4539-97f8-9c3be58fbf64\") " pod="openstack/keystone-d5e8-account-create-update-76hlk" Dec 11 08:28:19 crc kubenswrapper[4860]: I1211 08:28:19.157863 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dkg2w\" (UniqueName: \"kubernetes.io/projected/25bead29-f8d6-4539-97f8-9c3be58fbf64-kube-api-access-dkg2w\") pod \"keystone-d5e8-account-create-update-76hlk\" (UID: \"25bead29-f8d6-4539-97f8-9c3be58fbf64\") " pod="openstack/keystone-d5e8-account-create-update-76hlk" Dec 11 08:28:19 crc kubenswrapper[4860]: I1211 08:28:19.158148 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8jmpz\" (UniqueName: \"kubernetes.io/projected/79aae41e-c682-42c8-8bbc-5d060804d535-kube-api-access-8jmpz\") pod \"keystone-db-create-66fgz\" (UID: \"79aae41e-c682-42c8-8bbc-5d060804d535\") " pod="openstack/keystone-db-create-66fgz" Dec 11 08:28:19 crc kubenswrapper[4860]: I1211 08:28:19.259629 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8jmpz\" (UniqueName: \"kubernetes.io/projected/79aae41e-c682-42c8-8bbc-5d060804d535-kube-api-access-8jmpz\") pod \"keystone-db-create-66fgz\" (UID: \"79aae41e-c682-42c8-8bbc-5d060804d535\") " pod="openstack/keystone-db-create-66fgz" Dec 11 08:28:19 crc kubenswrapper[4860]: I1211 08:28:19.259733 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/79aae41e-c682-42c8-8bbc-5d060804d535-operator-scripts\") pod \"keystone-db-create-66fgz\" (UID: \"79aae41e-c682-42c8-8bbc-5d060804d535\") " pod="openstack/keystone-db-create-66fgz" Dec 11 08:28:19 crc kubenswrapper[4860]: I1211 08:28:19.259769 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/25bead29-f8d6-4539-97f8-9c3be58fbf64-operator-scripts\") pod \"keystone-d5e8-account-create-update-76hlk\" (UID: \"25bead29-f8d6-4539-97f8-9c3be58fbf64\") " pod="openstack/keystone-d5e8-account-create-update-76hlk" Dec 11 08:28:19 crc kubenswrapper[4860]: I1211 08:28:19.259801 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dkg2w\" (UniqueName: \"kubernetes.io/projected/25bead29-f8d6-4539-97f8-9c3be58fbf64-kube-api-access-dkg2w\") pod \"keystone-d5e8-account-create-update-76hlk\" (UID: \"25bead29-f8d6-4539-97f8-9c3be58fbf64\") " pod="openstack/keystone-d5e8-account-create-update-76hlk" Dec 11 08:28:19 crc kubenswrapper[4860]: I1211 08:28:19.260697 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/79aae41e-c682-42c8-8bbc-5d060804d535-operator-scripts\") pod \"keystone-db-create-66fgz\" (UID: \"79aae41e-c682-42c8-8bbc-5d060804d535\") " pod="openstack/keystone-db-create-66fgz" Dec 11 08:28:19 crc kubenswrapper[4860]: I1211 08:28:19.261099 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/25bead29-f8d6-4539-97f8-9c3be58fbf64-operator-scripts\") pod \"keystone-d5e8-account-create-update-76hlk\" (UID: \"25bead29-f8d6-4539-97f8-9c3be58fbf64\") " pod="openstack/keystone-d5e8-account-create-update-76hlk" Dec 11 08:28:19 crc kubenswrapper[4860]: I1211 08:28:19.290537 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-x7qkx"] Dec 11 08:28:19 crc kubenswrapper[4860]: I1211 08:28:19.292405 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-x7qkx" Dec 11 08:28:19 crc kubenswrapper[4860]: I1211 08:28:19.298781 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8jmpz\" (UniqueName: \"kubernetes.io/projected/79aae41e-c682-42c8-8bbc-5d060804d535-kube-api-access-8jmpz\") pod \"keystone-db-create-66fgz\" (UID: \"79aae41e-c682-42c8-8bbc-5d060804d535\") " pod="openstack/keystone-db-create-66fgz" Dec 11 08:28:19 crc kubenswrapper[4860]: I1211 08:28:19.299547 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dkg2w\" (UniqueName: \"kubernetes.io/projected/25bead29-f8d6-4539-97f8-9c3be58fbf64-kube-api-access-dkg2w\") pod \"keystone-d5e8-account-create-update-76hlk\" (UID: \"25bead29-f8d6-4539-97f8-9c3be58fbf64\") " pod="openstack/keystone-d5e8-account-create-update-76hlk" Dec 11 08:28:19 crc kubenswrapper[4860]: I1211 08:28:19.305476 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-x7qkx"] Dec 11 08:28:19 crc kubenswrapper[4860]: I1211 08:28:19.336701 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-66fgz" Dec 11 08:28:19 crc kubenswrapper[4860]: I1211 08:28:19.352984 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-d5e8-account-create-update-76hlk" Dec 11 08:28:19 crc kubenswrapper[4860]: I1211 08:28:19.361953 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d449e0eb-9165-4f6b-b848-c0c6ea8bfa35-operator-scripts\") pod \"placement-db-create-x7qkx\" (UID: \"d449e0eb-9165-4f6b-b848-c0c6ea8bfa35\") " pod="openstack/placement-db-create-x7qkx" Dec 11 08:28:19 crc kubenswrapper[4860]: I1211 08:28:19.362331 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8hfzj\" (UniqueName: \"kubernetes.io/projected/d449e0eb-9165-4f6b-b848-c0c6ea8bfa35-kube-api-access-8hfzj\") pod \"placement-db-create-x7qkx\" (UID: \"d449e0eb-9165-4f6b-b848-c0c6ea8bfa35\") " pod="openstack/placement-db-create-x7qkx" Dec 11 08:28:19 crc kubenswrapper[4860]: I1211 08:28:19.399384 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-f52d-account-create-update-8rklk"] Dec 11 08:28:19 crc kubenswrapper[4860]: I1211 08:28:19.404331 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-f52d-account-create-update-8rklk" Dec 11 08:28:19 crc kubenswrapper[4860]: I1211 08:28:19.407956 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Dec 11 08:28:19 crc kubenswrapper[4860]: I1211 08:28:19.444564 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-f52d-account-create-update-8rklk"] Dec 11 08:28:19 crc kubenswrapper[4860]: I1211 08:28:19.465006 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d449e0eb-9165-4f6b-b848-c0c6ea8bfa35-operator-scripts\") pod \"placement-db-create-x7qkx\" (UID: \"d449e0eb-9165-4f6b-b848-c0c6ea8bfa35\") " pod="openstack/placement-db-create-x7qkx" Dec 11 08:28:19 crc kubenswrapper[4860]: I1211 08:28:19.465075 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8hfzj\" (UniqueName: \"kubernetes.io/projected/d449e0eb-9165-4f6b-b848-c0c6ea8bfa35-kube-api-access-8hfzj\") pod \"placement-db-create-x7qkx\" (UID: \"d449e0eb-9165-4f6b-b848-c0c6ea8bfa35\") " pod="openstack/placement-db-create-x7qkx" Dec 11 08:28:19 crc kubenswrapper[4860]: I1211 08:28:19.469441 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d449e0eb-9165-4f6b-b848-c0c6ea8bfa35-operator-scripts\") pod \"placement-db-create-x7qkx\" (UID: \"d449e0eb-9165-4f6b-b848-c0c6ea8bfa35\") " pod="openstack/placement-db-create-x7qkx" Dec 11 08:28:19 crc kubenswrapper[4860]: I1211 08:28:19.503456 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8hfzj\" (UniqueName: \"kubernetes.io/projected/d449e0eb-9165-4f6b-b848-c0c6ea8bfa35-kube-api-access-8hfzj\") pod \"placement-db-create-x7qkx\" (UID: \"d449e0eb-9165-4f6b-b848-c0c6ea8bfa35\") " pod="openstack/placement-db-create-x7qkx" Dec 11 08:28:19 crc kubenswrapper[4860]: I1211 08:28:19.573252 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/35b0da4d-26cd-4018-877b-5e8cbe0d4176-operator-scripts\") pod \"placement-f52d-account-create-update-8rklk\" (UID: \"35b0da4d-26cd-4018-877b-5e8cbe0d4176\") " pod="openstack/placement-f52d-account-create-update-8rklk" Dec 11 08:28:19 crc kubenswrapper[4860]: I1211 08:28:19.573781 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-67sr6\" (UniqueName: \"kubernetes.io/projected/35b0da4d-26cd-4018-877b-5e8cbe0d4176-kube-api-access-67sr6\") pod \"placement-f52d-account-create-update-8rklk\" (UID: \"35b0da4d-26cd-4018-877b-5e8cbe0d4176\") " pod="openstack/placement-f52d-account-create-update-8rklk" Dec 11 08:28:19 crc kubenswrapper[4860]: I1211 08:28:19.613020 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c4edc2e0-ea25-46f4-a127-b25e87244b28" path="/var/lib/kubelet/pods/c4edc2e0-ea25-46f4-a127-b25e87244b28/volumes" Dec 11 08:28:19 crc kubenswrapper[4860]: I1211 08:28:19.614232 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-84lgh"] Dec 11 08:28:19 crc kubenswrapper[4860]: I1211 08:28:19.615870 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-84lgh" Dec 11 08:28:19 crc kubenswrapper[4860]: I1211 08:28:19.616010 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-84lgh"] Dec 11 08:28:19 crc kubenswrapper[4860]: I1211 08:28:19.676061 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/35b0da4d-26cd-4018-877b-5e8cbe0d4176-operator-scripts\") pod \"placement-f52d-account-create-update-8rklk\" (UID: \"35b0da4d-26cd-4018-877b-5e8cbe0d4176\") " pod="openstack/placement-f52d-account-create-update-8rklk" Dec 11 08:28:19 crc kubenswrapper[4860]: I1211 08:28:19.676117 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-67sr6\" (UniqueName: \"kubernetes.io/projected/35b0da4d-26cd-4018-877b-5e8cbe0d4176-kube-api-access-67sr6\") pod \"placement-f52d-account-create-update-8rklk\" (UID: \"35b0da4d-26cd-4018-877b-5e8cbe0d4176\") " pod="openstack/placement-f52d-account-create-update-8rklk" Dec 11 08:28:19 crc kubenswrapper[4860]: I1211 08:28:19.676899 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/35b0da4d-26cd-4018-877b-5e8cbe0d4176-operator-scripts\") pod \"placement-f52d-account-create-update-8rklk\" (UID: \"35b0da4d-26cd-4018-877b-5e8cbe0d4176\") " pod="openstack/placement-f52d-account-create-update-8rklk" Dec 11 08:28:19 crc kubenswrapper[4860]: I1211 08:28:19.707221 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-67sr6\" (UniqueName: \"kubernetes.io/projected/35b0da4d-26cd-4018-877b-5e8cbe0d4176-kube-api-access-67sr6\") pod \"placement-f52d-account-create-update-8rklk\" (UID: \"35b0da4d-26cd-4018-877b-5e8cbe0d4176\") " pod="openstack/placement-f52d-account-create-update-8rklk" Dec 11 08:28:19 crc kubenswrapper[4860]: I1211 08:28:19.743260 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-86d4-account-create-update-khw2q"] Dec 11 08:28:19 crc kubenswrapper[4860]: I1211 08:28:19.746450 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-86d4-account-create-update-khw2q" Dec 11 08:28:19 crc kubenswrapper[4860]: I1211 08:28:19.748086 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"97634ec0-7953-4e38-a07e-95967ce874d7","Type":"ContainerStarted","Data":"c08aab93a48a1273f41e2191e9a12ab22ada37fed914adea9f44a9eb9eee90c9"} Dec 11 08:28:19 crc kubenswrapper[4860]: I1211 08:28:19.748984 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 11 08:28:19 crc kubenswrapper[4860]: I1211 08:28:19.752852 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Dec 11 08:28:19 crc kubenswrapper[4860]: I1211 08:28:19.754849 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-86d4-account-create-update-khw2q"] Dec 11 08:28:19 crc kubenswrapper[4860]: I1211 08:28:19.764360 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a672e32c-a924-4080-a7d1-bf7012b6725b","Type":"ContainerStarted","Data":"11d8c116ec8e8e9e4925c92ce1ee629a62ed8fb1b74b4280537f9e8ffe936427"} Dec 11 08:28:19 crc kubenswrapper[4860]: I1211 08:28:19.769578 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"a004cb7b-ee08-44c0-9dc7-d0b69509fe13","Type":"ContainerStarted","Data":"c1b416ccd0ce4787808ac0bc1fe677baedfb2f3a87f1c9829d5cf793665cbe0c"} Dec 11 08:28:19 crc kubenswrapper[4860]: I1211 08:28:19.770250 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:28:19 crc kubenswrapper[4860]: I1211 08:28:19.778712 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/60da3a16-5fba-47fc-8aef-7c7e942fa69b-operator-scripts\") pod \"glance-db-create-84lgh\" (UID: \"60da3a16-5fba-47fc-8aef-7c7e942fa69b\") " pod="openstack/glance-db-create-84lgh" Dec 11 08:28:19 crc kubenswrapper[4860]: I1211 08:28:19.778920 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zfld9\" (UniqueName: \"kubernetes.io/projected/60da3a16-5fba-47fc-8aef-7c7e942fa69b-kube-api-access-zfld9\") pod \"glance-db-create-84lgh\" (UID: \"60da3a16-5fba-47fc-8aef-7c7e942fa69b\") " pod="openstack/glance-db-create-84lgh" Dec 11 08:28:19 crc kubenswrapper[4860]: I1211 08:28:19.794263 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-x7qkx" Dec 11 08:28:19 crc kubenswrapper[4860]: I1211 08:28:19.802943 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-f52d-account-create-update-8rklk" Dec 11 08:28:19 crc kubenswrapper[4860]: I1211 08:28:19.840975 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=38.182579003 podStartE2EDuration="55.840948273s" podCreationTimestamp="2025-12-11 08:27:24 +0000 UTC" firstStartedPulling="2025-12-11 08:27:27.087880514 +0000 UTC m=+979.816399569" lastFinishedPulling="2025-12-11 08:27:44.746249784 +0000 UTC m=+997.474768839" observedRunningTime="2025-12-11 08:28:19.801528436 +0000 UTC m=+1032.530047491" watchObservedRunningTime="2025-12-11 08:28:19.840948273 +0000 UTC m=+1032.569467328" Dec 11 08:28:19 crc kubenswrapper[4860]: I1211 08:28:19.856283 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=38.211307612 podStartE2EDuration="55.856236082s" podCreationTimestamp="2025-12-11 08:27:24 +0000 UTC" firstStartedPulling="2025-12-11 08:27:27.069438595 +0000 UTC m=+979.797957650" lastFinishedPulling="2025-12-11 08:27:44.714367065 +0000 UTC m=+997.442886120" observedRunningTime="2025-12-11 08:28:19.83926986 +0000 UTC m=+1032.567788915" watchObservedRunningTime="2025-12-11 08:28:19.856236082 +0000 UTC m=+1032.584755137" Dec 11 08:28:19 crc kubenswrapper[4860]: I1211 08:28:19.889794 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-66fgz"] Dec 11 08:28:19 crc kubenswrapper[4860]: I1211 08:28:19.899540 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/60da3a16-5fba-47fc-8aef-7c7e942fa69b-operator-scripts\") pod \"glance-db-create-84lgh\" (UID: \"60da3a16-5fba-47fc-8aef-7c7e942fa69b\") " pod="openstack/glance-db-create-84lgh" Dec 11 08:28:19 crc kubenswrapper[4860]: I1211 08:28:19.899878 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zfld9\" (UniqueName: \"kubernetes.io/projected/60da3a16-5fba-47fc-8aef-7c7e942fa69b-kube-api-access-zfld9\") pod \"glance-db-create-84lgh\" (UID: \"60da3a16-5fba-47fc-8aef-7c7e942fa69b\") " pod="openstack/glance-db-create-84lgh" Dec 11 08:28:19 crc kubenswrapper[4860]: I1211 08:28:19.899949 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f765a5cb-6b7f-40ce-95fe-8c1abf5ce4d5-operator-scripts\") pod \"glance-86d4-account-create-update-khw2q\" (UID: \"f765a5cb-6b7f-40ce-95fe-8c1abf5ce4d5\") " pod="openstack/glance-86d4-account-create-update-khw2q" Dec 11 08:28:19 crc kubenswrapper[4860]: I1211 08:28:19.900001 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b85f4\" (UniqueName: \"kubernetes.io/projected/f765a5cb-6b7f-40ce-95fe-8c1abf5ce4d5-kube-api-access-b85f4\") pod \"glance-86d4-account-create-update-khw2q\" (UID: \"f765a5cb-6b7f-40ce-95fe-8c1abf5ce4d5\") " pod="openstack/glance-86d4-account-create-update-khw2q" Dec 11 08:28:19 crc kubenswrapper[4860]: I1211 08:28:19.901970 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/60da3a16-5fba-47fc-8aef-7c7e942fa69b-operator-scripts\") pod \"glance-db-create-84lgh\" (UID: \"60da3a16-5fba-47fc-8aef-7c7e942fa69b\") " pod="openstack/glance-db-create-84lgh" Dec 11 08:28:19 crc kubenswrapper[4860]: I1211 08:28:19.922305 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zfld9\" (UniqueName: \"kubernetes.io/projected/60da3a16-5fba-47fc-8aef-7c7e942fa69b-kube-api-access-zfld9\") pod \"glance-db-create-84lgh\" (UID: \"60da3a16-5fba-47fc-8aef-7c7e942fa69b\") " pod="openstack/glance-db-create-84lgh" Dec 11 08:28:19 crc kubenswrapper[4860]: I1211 08:28:19.956187 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-84lgh" Dec 11 08:28:19 crc kubenswrapper[4860]: I1211 08:28:19.996989 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-d5e8-account-create-update-76hlk"] Dec 11 08:28:20 crc kubenswrapper[4860]: I1211 08:28:20.001233 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f765a5cb-6b7f-40ce-95fe-8c1abf5ce4d5-operator-scripts\") pod \"glance-86d4-account-create-update-khw2q\" (UID: \"f765a5cb-6b7f-40ce-95fe-8c1abf5ce4d5\") " pod="openstack/glance-86d4-account-create-update-khw2q" Dec 11 08:28:20 crc kubenswrapper[4860]: I1211 08:28:20.001288 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b85f4\" (UniqueName: \"kubernetes.io/projected/f765a5cb-6b7f-40ce-95fe-8c1abf5ce4d5-kube-api-access-b85f4\") pod \"glance-86d4-account-create-update-khw2q\" (UID: \"f765a5cb-6b7f-40ce-95fe-8c1abf5ce4d5\") " pod="openstack/glance-86d4-account-create-update-khw2q" Dec 11 08:28:20 crc kubenswrapper[4860]: I1211 08:28:20.002363 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f765a5cb-6b7f-40ce-95fe-8c1abf5ce4d5-operator-scripts\") pod \"glance-86d4-account-create-update-khw2q\" (UID: \"f765a5cb-6b7f-40ce-95fe-8c1abf5ce4d5\") " pod="openstack/glance-86d4-account-create-update-khw2q" Dec 11 08:28:20 crc kubenswrapper[4860]: I1211 08:28:20.029857 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b85f4\" (UniqueName: \"kubernetes.io/projected/f765a5cb-6b7f-40ce-95fe-8c1abf5ce4d5-kube-api-access-b85f4\") pod \"glance-86d4-account-create-update-khw2q\" (UID: \"f765a5cb-6b7f-40ce-95fe-8c1abf5ce4d5\") " pod="openstack/glance-86d4-account-create-update-khw2q" Dec 11 08:28:20 crc kubenswrapper[4860]: I1211 08:28:20.072036 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-86d4-account-create-update-khw2q" Dec 11 08:28:20 crc kubenswrapper[4860]: I1211 08:28:20.222979 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-f52d-account-create-update-8rklk"] Dec 11 08:28:20 crc kubenswrapper[4860]: W1211 08:28:20.255716 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod35b0da4d_26cd_4018_877b_5e8cbe0d4176.slice/crio-b671bf7eac98990890652922f8e231922fe7de341d64cc7d0f19aa7f6a5e412c WatchSource:0}: Error finding container b671bf7eac98990890652922f8e231922fe7de341d64cc7d0f19aa7f6a5e412c: Status 404 returned error can't find the container with id b671bf7eac98990890652922f8e231922fe7de341d64cc7d0f19aa7f6a5e412c Dec 11 08:28:20 crc kubenswrapper[4860]: I1211 08:28:20.265030 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-bphmx" Dec 11 08:28:20 crc kubenswrapper[4860]: I1211 08:28:20.417915 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/79bcab38-b87b-44ff-ba2b-6027fd20719d-dispersionconf\") pod \"79bcab38-b87b-44ff-ba2b-6027fd20719d\" (UID: \"79bcab38-b87b-44ff-ba2b-6027fd20719d\") " Dec 11 08:28:20 crc kubenswrapper[4860]: I1211 08:28:20.418018 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/79bcab38-b87b-44ff-ba2b-6027fd20719d-swiftconf\") pod \"79bcab38-b87b-44ff-ba2b-6027fd20719d\" (UID: \"79bcab38-b87b-44ff-ba2b-6027fd20719d\") " Dec 11 08:28:20 crc kubenswrapper[4860]: I1211 08:28:20.418084 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gjh4n\" (UniqueName: \"kubernetes.io/projected/79bcab38-b87b-44ff-ba2b-6027fd20719d-kube-api-access-gjh4n\") pod \"79bcab38-b87b-44ff-ba2b-6027fd20719d\" (UID: \"79bcab38-b87b-44ff-ba2b-6027fd20719d\") " Dec 11 08:28:20 crc kubenswrapper[4860]: I1211 08:28:20.418243 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/79bcab38-b87b-44ff-ba2b-6027fd20719d-etc-swift\") pod \"79bcab38-b87b-44ff-ba2b-6027fd20719d\" (UID: \"79bcab38-b87b-44ff-ba2b-6027fd20719d\") " Dec 11 08:28:20 crc kubenswrapper[4860]: I1211 08:28:20.418324 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/79bcab38-b87b-44ff-ba2b-6027fd20719d-ring-data-devices\") pod \"79bcab38-b87b-44ff-ba2b-6027fd20719d\" (UID: \"79bcab38-b87b-44ff-ba2b-6027fd20719d\") " Dec 11 08:28:20 crc kubenswrapper[4860]: I1211 08:28:20.418371 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79bcab38-b87b-44ff-ba2b-6027fd20719d-combined-ca-bundle\") pod \"79bcab38-b87b-44ff-ba2b-6027fd20719d\" (UID: \"79bcab38-b87b-44ff-ba2b-6027fd20719d\") " Dec 11 08:28:20 crc kubenswrapper[4860]: I1211 08:28:20.418398 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/79bcab38-b87b-44ff-ba2b-6027fd20719d-scripts\") pod \"79bcab38-b87b-44ff-ba2b-6027fd20719d\" (UID: \"79bcab38-b87b-44ff-ba2b-6027fd20719d\") " Dec 11 08:28:20 crc kubenswrapper[4860]: I1211 08:28:20.421397 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/79bcab38-b87b-44ff-ba2b-6027fd20719d-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "79bcab38-b87b-44ff-ba2b-6027fd20719d" (UID: "79bcab38-b87b-44ff-ba2b-6027fd20719d"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:28:20 crc kubenswrapper[4860]: I1211 08:28:20.422155 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/79bcab38-b87b-44ff-ba2b-6027fd20719d-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "79bcab38-b87b-44ff-ba2b-6027fd20719d" (UID: "79bcab38-b87b-44ff-ba2b-6027fd20719d"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:28:20 crc kubenswrapper[4860]: I1211 08:28:20.429831 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/79bcab38-b87b-44ff-ba2b-6027fd20719d-kube-api-access-gjh4n" (OuterVolumeSpecName: "kube-api-access-gjh4n") pod "79bcab38-b87b-44ff-ba2b-6027fd20719d" (UID: "79bcab38-b87b-44ff-ba2b-6027fd20719d"). InnerVolumeSpecName "kube-api-access-gjh4n". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:28:20 crc kubenswrapper[4860]: I1211 08:28:20.433289 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79bcab38-b87b-44ff-ba2b-6027fd20719d-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "79bcab38-b87b-44ff-ba2b-6027fd20719d" (UID: "79bcab38-b87b-44ff-ba2b-6027fd20719d"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:28:20 crc kubenswrapper[4860]: I1211 08:28:20.455603 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79bcab38-b87b-44ff-ba2b-6027fd20719d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "79bcab38-b87b-44ff-ba2b-6027fd20719d" (UID: "79bcab38-b87b-44ff-ba2b-6027fd20719d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:28:20 crc kubenswrapper[4860]: I1211 08:28:20.463979 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/79bcab38-b87b-44ff-ba2b-6027fd20719d-scripts" (OuterVolumeSpecName: "scripts") pod "79bcab38-b87b-44ff-ba2b-6027fd20719d" (UID: "79bcab38-b87b-44ff-ba2b-6027fd20719d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:28:20 crc kubenswrapper[4860]: I1211 08:28:20.474252 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79bcab38-b87b-44ff-ba2b-6027fd20719d-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "79bcab38-b87b-44ff-ba2b-6027fd20719d" (UID: "79bcab38-b87b-44ff-ba2b-6027fd20719d"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:28:20 crc kubenswrapper[4860]: I1211 08:28:20.515827 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-x7qkx"] Dec 11 08:28:20 crc kubenswrapper[4860]: I1211 08:28:20.520535 4860 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/79bcab38-b87b-44ff-ba2b-6027fd20719d-ring-data-devices\") on node \"crc\" DevicePath \"\"" Dec 11 08:28:20 crc kubenswrapper[4860]: I1211 08:28:20.520578 4860 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79bcab38-b87b-44ff-ba2b-6027fd20719d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 08:28:20 crc kubenswrapper[4860]: I1211 08:28:20.520589 4860 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/79bcab38-b87b-44ff-ba2b-6027fd20719d-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 08:28:20 crc kubenswrapper[4860]: I1211 08:28:20.520602 4860 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/79bcab38-b87b-44ff-ba2b-6027fd20719d-dispersionconf\") on node \"crc\" DevicePath \"\"" Dec 11 08:28:20 crc kubenswrapper[4860]: I1211 08:28:20.520613 4860 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/79bcab38-b87b-44ff-ba2b-6027fd20719d-swiftconf\") on node \"crc\" DevicePath \"\"" Dec 11 08:28:20 crc kubenswrapper[4860]: I1211 08:28:20.520623 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gjh4n\" (UniqueName: \"kubernetes.io/projected/79bcab38-b87b-44ff-ba2b-6027fd20719d-kube-api-access-gjh4n\") on node \"crc\" DevicePath \"\"" Dec 11 08:28:20 crc kubenswrapper[4860]: I1211 08:28:20.520633 4860 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/79bcab38-b87b-44ff-ba2b-6027fd20719d-etc-swift\") on node \"crc\" DevicePath \"\"" Dec 11 08:28:20 crc kubenswrapper[4860]: W1211 08:28:20.521323 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd449e0eb_9165_4f6b_b848_c0c6ea8bfa35.slice/crio-088e8eef8bd79e92fbcdd5d5fd254e667f400bdba2d3b98422e7a2bee5f3b93f WatchSource:0}: Error finding container 088e8eef8bd79e92fbcdd5d5fd254e667f400bdba2d3b98422e7a2bee5f3b93f: Status 404 returned error can't find the container with id 088e8eef8bd79e92fbcdd5d5fd254e667f400bdba2d3b98422e7a2bee5f3b93f Dec 11 08:28:20 crc kubenswrapper[4860]: I1211 08:28:20.654527 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-84lgh"] Dec 11 08:28:20 crc kubenswrapper[4860]: I1211 08:28:20.762909 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-86d4-account-create-update-khw2q"] Dec 11 08:28:20 crc kubenswrapper[4860]: I1211 08:28:20.788134 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-bphmx" Dec 11 08:28:20 crc kubenswrapper[4860]: I1211 08:28:20.788134 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-bphmx" event={"ID":"79bcab38-b87b-44ff-ba2b-6027fd20719d","Type":"ContainerDied","Data":"060cb1eff63ca3df9aae672b635a61c937f3ef8b11a707f07a7f3715f31ed27f"} Dec 11 08:28:20 crc kubenswrapper[4860]: I1211 08:28:20.788278 4860 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="060cb1eff63ca3df9aae672b635a61c937f3ef8b11a707f07a7f3715f31ed27f" Dec 11 08:28:20 crc kubenswrapper[4860]: I1211 08:28:20.791034 4860 generic.go:334] "Generic (PLEG): container finished" podID="35b0da4d-26cd-4018-877b-5e8cbe0d4176" containerID="a1dffde6d5b34e4277a3235833c9cfa23befa0d0541a1ecc3976f3b8de6a843f" exitCode=0 Dec 11 08:28:20 crc kubenswrapper[4860]: I1211 08:28:20.791127 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-f52d-account-create-update-8rklk" event={"ID":"35b0da4d-26cd-4018-877b-5e8cbe0d4176","Type":"ContainerDied","Data":"a1dffde6d5b34e4277a3235833c9cfa23befa0d0541a1ecc3976f3b8de6a843f"} Dec 11 08:28:20 crc kubenswrapper[4860]: I1211 08:28:20.791162 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-f52d-account-create-update-8rklk" event={"ID":"35b0da4d-26cd-4018-877b-5e8cbe0d4176","Type":"ContainerStarted","Data":"b671bf7eac98990890652922f8e231922fe7de341d64cc7d0f19aa7f6a5e412c"} Dec 11 08:28:20 crc kubenswrapper[4860]: I1211 08:28:20.794490 4860 generic.go:334] "Generic (PLEG): container finished" podID="25bead29-f8d6-4539-97f8-9c3be58fbf64" containerID="deafcd4e90a9c8eb4c53882a697b7855a4976e819e86959545ccf4411f6e5a3a" exitCode=0 Dec 11 08:28:20 crc kubenswrapper[4860]: I1211 08:28:20.794576 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-d5e8-account-create-update-76hlk" event={"ID":"25bead29-f8d6-4539-97f8-9c3be58fbf64","Type":"ContainerDied","Data":"deafcd4e90a9c8eb4c53882a697b7855a4976e819e86959545ccf4411f6e5a3a"} Dec 11 08:28:20 crc kubenswrapper[4860]: I1211 08:28:20.794607 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-d5e8-account-create-update-76hlk" event={"ID":"25bead29-f8d6-4539-97f8-9c3be58fbf64","Type":"ContainerStarted","Data":"8ce0cbd1eb5e34808c0daf6caea904563b72724313711427c7cb0ed94497aebd"} Dec 11 08:28:20 crc kubenswrapper[4860]: I1211 08:28:20.796029 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-x7qkx" event={"ID":"d449e0eb-9165-4f6b-b848-c0c6ea8bfa35","Type":"ContainerStarted","Data":"088e8eef8bd79e92fbcdd5d5fd254e667f400bdba2d3b98422e7a2bee5f3b93f"} Dec 11 08:28:20 crc kubenswrapper[4860]: I1211 08:28:20.800122 4860 generic.go:334] "Generic (PLEG): container finished" podID="79aae41e-c682-42c8-8bbc-5d060804d535" containerID="c54a5f6501d64c77e977792e254bfbe24fadb74783de561e2d747ad3ad7140dd" exitCode=0 Dec 11 08:28:20 crc kubenswrapper[4860]: I1211 08:28:20.800183 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-66fgz" event={"ID":"79aae41e-c682-42c8-8bbc-5d060804d535","Type":"ContainerDied","Data":"c54a5f6501d64c77e977792e254bfbe24fadb74783de561e2d747ad3ad7140dd"} Dec 11 08:28:20 crc kubenswrapper[4860]: I1211 08:28:20.800220 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-66fgz" event={"ID":"79aae41e-c682-42c8-8bbc-5d060804d535","Type":"ContainerStarted","Data":"55b3d6ae3a8e717b022d5fc0f69aa71e7cab9fc2c921d91d291665212b84b535"} Dec 11 08:28:21 crc kubenswrapper[4860]: W1211 08:28:21.083412 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod60da3a16_5fba_47fc_8aef_7c7e942fa69b.slice/crio-4778bd90c046c291cee02b37e7999f27ed91c02a22fb2e9d442f448583b5265b WatchSource:0}: Error finding container 4778bd90c046c291cee02b37e7999f27ed91c02a22fb2e9d442f448583b5265b: Status 404 returned error can't find the container with id 4778bd90c046c291cee02b37e7999f27ed91c02a22fb2e9d442f448583b5265b Dec 11 08:28:21 crc kubenswrapper[4860]: W1211 08:28:21.084387 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf765a5cb_6b7f_40ce_95fe_8c1abf5ce4d5.slice/crio-d6d5240cebaef46decb8120f827c8a26d1630ee1073dcc51cef7816a79444d71 WatchSource:0}: Error finding container d6d5240cebaef46decb8120f827c8a26d1630ee1073dcc51cef7816a79444d71: Status 404 returned error can't find the container with id d6d5240cebaef46decb8120f827c8a26d1630ee1073dcc51cef7816a79444d71 Dec 11 08:28:21 crc kubenswrapper[4860]: I1211 08:28:21.811874 4860 generic.go:334] "Generic (PLEG): container finished" podID="d449e0eb-9165-4f6b-b848-c0c6ea8bfa35" containerID="f1438a46a2a9690d43847281c607ef6c47f38f8e05dfbaabe34adf6a253bfafe" exitCode=0 Dec 11 08:28:21 crc kubenswrapper[4860]: I1211 08:28:21.812302 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-x7qkx" event={"ID":"d449e0eb-9165-4f6b-b848-c0c6ea8bfa35","Type":"ContainerDied","Data":"f1438a46a2a9690d43847281c607ef6c47f38f8e05dfbaabe34adf6a253bfafe"} Dec 11 08:28:21 crc kubenswrapper[4860]: I1211 08:28:21.814343 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-86d4-account-create-update-khw2q" event={"ID":"f765a5cb-6b7f-40ce-95fe-8c1abf5ce4d5","Type":"ContainerStarted","Data":"739ec4e15f43896e65eeb1f0e6bbf9b0b64946cd2a3a5547c6a8c39b795bf61d"} Dec 11 08:28:21 crc kubenswrapper[4860]: I1211 08:28:21.814376 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-86d4-account-create-update-khw2q" event={"ID":"f765a5cb-6b7f-40ce-95fe-8c1abf5ce4d5","Type":"ContainerStarted","Data":"d6d5240cebaef46decb8120f827c8a26d1630ee1073dcc51cef7816a79444d71"} Dec 11 08:28:21 crc kubenswrapper[4860]: I1211 08:28:21.817362 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-84lgh" event={"ID":"60da3a16-5fba-47fc-8aef-7c7e942fa69b","Type":"ContainerStarted","Data":"44765f35f50e19467b780e5977831f691d7c724b607c8daec431551703717522"} Dec 11 08:28:21 crc kubenswrapper[4860]: I1211 08:28:21.817402 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-84lgh" event={"ID":"60da3a16-5fba-47fc-8aef-7c7e942fa69b","Type":"ContainerStarted","Data":"4778bd90c046c291cee02b37e7999f27ed91c02a22fb2e9d442f448583b5265b"} Dec 11 08:28:21 crc kubenswrapper[4860]: I1211 08:28:21.833126 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a672e32c-a924-4080-a7d1-bf7012b6725b","Type":"ContainerStarted","Data":"33361c6d36555fa9b9ffcd63b16ce3d4898f93a9fa8e1d4c74bddedd6217b8a4"} Dec 11 08:28:21 crc kubenswrapper[4860]: I1211 08:28:21.833198 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a672e32c-a924-4080-a7d1-bf7012b6725b","Type":"ContainerStarted","Data":"628f5e1a358b89d8215748386531685ad66eadde7e6a66753720200561eb4038"} Dec 11 08:28:21 crc kubenswrapper[4860]: I1211 08:28:21.864816 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-create-84lgh" podStartSLOduration=2.864796513 podStartE2EDuration="2.864796513s" podCreationTimestamp="2025-12-11 08:28:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:28:21.860335072 +0000 UTC m=+1034.588854137" watchObservedRunningTime="2025-12-11 08:28:21.864796513 +0000 UTC m=+1034.593315568" Dec 11 08:28:22 crc kubenswrapper[4860]: I1211 08:28:22.480768 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-66fgz" Dec 11 08:28:22 crc kubenswrapper[4860]: I1211 08:28:22.486009 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-f52d-account-create-update-8rklk" Dec 11 08:28:22 crc kubenswrapper[4860]: I1211 08:28:22.490374 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-d5e8-account-create-update-76hlk" Dec 11 08:28:22 crc kubenswrapper[4860]: I1211 08:28:22.499405 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-86d4-account-create-update-khw2q" podStartSLOduration=3.499382987 podStartE2EDuration="3.499382987s" podCreationTimestamp="2025-12-11 08:28:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:28:21.878367658 +0000 UTC m=+1034.606886713" watchObservedRunningTime="2025-12-11 08:28:22.499382987 +0000 UTC m=+1035.227902042" Dec 11 08:28:22 crc kubenswrapper[4860]: I1211 08:28:22.576145 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-67sr6\" (UniqueName: \"kubernetes.io/projected/35b0da4d-26cd-4018-877b-5e8cbe0d4176-kube-api-access-67sr6\") pod \"35b0da4d-26cd-4018-877b-5e8cbe0d4176\" (UID: \"35b0da4d-26cd-4018-877b-5e8cbe0d4176\") " Dec 11 08:28:22 crc kubenswrapper[4860]: I1211 08:28:22.576244 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/79aae41e-c682-42c8-8bbc-5d060804d535-operator-scripts\") pod \"79aae41e-c682-42c8-8bbc-5d060804d535\" (UID: \"79aae41e-c682-42c8-8bbc-5d060804d535\") " Dec 11 08:28:22 crc kubenswrapper[4860]: I1211 08:28:22.576272 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8jmpz\" (UniqueName: \"kubernetes.io/projected/79aae41e-c682-42c8-8bbc-5d060804d535-kube-api-access-8jmpz\") pod \"79aae41e-c682-42c8-8bbc-5d060804d535\" (UID: \"79aae41e-c682-42c8-8bbc-5d060804d535\") " Dec 11 08:28:22 crc kubenswrapper[4860]: I1211 08:28:22.576328 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dkg2w\" (UniqueName: \"kubernetes.io/projected/25bead29-f8d6-4539-97f8-9c3be58fbf64-kube-api-access-dkg2w\") pod \"25bead29-f8d6-4539-97f8-9c3be58fbf64\" (UID: \"25bead29-f8d6-4539-97f8-9c3be58fbf64\") " Dec 11 08:28:22 crc kubenswrapper[4860]: I1211 08:28:22.576428 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/25bead29-f8d6-4539-97f8-9c3be58fbf64-operator-scripts\") pod \"25bead29-f8d6-4539-97f8-9c3be58fbf64\" (UID: \"25bead29-f8d6-4539-97f8-9c3be58fbf64\") " Dec 11 08:28:22 crc kubenswrapper[4860]: I1211 08:28:22.576501 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/35b0da4d-26cd-4018-877b-5e8cbe0d4176-operator-scripts\") pod \"35b0da4d-26cd-4018-877b-5e8cbe0d4176\" (UID: \"35b0da4d-26cd-4018-877b-5e8cbe0d4176\") " Dec 11 08:28:22 crc kubenswrapper[4860]: I1211 08:28:22.577021 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/79aae41e-c682-42c8-8bbc-5d060804d535-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "79aae41e-c682-42c8-8bbc-5d060804d535" (UID: "79aae41e-c682-42c8-8bbc-5d060804d535"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:28:22 crc kubenswrapper[4860]: I1211 08:28:22.577252 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/35b0da4d-26cd-4018-877b-5e8cbe0d4176-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "35b0da4d-26cd-4018-877b-5e8cbe0d4176" (UID: "35b0da4d-26cd-4018-877b-5e8cbe0d4176"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:28:22 crc kubenswrapper[4860]: I1211 08:28:22.577612 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25bead29-f8d6-4539-97f8-9c3be58fbf64-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "25bead29-f8d6-4539-97f8-9c3be58fbf64" (UID: "25bead29-f8d6-4539-97f8-9c3be58fbf64"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:28:22 crc kubenswrapper[4860]: I1211 08:28:22.583563 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/79aae41e-c682-42c8-8bbc-5d060804d535-kube-api-access-8jmpz" (OuterVolumeSpecName: "kube-api-access-8jmpz") pod "79aae41e-c682-42c8-8bbc-5d060804d535" (UID: "79aae41e-c682-42c8-8bbc-5d060804d535"). InnerVolumeSpecName "kube-api-access-8jmpz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:28:22 crc kubenswrapper[4860]: I1211 08:28:22.583620 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25bead29-f8d6-4539-97f8-9c3be58fbf64-kube-api-access-dkg2w" (OuterVolumeSpecName: "kube-api-access-dkg2w") pod "25bead29-f8d6-4539-97f8-9c3be58fbf64" (UID: "25bead29-f8d6-4539-97f8-9c3be58fbf64"). InnerVolumeSpecName "kube-api-access-dkg2w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:28:22 crc kubenswrapper[4860]: I1211 08:28:22.583765 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35b0da4d-26cd-4018-877b-5e8cbe0d4176-kube-api-access-67sr6" (OuterVolumeSpecName: "kube-api-access-67sr6") pod "35b0da4d-26cd-4018-877b-5e8cbe0d4176" (UID: "35b0da4d-26cd-4018-877b-5e8cbe0d4176"). InnerVolumeSpecName "kube-api-access-67sr6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:28:22 crc kubenswrapper[4860]: I1211 08:28:22.678127 4860 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/79aae41e-c682-42c8-8bbc-5d060804d535-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 08:28:22 crc kubenswrapper[4860]: I1211 08:28:22.678181 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8jmpz\" (UniqueName: \"kubernetes.io/projected/79aae41e-c682-42c8-8bbc-5d060804d535-kube-api-access-8jmpz\") on node \"crc\" DevicePath \"\"" Dec 11 08:28:22 crc kubenswrapper[4860]: I1211 08:28:22.678194 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dkg2w\" (UniqueName: \"kubernetes.io/projected/25bead29-f8d6-4539-97f8-9c3be58fbf64-kube-api-access-dkg2w\") on node \"crc\" DevicePath \"\"" Dec 11 08:28:22 crc kubenswrapper[4860]: I1211 08:28:22.678203 4860 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/25bead29-f8d6-4539-97f8-9c3be58fbf64-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 08:28:22 crc kubenswrapper[4860]: I1211 08:28:22.678212 4860 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/35b0da4d-26cd-4018-877b-5e8cbe0d4176-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 08:28:22 crc kubenswrapper[4860]: I1211 08:28:22.678221 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-67sr6\" (UniqueName: \"kubernetes.io/projected/35b0da4d-26cd-4018-877b-5e8cbe0d4176-kube-api-access-67sr6\") on node \"crc\" DevicePath \"\"" Dec 11 08:28:22 crc kubenswrapper[4860]: I1211 08:28:22.842323 4860 generic.go:334] "Generic (PLEG): container finished" podID="f765a5cb-6b7f-40ce-95fe-8c1abf5ce4d5" containerID="739ec4e15f43896e65eeb1f0e6bbf9b0b64946cd2a3a5547c6a8c39b795bf61d" exitCode=0 Dec 11 08:28:22 crc kubenswrapper[4860]: I1211 08:28:22.842378 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-86d4-account-create-update-khw2q" event={"ID":"f765a5cb-6b7f-40ce-95fe-8c1abf5ce4d5","Type":"ContainerDied","Data":"739ec4e15f43896e65eeb1f0e6bbf9b0b64946cd2a3a5547c6a8c39b795bf61d"} Dec 11 08:28:22 crc kubenswrapper[4860]: I1211 08:28:22.844356 4860 generic.go:334] "Generic (PLEG): container finished" podID="60da3a16-5fba-47fc-8aef-7c7e942fa69b" containerID="44765f35f50e19467b780e5977831f691d7c724b607c8daec431551703717522" exitCode=0 Dec 11 08:28:22 crc kubenswrapper[4860]: I1211 08:28:22.844443 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-84lgh" event={"ID":"60da3a16-5fba-47fc-8aef-7c7e942fa69b","Type":"ContainerDied","Data":"44765f35f50e19467b780e5977831f691d7c724b607c8daec431551703717522"} Dec 11 08:28:22 crc kubenswrapper[4860]: I1211 08:28:22.845747 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-66fgz" event={"ID":"79aae41e-c682-42c8-8bbc-5d060804d535","Type":"ContainerDied","Data":"55b3d6ae3a8e717b022d5fc0f69aa71e7cab9fc2c921d91d291665212b84b535"} Dec 11 08:28:22 crc kubenswrapper[4860]: I1211 08:28:22.845792 4860 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="55b3d6ae3a8e717b022d5fc0f69aa71e7cab9fc2c921d91d291665212b84b535" Dec 11 08:28:22 crc kubenswrapper[4860]: I1211 08:28:22.845842 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-66fgz" Dec 11 08:28:22 crc kubenswrapper[4860]: I1211 08:28:22.851910 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a672e32c-a924-4080-a7d1-bf7012b6725b","Type":"ContainerStarted","Data":"ef4cfc17d91c6dceae084780cd4f4bedc94dbcb05b541959a3e95d3c896cb116"} Dec 11 08:28:22 crc kubenswrapper[4860]: I1211 08:28:22.851953 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a672e32c-a924-4080-a7d1-bf7012b6725b","Type":"ContainerStarted","Data":"cec8e16eeb2818ee1567a9e1ab5106481b13e66f8c7b60ca0d8e2585d725eb3b"} Dec 11 08:28:22 crc kubenswrapper[4860]: I1211 08:28:22.853218 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-f52d-account-create-update-8rklk" event={"ID":"35b0da4d-26cd-4018-877b-5e8cbe0d4176","Type":"ContainerDied","Data":"b671bf7eac98990890652922f8e231922fe7de341d64cc7d0f19aa7f6a5e412c"} Dec 11 08:28:22 crc kubenswrapper[4860]: I1211 08:28:22.853247 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-f52d-account-create-update-8rklk" Dec 11 08:28:22 crc kubenswrapper[4860]: I1211 08:28:22.853250 4860 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b671bf7eac98990890652922f8e231922fe7de341d64cc7d0f19aa7f6a5e412c" Dec 11 08:28:22 crc kubenswrapper[4860]: I1211 08:28:22.856709 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-d5e8-account-create-update-76hlk" Dec 11 08:28:22 crc kubenswrapper[4860]: I1211 08:28:22.856691 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-d5e8-account-create-update-76hlk" event={"ID":"25bead29-f8d6-4539-97f8-9c3be58fbf64","Type":"ContainerDied","Data":"8ce0cbd1eb5e34808c0daf6caea904563b72724313711427c7cb0ed94497aebd"} Dec 11 08:28:22 crc kubenswrapper[4860]: I1211 08:28:22.856756 4860 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8ce0cbd1eb5e34808c0daf6caea904563b72724313711427c7cb0ed94497aebd" Dec 11 08:28:23 crc kubenswrapper[4860]: I1211 08:28:23.226635 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-x7qkx" Dec 11 08:28:23 crc kubenswrapper[4860]: I1211 08:28:23.405120 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d449e0eb-9165-4f6b-b848-c0c6ea8bfa35-operator-scripts\") pod \"d449e0eb-9165-4f6b-b848-c0c6ea8bfa35\" (UID: \"d449e0eb-9165-4f6b-b848-c0c6ea8bfa35\") " Dec 11 08:28:23 crc kubenswrapper[4860]: I1211 08:28:23.405359 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8hfzj\" (UniqueName: \"kubernetes.io/projected/d449e0eb-9165-4f6b-b848-c0c6ea8bfa35-kube-api-access-8hfzj\") pod \"d449e0eb-9165-4f6b-b848-c0c6ea8bfa35\" (UID: \"d449e0eb-9165-4f6b-b848-c0c6ea8bfa35\") " Dec 11 08:28:23 crc kubenswrapper[4860]: I1211 08:28:23.407554 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d449e0eb-9165-4f6b-b848-c0c6ea8bfa35-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d449e0eb-9165-4f6b-b848-c0c6ea8bfa35" (UID: "d449e0eb-9165-4f6b-b848-c0c6ea8bfa35"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:28:23 crc kubenswrapper[4860]: I1211 08:28:23.413983 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d449e0eb-9165-4f6b-b848-c0c6ea8bfa35-kube-api-access-8hfzj" (OuterVolumeSpecName: "kube-api-access-8hfzj") pod "d449e0eb-9165-4f6b-b848-c0c6ea8bfa35" (UID: "d449e0eb-9165-4f6b-b848-c0c6ea8bfa35"). InnerVolumeSpecName "kube-api-access-8hfzj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:28:23 crc kubenswrapper[4860]: I1211 08:28:23.507069 4860 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d449e0eb-9165-4f6b-b848-c0c6ea8bfa35-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 08:28:23 crc kubenswrapper[4860]: I1211 08:28:23.507108 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8hfzj\" (UniqueName: \"kubernetes.io/projected/d449e0eb-9165-4f6b-b848-c0c6ea8bfa35-kube-api-access-8hfzj\") on node \"crc\" DevicePath \"\"" Dec 11 08:28:23 crc kubenswrapper[4860]: I1211 08:28:23.870457 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-x7qkx" event={"ID":"d449e0eb-9165-4f6b-b848-c0c6ea8bfa35","Type":"ContainerDied","Data":"088e8eef8bd79e92fbcdd5d5fd254e667f400bdba2d3b98422e7a2bee5f3b93f"} Dec 11 08:28:23 crc kubenswrapper[4860]: I1211 08:28:23.870829 4860 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="088e8eef8bd79e92fbcdd5d5fd254e667f400bdba2d3b98422e7a2bee5f3b93f" Dec 11 08:28:23 crc kubenswrapper[4860]: I1211 08:28:23.870503 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-x7qkx" Dec 11 08:28:23 crc kubenswrapper[4860]: I1211 08:28:23.883401 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a672e32c-a924-4080-a7d1-bf7012b6725b","Type":"ContainerStarted","Data":"dc9821d4824a4be13e65f2edd07efe5b4f5b8297d639aeb69b445082c3a94e44"} Dec 11 08:28:23 crc kubenswrapper[4860]: I1211 08:28:23.883477 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a672e32c-a924-4080-a7d1-bf7012b6725b","Type":"ContainerStarted","Data":"3696b631fc9231c176921ae84a22e0e077492ec93dd7174a4dc97c23f74b17cc"} Dec 11 08:28:24 crc kubenswrapper[4860]: I1211 08:28:24.334225 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-84lgh" Dec 11 08:28:24 crc kubenswrapper[4860]: I1211 08:28:24.340729 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-86d4-account-create-update-khw2q" Dec 11 08:28:24 crc kubenswrapper[4860]: I1211 08:28:24.422788 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zfld9\" (UniqueName: \"kubernetes.io/projected/60da3a16-5fba-47fc-8aef-7c7e942fa69b-kube-api-access-zfld9\") pod \"60da3a16-5fba-47fc-8aef-7c7e942fa69b\" (UID: \"60da3a16-5fba-47fc-8aef-7c7e942fa69b\") " Dec 11 08:28:24 crc kubenswrapper[4860]: I1211 08:28:24.422919 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/60da3a16-5fba-47fc-8aef-7c7e942fa69b-operator-scripts\") pod \"60da3a16-5fba-47fc-8aef-7c7e942fa69b\" (UID: \"60da3a16-5fba-47fc-8aef-7c7e942fa69b\") " Dec 11 08:28:24 crc kubenswrapper[4860]: I1211 08:28:24.423022 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f765a5cb-6b7f-40ce-95fe-8c1abf5ce4d5-operator-scripts\") pod \"f765a5cb-6b7f-40ce-95fe-8c1abf5ce4d5\" (UID: \"f765a5cb-6b7f-40ce-95fe-8c1abf5ce4d5\") " Dec 11 08:28:24 crc kubenswrapper[4860]: I1211 08:28:24.423066 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b85f4\" (UniqueName: \"kubernetes.io/projected/f765a5cb-6b7f-40ce-95fe-8c1abf5ce4d5-kube-api-access-b85f4\") pod \"f765a5cb-6b7f-40ce-95fe-8c1abf5ce4d5\" (UID: \"f765a5cb-6b7f-40ce-95fe-8c1abf5ce4d5\") " Dec 11 08:28:24 crc kubenswrapper[4860]: I1211 08:28:24.423937 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/60da3a16-5fba-47fc-8aef-7c7e942fa69b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "60da3a16-5fba-47fc-8aef-7c7e942fa69b" (UID: "60da3a16-5fba-47fc-8aef-7c7e942fa69b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:28:24 crc kubenswrapper[4860]: I1211 08:28:24.424015 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f765a5cb-6b7f-40ce-95fe-8c1abf5ce4d5-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f765a5cb-6b7f-40ce-95fe-8c1abf5ce4d5" (UID: "f765a5cb-6b7f-40ce-95fe-8c1abf5ce4d5"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:28:24 crc kubenswrapper[4860]: I1211 08:28:24.433897 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/60da3a16-5fba-47fc-8aef-7c7e942fa69b-kube-api-access-zfld9" (OuterVolumeSpecName: "kube-api-access-zfld9") pod "60da3a16-5fba-47fc-8aef-7c7e942fa69b" (UID: "60da3a16-5fba-47fc-8aef-7c7e942fa69b"). InnerVolumeSpecName "kube-api-access-zfld9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:28:24 crc kubenswrapper[4860]: I1211 08:28:24.435355 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f765a5cb-6b7f-40ce-95fe-8c1abf5ce4d5-kube-api-access-b85f4" (OuterVolumeSpecName: "kube-api-access-b85f4") pod "f765a5cb-6b7f-40ce-95fe-8c1abf5ce4d5" (UID: "f765a5cb-6b7f-40ce-95fe-8c1abf5ce4d5"). InnerVolumeSpecName "kube-api-access-b85f4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:28:24 crc kubenswrapper[4860]: I1211 08:28:24.445832 4860 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-dbdtz" podUID="e834572f-9650-45d3-b978-c39141b44b74" containerName="ovn-controller" probeResult="failure" output=< Dec 11 08:28:24 crc kubenswrapper[4860]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 11 08:28:24 crc kubenswrapper[4860]: > Dec 11 08:28:24 crc kubenswrapper[4860]: I1211 08:28:24.525394 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zfld9\" (UniqueName: \"kubernetes.io/projected/60da3a16-5fba-47fc-8aef-7c7e942fa69b-kube-api-access-zfld9\") on node \"crc\" DevicePath \"\"" Dec 11 08:28:24 crc kubenswrapper[4860]: I1211 08:28:24.525455 4860 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/60da3a16-5fba-47fc-8aef-7c7e942fa69b-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 08:28:24 crc kubenswrapper[4860]: I1211 08:28:24.525471 4860 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f765a5cb-6b7f-40ce-95fe-8c1abf5ce4d5-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 08:28:24 crc kubenswrapper[4860]: I1211 08:28:24.525483 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b85f4\" (UniqueName: \"kubernetes.io/projected/f765a5cb-6b7f-40ce-95fe-8c1abf5ce4d5-kube-api-access-b85f4\") on node \"crc\" DevicePath \"\"" Dec 11 08:28:24 crc kubenswrapper[4860]: I1211 08:28:24.661325 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Dec 11 08:28:24 crc kubenswrapper[4860]: I1211 08:28:24.930403 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-86d4-account-create-update-khw2q" Dec 11 08:28:24 crc kubenswrapper[4860]: I1211 08:28:24.931796 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-86d4-account-create-update-khw2q" event={"ID":"f765a5cb-6b7f-40ce-95fe-8c1abf5ce4d5","Type":"ContainerDied","Data":"d6d5240cebaef46decb8120f827c8a26d1630ee1073dcc51cef7816a79444d71"} Dec 11 08:28:24 crc kubenswrapper[4860]: I1211 08:28:24.931836 4860 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d6d5240cebaef46decb8120f827c8a26d1630ee1073dcc51cef7816a79444d71" Dec 11 08:28:24 crc kubenswrapper[4860]: I1211 08:28:24.935020 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-84lgh" event={"ID":"60da3a16-5fba-47fc-8aef-7c7e942fa69b","Type":"ContainerDied","Data":"4778bd90c046c291cee02b37e7999f27ed91c02a22fb2e9d442f448583b5265b"} Dec 11 08:28:24 crc kubenswrapper[4860]: I1211 08:28:24.935098 4860 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4778bd90c046c291cee02b37e7999f27ed91c02a22fb2e9d442f448583b5265b" Dec 11 08:28:24 crc kubenswrapper[4860]: I1211 08:28:24.935250 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-84lgh" Dec 11 08:28:24 crc kubenswrapper[4860]: I1211 08:28:24.951816 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a672e32c-a924-4080-a7d1-bf7012b6725b","Type":"ContainerStarted","Data":"c3af1616746e309807aa2b3eb890312df703b4a8a51356265f7e5267380014c6"} Dec 11 08:28:24 crc kubenswrapper[4860]: I1211 08:28:24.951906 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a672e32c-a924-4080-a7d1-bf7012b6725b","Type":"ContainerStarted","Data":"c3264a921bca421184cda226d366a5bfd3a16f49c3c46edfcb350929bd2299a3"} Dec 11 08:28:25 crc kubenswrapper[4860]: I1211 08:28:25.969194 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a672e32c-a924-4080-a7d1-bf7012b6725b","Type":"ContainerStarted","Data":"e1a9e83f4a8a2905a6cb2a9af8998619eb7cf3f277835f92940b18e0c930aa5c"} Dec 11 08:28:26 crc kubenswrapper[4860]: I1211 08:28:26.986355 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a672e32c-a924-4080-a7d1-bf7012b6725b","Type":"ContainerStarted","Data":"e22b3389599fd4aeff5751a53e0b8cc9c4c52148e2119845e53901100a3e3d5f"} Dec 11 08:28:26 crc kubenswrapper[4860]: I1211 08:28:26.986773 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a672e32c-a924-4080-a7d1-bf7012b6725b","Type":"ContainerStarted","Data":"53f2f85be00b6ec905a2e69d045084ea396b754895f7f64db182e90f420f46e5"} Dec 11 08:28:29 crc kubenswrapper[4860]: I1211 08:28:29.415917 4860 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-dbdtz" podUID="e834572f-9650-45d3-b978-c39141b44b74" containerName="ovn-controller" probeResult="failure" output=< Dec 11 08:28:29 crc kubenswrapper[4860]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 11 08:28:29 crc kubenswrapper[4860]: > Dec 11 08:28:29 crc kubenswrapper[4860]: I1211 08:28:29.472009 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-q86ch" Dec 11 08:28:29 crc kubenswrapper[4860]: I1211 08:28:29.477461 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-q86ch" Dec 11 08:28:29 crc kubenswrapper[4860]: I1211 08:28:29.724720 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-dbdtz-config-svfgb"] Dec 11 08:28:29 crc kubenswrapper[4860]: E1211 08:28:29.725427 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25bead29-f8d6-4539-97f8-9c3be58fbf64" containerName="mariadb-account-create-update" Dec 11 08:28:29 crc kubenswrapper[4860]: I1211 08:28:29.725448 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="25bead29-f8d6-4539-97f8-9c3be58fbf64" containerName="mariadb-account-create-update" Dec 11 08:28:29 crc kubenswrapper[4860]: E1211 08:28:29.725476 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d449e0eb-9165-4f6b-b848-c0c6ea8bfa35" containerName="mariadb-database-create" Dec 11 08:28:29 crc kubenswrapper[4860]: I1211 08:28:29.725484 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="d449e0eb-9165-4f6b-b848-c0c6ea8bfa35" containerName="mariadb-database-create" Dec 11 08:28:29 crc kubenswrapper[4860]: E1211 08:28:29.725505 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f765a5cb-6b7f-40ce-95fe-8c1abf5ce4d5" containerName="mariadb-account-create-update" Dec 11 08:28:29 crc kubenswrapper[4860]: I1211 08:28:29.725514 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="f765a5cb-6b7f-40ce-95fe-8c1abf5ce4d5" containerName="mariadb-account-create-update" Dec 11 08:28:29 crc kubenswrapper[4860]: E1211 08:28:29.725528 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79aae41e-c682-42c8-8bbc-5d060804d535" containerName="mariadb-database-create" Dec 11 08:28:29 crc kubenswrapper[4860]: I1211 08:28:29.725535 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="79aae41e-c682-42c8-8bbc-5d060804d535" containerName="mariadb-database-create" Dec 11 08:28:29 crc kubenswrapper[4860]: E1211 08:28:29.725548 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35b0da4d-26cd-4018-877b-5e8cbe0d4176" containerName="mariadb-account-create-update" Dec 11 08:28:29 crc kubenswrapper[4860]: I1211 08:28:29.725555 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="35b0da4d-26cd-4018-877b-5e8cbe0d4176" containerName="mariadb-account-create-update" Dec 11 08:28:29 crc kubenswrapper[4860]: E1211 08:28:29.725570 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60da3a16-5fba-47fc-8aef-7c7e942fa69b" containerName="mariadb-database-create" Dec 11 08:28:29 crc kubenswrapper[4860]: I1211 08:28:29.725577 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="60da3a16-5fba-47fc-8aef-7c7e942fa69b" containerName="mariadb-database-create" Dec 11 08:28:29 crc kubenswrapper[4860]: E1211 08:28:29.725590 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79bcab38-b87b-44ff-ba2b-6027fd20719d" containerName="swift-ring-rebalance" Dec 11 08:28:29 crc kubenswrapper[4860]: I1211 08:28:29.725599 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="79bcab38-b87b-44ff-ba2b-6027fd20719d" containerName="swift-ring-rebalance" Dec 11 08:28:29 crc kubenswrapper[4860]: I1211 08:28:29.725988 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="79aae41e-c682-42c8-8bbc-5d060804d535" containerName="mariadb-database-create" Dec 11 08:28:29 crc kubenswrapper[4860]: I1211 08:28:29.726018 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="79bcab38-b87b-44ff-ba2b-6027fd20719d" containerName="swift-ring-rebalance" Dec 11 08:28:29 crc kubenswrapper[4860]: I1211 08:28:29.726036 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="f765a5cb-6b7f-40ce-95fe-8c1abf5ce4d5" containerName="mariadb-account-create-update" Dec 11 08:28:29 crc kubenswrapper[4860]: I1211 08:28:29.726049 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="d449e0eb-9165-4f6b-b848-c0c6ea8bfa35" containerName="mariadb-database-create" Dec 11 08:28:29 crc kubenswrapper[4860]: I1211 08:28:29.726064 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="35b0da4d-26cd-4018-877b-5e8cbe0d4176" containerName="mariadb-account-create-update" Dec 11 08:28:29 crc kubenswrapper[4860]: I1211 08:28:29.726081 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="60da3a16-5fba-47fc-8aef-7c7e942fa69b" containerName="mariadb-database-create" Dec 11 08:28:29 crc kubenswrapper[4860]: I1211 08:28:29.726096 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="25bead29-f8d6-4539-97f8-9c3be58fbf64" containerName="mariadb-account-create-update" Dec 11 08:28:29 crc kubenswrapper[4860]: I1211 08:28:29.726670 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-dbdtz-config-svfgb" Dec 11 08:28:29 crc kubenswrapper[4860]: I1211 08:28:29.729337 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Dec 11 08:28:29 crc kubenswrapper[4860]: I1211 08:28:29.741602 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-dbdtz-config-svfgb"] Dec 11 08:28:29 crc kubenswrapper[4860]: I1211 08:28:29.839428 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/daa404d6-76e8-4568-9785-bc7743817b18-var-run\") pod \"ovn-controller-dbdtz-config-svfgb\" (UID: \"daa404d6-76e8-4568-9785-bc7743817b18\") " pod="openstack/ovn-controller-dbdtz-config-svfgb" Dec 11 08:28:29 crc kubenswrapper[4860]: I1211 08:28:29.839497 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2wrwl\" (UniqueName: \"kubernetes.io/projected/daa404d6-76e8-4568-9785-bc7743817b18-kube-api-access-2wrwl\") pod \"ovn-controller-dbdtz-config-svfgb\" (UID: \"daa404d6-76e8-4568-9785-bc7743817b18\") " pod="openstack/ovn-controller-dbdtz-config-svfgb" Dec 11 08:28:29 crc kubenswrapper[4860]: I1211 08:28:29.839561 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/daa404d6-76e8-4568-9785-bc7743817b18-var-log-ovn\") pod \"ovn-controller-dbdtz-config-svfgb\" (UID: \"daa404d6-76e8-4568-9785-bc7743817b18\") " pod="openstack/ovn-controller-dbdtz-config-svfgb" Dec 11 08:28:29 crc kubenswrapper[4860]: I1211 08:28:29.839580 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/daa404d6-76e8-4568-9785-bc7743817b18-var-run-ovn\") pod \"ovn-controller-dbdtz-config-svfgb\" (UID: \"daa404d6-76e8-4568-9785-bc7743817b18\") " pod="openstack/ovn-controller-dbdtz-config-svfgb" Dec 11 08:28:29 crc kubenswrapper[4860]: I1211 08:28:29.839621 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/daa404d6-76e8-4568-9785-bc7743817b18-additional-scripts\") pod \"ovn-controller-dbdtz-config-svfgb\" (UID: \"daa404d6-76e8-4568-9785-bc7743817b18\") " pod="openstack/ovn-controller-dbdtz-config-svfgb" Dec 11 08:28:29 crc kubenswrapper[4860]: I1211 08:28:29.839671 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/daa404d6-76e8-4568-9785-bc7743817b18-scripts\") pod \"ovn-controller-dbdtz-config-svfgb\" (UID: \"daa404d6-76e8-4568-9785-bc7743817b18\") " pod="openstack/ovn-controller-dbdtz-config-svfgb" Dec 11 08:28:29 crc kubenswrapper[4860]: I1211 08:28:29.941495 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/daa404d6-76e8-4568-9785-bc7743817b18-var-run-ovn\") pod \"ovn-controller-dbdtz-config-svfgb\" (UID: \"daa404d6-76e8-4568-9785-bc7743817b18\") " pod="openstack/ovn-controller-dbdtz-config-svfgb" Dec 11 08:28:29 crc kubenswrapper[4860]: I1211 08:28:29.941610 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/daa404d6-76e8-4568-9785-bc7743817b18-additional-scripts\") pod \"ovn-controller-dbdtz-config-svfgb\" (UID: \"daa404d6-76e8-4568-9785-bc7743817b18\") " pod="openstack/ovn-controller-dbdtz-config-svfgb" Dec 11 08:28:29 crc kubenswrapper[4860]: I1211 08:28:29.941684 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/daa404d6-76e8-4568-9785-bc7743817b18-scripts\") pod \"ovn-controller-dbdtz-config-svfgb\" (UID: \"daa404d6-76e8-4568-9785-bc7743817b18\") " pod="openstack/ovn-controller-dbdtz-config-svfgb" Dec 11 08:28:29 crc kubenswrapper[4860]: I1211 08:28:29.941734 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/daa404d6-76e8-4568-9785-bc7743817b18-var-run\") pod \"ovn-controller-dbdtz-config-svfgb\" (UID: \"daa404d6-76e8-4568-9785-bc7743817b18\") " pod="openstack/ovn-controller-dbdtz-config-svfgb" Dec 11 08:28:29 crc kubenswrapper[4860]: I1211 08:28:29.941791 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2wrwl\" (UniqueName: \"kubernetes.io/projected/daa404d6-76e8-4568-9785-bc7743817b18-kube-api-access-2wrwl\") pod \"ovn-controller-dbdtz-config-svfgb\" (UID: \"daa404d6-76e8-4568-9785-bc7743817b18\") " pod="openstack/ovn-controller-dbdtz-config-svfgb" Dec 11 08:28:29 crc kubenswrapper[4860]: I1211 08:28:29.941867 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/daa404d6-76e8-4568-9785-bc7743817b18-var-log-ovn\") pod \"ovn-controller-dbdtz-config-svfgb\" (UID: \"daa404d6-76e8-4568-9785-bc7743817b18\") " pod="openstack/ovn-controller-dbdtz-config-svfgb" Dec 11 08:28:29 crc kubenswrapper[4860]: I1211 08:28:29.941867 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/daa404d6-76e8-4568-9785-bc7743817b18-var-run-ovn\") pod \"ovn-controller-dbdtz-config-svfgb\" (UID: \"daa404d6-76e8-4568-9785-bc7743817b18\") " pod="openstack/ovn-controller-dbdtz-config-svfgb" Dec 11 08:28:29 crc kubenswrapper[4860]: I1211 08:28:29.941891 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/daa404d6-76e8-4568-9785-bc7743817b18-var-run\") pod \"ovn-controller-dbdtz-config-svfgb\" (UID: \"daa404d6-76e8-4568-9785-bc7743817b18\") " pod="openstack/ovn-controller-dbdtz-config-svfgb" Dec 11 08:28:29 crc kubenswrapper[4860]: I1211 08:28:29.941970 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/daa404d6-76e8-4568-9785-bc7743817b18-var-log-ovn\") pod \"ovn-controller-dbdtz-config-svfgb\" (UID: \"daa404d6-76e8-4568-9785-bc7743817b18\") " pod="openstack/ovn-controller-dbdtz-config-svfgb" Dec 11 08:28:29 crc kubenswrapper[4860]: I1211 08:28:29.942594 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/daa404d6-76e8-4568-9785-bc7743817b18-additional-scripts\") pod \"ovn-controller-dbdtz-config-svfgb\" (UID: \"daa404d6-76e8-4568-9785-bc7743817b18\") " pod="openstack/ovn-controller-dbdtz-config-svfgb" Dec 11 08:28:29 crc kubenswrapper[4860]: I1211 08:28:29.944835 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/daa404d6-76e8-4568-9785-bc7743817b18-scripts\") pod \"ovn-controller-dbdtz-config-svfgb\" (UID: \"daa404d6-76e8-4568-9785-bc7743817b18\") " pod="openstack/ovn-controller-dbdtz-config-svfgb" Dec 11 08:28:29 crc kubenswrapper[4860]: I1211 08:28:29.962016 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2wrwl\" (UniqueName: \"kubernetes.io/projected/daa404d6-76e8-4568-9785-bc7743817b18-kube-api-access-2wrwl\") pod \"ovn-controller-dbdtz-config-svfgb\" (UID: \"daa404d6-76e8-4568-9785-bc7743817b18\") " pod="openstack/ovn-controller-dbdtz-config-svfgb" Dec 11 08:28:30 crc kubenswrapper[4860]: I1211 08:28:30.025674 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a672e32c-a924-4080-a7d1-bf7012b6725b","Type":"ContainerStarted","Data":"47d31344cc783a81a4dbf14d8fe8c240712c564ea45a7e7c29f5d756636c0346"} Dec 11 08:28:30 crc kubenswrapper[4860]: I1211 08:28:30.025743 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a672e32c-a924-4080-a7d1-bf7012b6725b","Type":"ContainerStarted","Data":"827566689c961cde038f390d1277b8d6b158728d678de19e0698de6822c3c7b4"} Dec 11 08:28:30 crc kubenswrapper[4860]: I1211 08:28:30.025764 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a672e32c-a924-4080-a7d1-bf7012b6725b","Type":"ContainerStarted","Data":"2c6ed60993ebaac30370d4231aae9f0de2d954942a529c20894071e7cf8b45c1"} Dec 11 08:28:30 crc kubenswrapper[4860]: I1211 08:28:30.025781 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a672e32c-a924-4080-a7d1-bf7012b6725b","Type":"ContainerStarted","Data":"4dcefbf79416580d714d143645bd61cd80ddc57d781ff4eabff99a1988c7392d"} Dec 11 08:28:30 crc kubenswrapper[4860]: I1211 08:28:30.038832 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-4fsqr"] Dec 11 08:28:30 crc kubenswrapper[4860]: I1211 08:28:30.040895 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-4fsqr" Dec 11 08:28:30 crc kubenswrapper[4860]: I1211 08:28:30.047986 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Dec 11 08:28:30 crc kubenswrapper[4860]: I1211 08:28:30.048430 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-9qb2r" Dec 11 08:28:30 crc kubenswrapper[4860]: I1211 08:28:30.058839 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-dbdtz-config-svfgb" Dec 11 08:28:30 crc kubenswrapper[4860]: I1211 08:28:30.072810 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-4fsqr"] Dec 11 08:28:30 crc kubenswrapper[4860]: I1211 08:28:30.080224 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=22.224415652 podStartE2EDuration="29.080199171s" podCreationTimestamp="2025-12-11 08:28:01 +0000 UTC" firstStartedPulling="2025-12-11 08:28:18.843055123 +0000 UTC m=+1031.571574188" lastFinishedPulling="2025-12-11 08:28:25.698838652 +0000 UTC m=+1038.427357707" observedRunningTime="2025-12-11 08:28:30.072871581 +0000 UTC m=+1042.801390646" watchObservedRunningTime="2025-12-11 08:28:30.080199171 +0000 UTC m=+1042.808718236" Dec 11 08:28:30 crc kubenswrapper[4860]: I1211 08:28:30.147353 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/cf433d5d-8c95-4d72-864c-3b74f315dd2b-db-sync-config-data\") pod \"glance-db-sync-4fsqr\" (UID: \"cf433d5d-8c95-4d72-864c-3b74f315dd2b\") " pod="openstack/glance-db-sync-4fsqr" Dec 11 08:28:30 crc kubenswrapper[4860]: I1211 08:28:30.147401 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cmvl7\" (UniqueName: \"kubernetes.io/projected/cf433d5d-8c95-4d72-864c-3b74f315dd2b-kube-api-access-cmvl7\") pod \"glance-db-sync-4fsqr\" (UID: \"cf433d5d-8c95-4d72-864c-3b74f315dd2b\") " pod="openstack/glance-db-sync-4fsqr" Dec 11 08:28:30 crc kubenswrapper[4860]: I1211 08:28:30.147440 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf433d5d-8c95-4d72-864c-3b74f315dd2b-config-data\") pod \"glance-db-sync-4fsqr\" (UID: \"cf433d5d-8c95-4d72-864c-3b74f315dd2b\") " pod="openstack/glance-db-sync-4fsqr" Dec 11 08:28:30 crc kubenswrapper[4860]: I1211 08:28:30.147484 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf433d5d-8c95-4d72-864c-3b74f315dd2b-combined-ca-bundle\") pod \"glance-db-sync-4fsqr\" (UID: \"cf433d5d-8c95-4d72-864c-3b74f315dd2b\") " pod="openstack/glance-db-sync-4fsqr" Dec 11 08:28:30 crc kubenswrapper[4860]: I1211 08:28:30.251961 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/cf433d5d-8c95-4d72-864c-3b74f315dd2b-db-sync-config-data\") pod \"glance-db-sync-4fsqr\" (UID: \"cf433d5d-8c95-4d72-864c-3b74f315dd2b\") " pod="openstack/glance-db-sync-4fsqr" Dec 11 08:28:30 crc kubenswrapper[4860]: I1211 08:28:30.252347 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cmvl7\" (UniqueName: \"kubernetes.io/projected/cf433d5d-8c95-4d72-864c-3b74f315dd2b-kube-api-access-cmvl7\") pod \"glance-db-sync-4fsqr\" (UID: \"cf433d5d-8c95-4d72-864c-3b74f315dd2b\") " pod="openstack/glance-db-sync-4fsqr" Dec 11 08:28:30 crc kubenswrapper[4860]: I1211 08:28:30.252406 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf433d5d-8c95-4d72-864c-3b74f315dd2b-config-data\") pod \"glance-db-sync-4fsqr\" (UID: \"cf433d5d-8c95-4d72-864c-3b74f315dd2b\") " pod="openstack/glance-db-sync-4fsqr" Dec 11 08:28:30 crc kubenswrapper[4860]: I1211 08:28:30.252479 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf433d5d-8c95-4d72-864c-3b74f315dd2b-combined-ca-bundle\") pod \"glance-db-sync-4fsqr\" (UID: \"cf433d5d-8c95-4d72-864c-3b74f315dd2b\") " pod="openstack/glance-db-sync-4fsqr" Dec 11 08:28:30 crc kubenswrapper[4860]: I1211 08:28:30.258844 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf433d5d-8c95-4d72-864c-3b74f315dd2b-combined-ca-bundle\") pod \"glance-db-sync-4fsqr\" (UID: \"cf433d5d-8c95-4d72-864c-3b74f315dd2b\") " pod="openstack/glance-db-sync-4fsqr" Dec 11 08:28:30 crc kubenswrapper[4860]: I1211 08:28:30.259007 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf433d5d-8c95-4d72-864c-3b74f315dd2b-config-data\") pod \"glance-db-sync-4fsqr\" (UID: \"cf433d5d-8c95-4d72-864c-3b74f315dd2b\") " pod="openstack/glance-db-sync-4fsqr" Dec 11 08:28:30 crc kubenswrapper[4860]: I1211 08:28:30.259399 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/cf433d5d-8c95-4d72-864c-3b74f315dd2b-db-sync-config-data\") pod \"glance-db-sync-4fsqr\" (UID: \"cf433d5d-8c95-4d72-864c-3b74f315dd2b\") " pod="openstack/glance-db-sync-4fsqr" Dec 11 08:28:30 crc kubenswrapper[4860]: I1211 08:28:30.272766 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cmvl7\" (UniqueName: \"kubernetes.io/projected/cf433d5d-8c95-4d72-864c-3b74f315dd2b-kube-api-access-cmvl7\") pod \"glance-db-sync-4fsqr\" (UID: \"cf433d5d-8c95-4d72-864c-3b74f315dd2b\") " pod="openstack/glance-db-sync-4fsqr" Dec 11 08:28:30 crc kubenswrapper[4860]: I1211 08:28:30.360324 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-4fsqr" Dec 11 08:28:30 crc kubenswrapper[4860]: I1211 08:28:30.388318 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6d5b6d6b67-pxxrf"] Dec 11 08:28:30 crc kubenswrapper[4860]: I1211 08:28:30.389983 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d5b6d6b67-pxxrf" Dec 11 08:28:30 crc kubenswrapper[4860]: I1211 08:28:30.399896 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Dec 11 08:28:30 crc kubenswrapper[4860]: I1211 08:28:30.411677 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d5b6d6b67-pxxrf"] Dec 11 08:28:30 crc kubenswrapper[4860]: I1211 08:28:30.425474 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-dbdtz-config-svfgb"] Dec 11 08:28:30 crc kubenswrapper[4860]: I1211 08:28:30.459397 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d59cd3a6-bd2d-4c74-97d0-265b3fc56dfd-ovsdbserver-sb\") pod \"dnsmasq-dns-6d5b6d6b67-pxxrf\" (UID: \"d59cd3a6-bd2d-4c74-97d0-265b3fc56dfd\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-pxxrf" Dec 11 08:28:30 crc kubenswrapper[4860]: I1211 08:28:30.459485 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d59cd3a6-bd2d-4c74-97d0-265b3fc56dfd-ovsdbserver-nb\") pod \"dnsmasq-dns-6d5b6d6b67-pxxrf\" (UID: \"d59cd3a6-bd2d-4c74-97d0-265b3fc56dfd\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-pxxrf" Dec 11 08:28:30 crc kubenswrapper[4860]: I1211 08:28:30.459609 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d59cd3a6-bd2d-4c74-97d0-265b3fc56dfd-dns-svc\") pod \"dnsmasq-dns-6d5b6d6b67-pxxrf\" (UID: \"d59cd3a6-bd2d-4c74-97d0-265b3fc56dfd\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-pxxrf" Dec 11 08:28:30 crc kubenswrapper[4860]: I1211 08:28:30.459670 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d59cd3a6-bd2d-4c74-97d0-265b3fc56dfd-dns-swift-storage-0\") pod \"dnsmasq-dns-6d5b6d6b67-pxxrf\" (UID: \"d59cd3a6-bd2d-4c74-97d0-265b3fc56dfd\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-pxxrf" Dec 11 08:28:30 crc kubenswrapper[4860]: I1211 08:28:30.459868 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zz27s\" (UniqueName: \"kubernetes.io/projected/d59cd3a6-bd2d-4c74-97d0-265b3fc56dfd-kube-api-access-zz27s\") pod \"dnsmasq-dns-6d5b6d6b67-pxxrf\" (UID: \"d59cd3a6-bd2d-4c74-97d0-265b3fc56dfd\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-pxxrf" Dec 11 08:28:30 crc kubenswrapper[4860]: I1211 08:28:30.459938 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d59cd3a6-bd2d-4c74-97d0-265b3fc56dfd-config\") pod \"dnsmasq-dns-6d5b6d6b67-pxxrf\" (UID: \"d59cd3a6-bd2d-4c74-97d0-265b3fc56dfd\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-pxxrf" Dec 11 08:28:30 crc kubenswrapper[4860]: I1211 08:28:30.562136 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d59cd3a6-bd2d-4c74-97d0-265b3fc56dfd-ovsdbserver-nb\") pod \"dnsmasq-dns-6d5b6d6b67-pxxrf\" (UID: \"d59cd3a6-bd2d-4c74-97d0-265b3fc56dfd\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-pxxrf" Dec 11 08:28:30 crc kubenswrapper[4860]: I1211 08:28:30.562452 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d59cd3a6-bd2d-4c74-97d0-265b3fc56dfd-dns-svc\") pod \"dnsmasq-dns-6d5b6d6b67-pxxrf\" (UID: \"d59cd3a6-bd2d-4c74-97d0-265b3fc56dfd\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-pxxrf" Dec 11 08:28:30 crc kubenswrapper[4860]: I1211 08:28:30.562477 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d59cd3a6-bd2d-4c74-97d0-265b3fc56dfd-dns-swift-storage-0\") pod \"dnsmasq-dns-6d5b6d6b67-pxxrf\" (UID: \"d59cd3a6-bd2d-4c74-97d0-265b3fc56dfd\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-pxxrf" Dec 11 08:28:30 crc kubenswrapper[4860]: I1211 08:28:30.562529 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zz27s\" (UniqueName: \"kubernetes.io/projected/d59cd3a6-bd2d-4c74-97d0-265b3fc56dfd-kube-api-access-zz27s\") pod \"dnsmasq-dns-6d5b6d6b67-pxxrf\" (UID: \"d59cd3a6-bd2d-4c74-97d0-265b3fc56dfd\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-pxxrf" Dec 11 08:28:30 crc kubenswrapper[4860]: I1211 08:28:30.562555 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d59cd3a6-bd2d-4c74-97d0-265b3fc56dfd-config\") pod \"dnsmasq-dns-6d5b6d6b67-pxxrf\" (UID: \"d59cd3a6-bd2d-4c74-97d0-265b3fc56dfd\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-pxxrf" Dec 11 08:28:30 crc kubenswrapper[4860]: I1211 08:28:30.562619 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d59cd3a6-bd2d-4c74-97d0-265b3fc56dfd-ovsdbserver-sb\") pod \"dnsmasq-dns-6d5b6d6b67-pxxrf\" (UID: \"d59cd3a6-bd2d-4c74-97d0-265b3fc56dfd\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-pxxrf" Dec 11 08:28:30 crc kubenswrapper[4860]: I1211 08:28:30.563249 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d59cd3a6-bd2d-4c74-97d0-265b3fc56dfd-ovsdbserver-nb\") pod \"dnsmasq-dns-6d5b6d6b67-pxxrf\" (UID: \"d59cd3a6-bd2d-4c74-97d0-265b3fc56dfd\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-pxxrf" Dec 11 08:28:30 crc kubenswrapper[4860]: I1211 08:28:30.563315 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d59cd3a6-bd2d-4c74-97d0-265b3fc56dfd-dns-swift-storage-0\") pod \"dnsmasq-dns-6d5b6d6b67-pxxrf\" (UID: \"d59cd3a6-bd2d-4c74-97d0-265b3fc56dfd\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-pxxrf" Dec 11 08:28:30 crc kubenswrapper[4860]: I1211 08:28:30.563492 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d59cd3a6-bd2d-4c74-97d0-265b3fc56dfd-config\") pod \"dnsmasq-dns-6d5b6d6b67-pxxrf\" (UID: \"d59cd3a6-bd2d-4c74-97d0-265b3fc56dfd\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-pxxrf" Dec 11 08:28:30 crc kubenswrapper[4860]: I1211 08:28:30.563529 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d59cd3a6-bd2d-4c74-97d0-265b3fc56dfd-dns-svc\") pod \"dnsmasq-dns-6d5b6d6b67-pxxrf\" (UID: \"d59cd3a6-bd2d-4c74-97d0-265b3fc56dfd\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-pxxrf" Dec 11 08:28:30 crc kubenswrapper[4860]: I1211 08:28:30.566557 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d59cd3a6-bd2d-4c74-97d0-265b3fc56dfd-ovsdbserver-sb\") pod \"dnsmasq-dns-6d5b6d6b67-pxxrf\" (UID: \"d59cd3a6-bd2d-4c74-97d0-265b3fc56dfd\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-pxxrf" Dec 11 08:28:30 crc kubenswrapper[4860]: I1211 08:28:30.605410 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zz27s\" (UniqueName: \"kubernetes.io/projected/d59cd3a6-bd2d-4c74-97d0-265b3fc56dfd-kube-api-access-zz27s\") pod \"dnsmasq-dns-6d5b6d6b67-pxxrf\" (UID: \"d59cd3a6-bd2d-4c74-97d0-265b3fc56dfd\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-pxxrf" Dec 11 08:28:30 crc kubenswrapper[4860]: I1211 08:28:30.846720 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d5b6d6b67-pxxrf" Dec 11 08:28:30 crc kubenswrapper[4860]: I1211 08:28:30.966883 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-4fsqr"] Dec 11 08:28:31 crc kubenswrapper[4860]: I1211 08:28:31.058428 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-4fsqr" event={"ID":"cf433d5d-8c95-4d72-864c-3b74f315dd2b","Type":"ContainerStarted","Data":"3fd59619fd28c03de0bc2164837abea1e39a9f1b0fcb738f470203d59015276b"} Dec 11 08:28:31 crc kubenswrapper[4860]: I1211 08:28:31.068675 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-dbdtz-config-svfgb" event={"ID":"daa404d6-76e8-4568-9785-bc7743817b18","Type":"ContainerStarted","Data":"ab9f83a10a88c0bcbaba9d0b040e40de344ecc46ca54bb9cd5a1f2984b476452"} Dec 11 08:28:31 crc kubenswrapper[4860]: I1211 08:28:31.069836 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-dbdtz-config-svfgb" event={"ID":"daa404d6-76e8-4568-9785-bc7743817b18","Type":"ContainerStarted","Data":"042e4fb7b0069346e101c32fa6bb315d245f50b3fb78ca9f60fe3db1d59936c2"} Dec 11 08:28:31 crc kubenswrapper[4860]: I1211 08:28:31.099848 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-dbdtz-config-svfgb" podStartSLOduration=2.099819399 podStartE2EDuration="2.099819399s" podCreationTimestamp="2025-12-11 08:28:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:28:31.08394074 +0000 UTC m=+1043.812459815" watchObservedRunningTime="2025-12-11 08:28:31.099819399 +0000 UTC m=+1043.828338454" Dec 11 08:28:31 crc kubenswrapper[4860]: I1211 08:28:31.360902 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d5b6d6b67-pxxrf"] Dec 11 08:28:32 crc kubenswrapper[4860]: I1211 08:28:32.077601 4860 generic.go:334] "Generic (PLEG): container finished" podID="d59cd3a6-bd2d-4c74-97d0-265b3fc56dfd" containerID="4baa691e265690619f74abc8c059724e292805c92021c7c8964d08fdb9c12671" exitCode=0 Dec 11 08:28:32 crc kubenswrapper[4860]: I1211 08:28:32.077702 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d5b6d6b67-pxxrf" event={"ID":"d59cd3a6-bd2d-4c74-97d0-265b3fc56dfd","Type":"ContainerDied","Data":"4baa691e265690619f74abc8c059724e292805c92021c7c8964d08fdb9c12671"} Dec 11 08:28:32 crc kubenswrapper[4860]: I1211 08:28:32.077731 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d5b6d6b67-pxxrf" event={"ID":"d59cd3a6-bd2d-4c74-97d0-265b3fc56dfd","Type":"ContainerStarted","Data":"a3afb983675c4e177f06e0451a9e72be824cb1f7e1200cae5923a1122ff43b00"} Dec 11 08:28:32 crc kubenswrapper[4860]: I1211 08:28:32.098900 4860 generic.go:334] "Generic (PLEG): container finished" podID="daa404d6-76e8-4568-9785-bc7743817b18" containerID="ab9f83a10a88c0bcbaba9d0b040e40de344ecc46ca54bb9cd5a1f2984b476452" exitCode=0 Dec 11 08:28:32 crc kubenswrapper[4860]: I1211 08:28:32.098951 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-dbdtz-config-svfgb" event={"ID":"daa404d6-76e8-4568-9785-bc7743817b18","Type":"ContainerDied","Data":"ab9f83a10a88c0bcbaba9d0b040e40de344ecc46ca54bb9cd5a1f2984b476452"} Dec 11 08:28:33 crc kubenswrapper[4860]: I1211 08:28:33.115611 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d5b6d6b67-pxxrf" event={"ID":"d59cd3a6-bd2d-4c74-97d0-265b3fc56dfd","Type":"ContainerStarted","Data":"8458e72beef31344067e7052e8cb6285a95889bd51a203b4e3d4a97d4c4e0d2c"} Dec 11 08:28:33 crc kubenswrapper[4860]: I1211 08:28:33.115755 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6d5b6d6b67-pxxrf" Dec 11 08:28:33 crc kubenswrapper[4860]: I1211 08:28:33.146303 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6d5b6d6b67-pxxrf" podStartSLOduration=3.146283791 podStartE2EDuration="3.146283791s" podCreationTimestamp="2025-12-11 08:28:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:28:33.138383763 +0000 UTC m=+1045.866902818" watchObservedRunningTime="2025-12-11 08:28:33.146283791 +0000 UTC m=+1045.874802846" Dec 11 08:28:33 crc kubenswrapper[4860]: I1211 08:28:33.493088 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-dbdtz-config-svfgb" Dec 11 08:28:33 crc kubenswrapper[4860]: I1211 08:28:33.519782 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/daa404d6-76e8-4568-9785-bc7743817b18-var-run\") pod \"daa404d6-76e8-4568-9785-bc7743817b18\" (UID: \"daa404d6-76e8-4568-9785-bc7743817b18\") " Dec 11 08:28:33 crc kubenswrapper[4860]: I1211 08:28:33.519866 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/daa404d6-76e8-4568-9785-bc7743817b18-additional-scripts\") pod \"daa404d6-76e8-4568-9785-bc7743817b18\" (UID: \"daa404d6-76e8-4568-9785-bc7743817b18\") " Dec 11 08:28:33 crc kubenswrapper[4860]: I1211 08:28:33.519893 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/daa404d6-76e8-4568-9785-bc7743817b18-var-run-ovn\") pod \"daa404d6-76e8-4568-9785-bc7743817b18\" (UID: \"daa404d6-76e8-4568-9785-bc7743817b18\") " Dec 11 08:28:33 crc kubenswrapper[4860]: I1211 08:28:33.519980 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/daa404d6-76e8-4568-9785-bc7743817b18-scripts\") pod \"daa404d6-76e8-4568-9785-bc7743817b18\" (UID: \"daa404d6-76e8-4568-9785-bc7743817b18\") " Dec 11 08:28:33 crc kubenswrapper[4860]: I1211 08:28:33.520013 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/daa404d6-76e8-4568-9785-bc7743817b18-var-log-ovn\") pod \"daa404d6-76e8-4568-9785-bc7743817b18\" (UID: \"daa404d6-76e8-4568-9785-bc7743817b18\") " Dec 11 08:28:33 crc kubenswrapper[4860]: I1211 08:28:33.520035 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2wrwl\" (UniqueName: \"kubernetes.io/projected/daa404d6-76e8-4568-9785-bc7743817b18-kube-api-access-2wrwl\") pod \"daa404d6-76e8-4568-9785-bc7743817b18\" (UID: \"daa404d6-76e8-4568-9785-bc7743817b18\") " Dec 11 08:28:33 crc kubenswrapper[4860]: I1211 08:28:33.520935 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/daa404d6-76e8-4568-9785-bc7743817b18-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "daa404d6-76e8-4568-9785-bc7743817b18" (UID: "daa404d6-76e8-4568-9785-bc7743817b18"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 08:28:33 crc kubenswrapper[4860]: I1211 08:28:33.520969 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/daa404d6-76e8-4568-9785-bc7743817b18-var-run" (OuterVolumeSpecName: "var-run") pod "daa404d6-76e8-4568-9785-bc7743817b18" (UID: "daa404d6-76e8-4568-9785-bc7743817b18"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 08:28:33 crc kubenswrapper[4860]: I1211 08:28:33.521848 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/daa404d6-76e8-4568-9785-bc7743817b18-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "daa404d6-76e8-4568-9785-bc7743817b18" (UID: "daa404d6-76e8-4568-9785-bc7743817b18"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:28:33 crc kubenswrapper[4860]: I1211 08:28:33.521899 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/daa404d6-76e8-4568-9785-bc7743817b18-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "daa404d6-76e8-4568-9785-bc7743817b18" (UID: "daa404d6-76e8-4568-9785-bc7743817b18"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 08:28:33 crc kubenswrapper[4860]: I1211 08:28:33.522161 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/daa404d6-76e8-4568-9785-bc7743817b18-scripts" (OuterVolumeSpecName: "scripts") pod "daa404d6-76e8-4568-9785-bc7743817b18" (UID: "daa404d6-76e8-4568-9785-bc7743817b18"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:28:33 crc kubenswrapper[4860]: I1211 08:28:33.529205 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/daa404d6-76e8-4568-9785-bc7743817b18-kube-api-access-2wrwl" (OuterVolumeSpecName: "kube-api-access-2wrwl") pod "daa404d6-76e8-4568-9785-bc7743817b18" (UID: "daa404d6-76e8-4568-9785-bc7743817b18"). InnerVolumeSpecName "kube-api-access-2wrwl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:28:33 crc kubenswrapper[4860]: I1211 08:28:33.621719 4860 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/daa404d6-76e8-4568-9785-bc7743817b18-additional-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 08:28:33 crc kubenswrapper[4860]: I1211 08:28:33.621759 4860 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/daa404d6-76e8-4568-9785-bc7743817b18-var-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 11 08:28:33 crc kubenswrapper[4860]: I1211 08:28:33.621774 4860 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/daa404d6-76e8-4568-9785-bc7743817b18-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 08:28:33 crc kubenswrapper[4860]: I1211 08:28:33.621787 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2wrwl\" (UniqueName: \"kubernetes.io/projected/daa404d6-76e8-4568-9785-bc7743817b18-kube-api-access-2wrwl\") on node \"crc\" DevicePath \"\"" Dec 11 08:28:33 crc kubenswrapper[4860]: I1211 08:28:33.621801 4860 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/daa404d6-76e8-4568-9785-bc7743817b18-var-log-ovn\") on node \"crc\" DevicePath \"\"" Dec 11 08:28:33 crc kubenswrapper[4860]: I1211 08:28:33.621818 4860 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/daa404d6-76e8-4568-9785-bc7743817b18-var-run\") on node \"crc\" DevicePath \"\"" Dec 11 08:28:34 crc kubenswrapper[4860]: I1211 08:28:34.126950 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-dbdtz-config-svfgb" Dec 11 08:28:34 crc kubenswrapper[4860]: I1211 08:28:34.127522 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-dbdtz-config-svfgb" event={"ID":"daa404d6-76e8-4568-9785-bc7743817b18","Type":"ContainerDied","Data":"042e4fb7b0069346e101c32fa6bb315d245f50b3fb78ca9f60fe3db1d59936c2"} Dec 11 08:28:34 crc kubenswrapper[4860]: I1211 08:28:34.127552 4860 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="042e4fb7b0069346e101c32fa6bb315d245f50b3fb78ca9f60fe3db1d59936c2" Dec 11 08:28:34 crc kubenswrapper[4860]: I1211 08:28:34.211757 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-dbdtz-config-svfgb"] Dec 11 08:28:34 crc kubenswrapper[4860]: I1211 08:28:34.227772 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-dbdtz-config-svfgb"] Dec 11 08:28:34 crc kubenswrapper[4860]: I1211 08:28:34.416806 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-dbdtz" Dec 11 08:28:35 crc kubenswrapper[4860]: I1211 08:28:35.593092 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="daa404d6-76e8-4568-9785-bc7743817b18" path="/var/lib/kubelet/pods/daa404d6-76e8-4568-9785-bc7743817b18/volumes" Dec 11 08:28:36 crc kubenswrapper[4860]: I1211 08:28:36.286868 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 11 08:28:36 crc kubenswrapper[4860]: I1211 08:28:36.395014 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:28:36 crc kubenswrapper[4860]: I1211 08:28:36.717045 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-c4mc4"] Dec 11 08:28:36 crc kubenswrapper[4860]: E1211 08:28:36.718132 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="daa404d6-76e8-4568-9785-bc7743817b18" containerName="ovn-config" Dec 11 08:28:36 crc kubenswrapper[4860]: I1211 08:28:36.718154 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="daa404d6-76e8-4568-9785-bc7743817b18" containerName="ovn-config" Dec 11 08:28:36 crc kubenswrapper[4860]: I1211 08:28:36.718369 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="daa404d6-76e8-4568-9785-bc7743817b18" containerName="ovn-config" Dec 11 08:28:36 crc kubenswrapper[4860]: I1211 08:28:36.721660 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-c4mc4" Dec 11 08:28:36 crc kubenswrapper[4860]: I1211 08:28:36.742775 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-c4mc4"] Dec 11 08:28:36 crc kubenswrapper[4860]: I1211 08:28:36.775878 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-6822-account-create-update-tjqd7"] Dec 11 08:28:36 crc kubenswrapper[4860]: I1211 08:28:36.777674 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-6822-account-create-update-tjqd7" Dec 11 08:28:36 crc kubenswrapper[4860]: I1211 08:28:36.781794 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Dec 11 08:28:36 crc kubenswrapper[4860]: I1211 08:28:36.863122 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-6822-account-create-update-tjqd7"] Dec 11 08:28:36 crc kubenswrapper[4860]: I1211 08:28:36.879682 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-dqvkz"] Dec 11 08:28:36 crc kubenswrapper[4860]: I1211 08:28:36.893804 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sl594\" (UniqueName: \"kubernetes.io/projected/e8941d7b-8d06-4ec6-8da3-101e001f824f-kube-api-access-sl594\") pod \"barbican-6822-account-create-update-tjqd7\" (UID: \"e8941d7b-8d06-4ec6-8da3-101e001f824f\") " pod="openstack/barbican-6822-account-create-update-tjqd7" Dec 11 08:28:36 crc kubenswrapper[4860]: I1211 08:28:36.897873 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-7567-account-create-update-5mzbx"] Dec 11 08:28:36 crc kubenswrapper[4860]: I1211 08:28:36.898187 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-dqvkz" Dec 11 08:28:36 crc kubenswrapper[4860]: I1211 08:28:36.901018 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-7567-account-create-update-5mzbx" Dec 11 08:28:36 crc kubenswrapper[4860]: I1211 08:28:36.901911 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ngc56\" (UniqueName: \"kubernetes.io/projected/7bea23d9-3be0-4a68-8ec1-8f21bd7feb80-kube-api-access-ngc56\") pod \"cinder-db-create-c4mc4\" (UID: \"7bea23d9-3be0-4a68-8ec1-8f21bd7feb80\") " pod="openstack/cinder-db-create-c4mc4" Dec 11 08:28:36 crc kubenswrapper[4860]: I1211 08:28:36.902221 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e8941d7b-8d06-4ec6-8da3-101e001f824f-operator-scripts\") pod \"barbican-6822-account-create-update-tjqd7\" (UID: \"e8941d7b-8d06-4ec6-8da3-101e001f824f\") " pod="openstack/barbican-6822-account-create-update-tjqd7" Dec 11 08:28:36 crc kubenswrapper[4860]: I1211 08:28:36.902459 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7bea23d9-3be0-4a68-8ec1-8f21bd7feb80-operator-scripts\") pod \"cinder-db-create-c4mc4\" (UID: \"7bea23d9-3be0-4a68-8ec1-8f21bd7feb80\") " pod="openstack/cinder-db-create-c4mc4" Dec 11 08:28:36 crc kubenswrapper[4860]: I1211 08:28:36.903213 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Dec 11 08:28:36 crc kubenswrapper[4860]: I1211 08:28:36.918488 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-dqvkz"] Dec 11 08:28:36 crc kubenswrapper[4860]: I1211 08:28:36.939539 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-7567-account-create-update-5mzbx"] Dec 11 08:28:37 crc kubenswrapper[4860]: I1211 08:28:37.004633 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e8941d7b-8d06-4ec6-8da3-101e001f824f-operator-scripts\") pod \"barbican-6822-account-create-update-tjqd7\" (UID: \"e8941d7b-8d06-4ec6-8da3-101e001f824f\") " pod="openstack/barbican-6822-account-create-update-tjqd7" Dec 11 08:28:37 crc kubenswrapper[4860]: I1211 08:28:37.004743 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7bea23d9-3be0-4a68-8ec1-8f21bd7feb80-operator-scripts\") pod \"cinder-db-create-c4mc4\" (UID: \"7bea23d9-3be0-4a68-8ec1-8f21bd7feb80\") " pod="openstack/cinder-db-create-c4mc4" Dec 11 08:28:37 crc kubenswrapper[4860]: I1211 08:28:37.004788 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ppvdj\" (UniqueName: \"kubernetes.io/projected/54f31848-ee29-43be-8363-0242dbe36164-kube-api-access-ppvdj\") pod \"cinder-7567-account-create-update-5mzbx\" (UID: \"54f31848-ee29-43be-8363-0242dbe36164\") " pod="openstack/cinder-7567-account-create-update-5mzbx" Dec 11 08:28:37 crc kubenswrapper[4860]: I1211 08:28:37.004882 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/54f31848-ee29-43be-8363-0242dbe36164-operator-scripts\") pod \"cinder-7567-account-create-update-5mzbx\" (UID: \"54f31848-ee29-43be-8363-0242dbe36164\") " pod="openstack/cinder-7567-account-create-update-5mzbx" Dec 11 08:28:37 crc kubenswrapper[4860]: I1211 08:28:37.004920 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sl594\" (UniqueName: \"kubernetes.io/projected/e8941d7b-8d06-4ec6-8da3-101e001f824f-kube-api-access-sl594\") pod \"barbican-6822-account-create-update-tjqd7\" (UID: \"e8941d7b-8d06-4ec6-8da3-101e001f824f\") " pod="openstack/barbican-6822-account-create-update-tjqd7" Dec 11 08:28:37 crc kubenswrapper[4860]: I1211 08:28:37.004942 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/07fbfcf2-916b-4cd8-b603-48c55736a6b0-operator-scripts\") pod \"barbican-db-create-dqvkz\" (UID: \"07fbfcf2-916b-4cd8-b603-48c55736a6b0\") " pod="openstack/barbican-db-create-dqvkz" Dec 11 08:28:37 crc kubenswrapper[4860]: I1211 08:28:37.004964 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ngc56\" (UniqueName: \"kubernetes.io/projected/7bea23d9-3be0-4a68-8ec1-8f21bd7feb80-kube-api-access-ngc56\") pod \"cinder-db-create-c4mc4\" (UID: \"7bea23d9-3be0-4a68-8ec1-8f21bd7feb80\") " pod="openstack/cinder-db-create-c4mc4" Dec 11 08:28:37 crc kubenswrapper[4860]: I1211 08:28:37.004986 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5g6b7\" (UniqueName: \"kubernetes.io/projected/07fbfcf2-916b-4cd8-b603-48c55736a6b0-kube-api-access-5g6b7\") pod \"barbican-db-create-dqvkz\" (UID: \"07fbfcf2-916b-4cd8-b603-48c55736a6b0\") " pod="openstack/barbican-db-create-dqvkz" Dec 11 08:28:37 crc kubenswrapper[4860]: I1211 08:28:37.006024 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e8941d7b-8d06-4ec6-8da3-101e001f824f-operator-scripts\") pod \"barbican-6822-account-create-update-tjqd7\" (UID: \"e8941d7b-8d06-4ec6-8da3-101e001f824f\") " pod="openstack/barbican-6822-account-create-update-tjqd7" Dec 11 08:28:37 crc kubenswrapper[4860]: I1211 08:28:37.006750 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7bea23d9-3be0-4a68-8ec1-8f21bd7feb80-operator-scripts\") pod \"cinder-db-create-c4mc4\" (UID: \"7bea23d9-3be0-4a68-8ec1-8f21bd7feb80\") " pod="openstack/cinder-db-create-c4mc4" Dec 11 08:28:37 crc kubenswrapper[4860]: I1211 08:28:37.013371 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-wkrv5"] Dec 11 08:28:37 crc kubenswrapper[4860]: I1211 08:28:37.014873 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-wkrv5" Dec 11 08:28:37 crc kubenswrapper[4860]: I1211 08:28:37.021569 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 11 08:28:37 crc kubenswrapper[4860]: I1211 08:28:37.021869 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 11 08:28:37 crc kubenswrapper[4860]: I1211 08:28:37.022476 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-qjrrt" Dec 11 08:28:37 crc kubenswrapper[4860]: I1211 08:28:37.028573 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 11 08:28:37 crc kubenswrapper[4860]: I1211 08:28:37.031025 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-wkrv5"] Dec 11 08:28:37 crc kubenswrapper[4860]: I1211 08:28:37.040077 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-970d-account-create-update-65w8k"] Dec 11 08:28:37 crc kubenswrapper[4860]: I1211 08:28:37.042613 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-970d-account-create-update-65w8k" Dec 11 08:28:37 crc kubenswrapper[4860]: I1211 08:28:37.045272 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ngc56\" (UniqueName: \"kubernetes.io/projected/7bea23d9-3be0-4a68-8ec1-8f21bd7feb80-kube-api-access-ngc56\") pod \"cinder-db-create-c4mc4\" (UID: \"7bea23d9-3be0-4a68-8ec1-8f21bd7feb80\") " pod="openstack/cinder-db-create-c4mc4" Dec 11 08:28:37 crc kubenswrapper[4860]: I1211 08:28:37.045988 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Dec 11 08:28:37 crc kubenswrapper[4860]: I1211 08:28:37.059257 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-970d-account-create-update-65w8k"] Dec 11 08:28:37 crc kubenswrapper[4860]: I1211 08:28:37.060589 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-c4mc4" Dec 11 08:28:37 crc kubenswrapper[4860]: I1211 08:28:37.110812 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5af63911-2f28-40ca-9d02-721f5871f52d-operator-scripts\") pod \"neutron-970d-account-create-update-65w8k\" (UID: \"5af63911-2f28-40ca-9d02-721f5871f52d\") " pod="openstack/neutron-970d-account-create-update-65w8k" Dec 11 08:28:37 crc kubenswrapper[4860]: I1211 08:28:37.110926 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9rw9p\" (UniqueName: \"kubernetes.io/projected/5af63911-2f28-40ca-9d02-721f5871f52d-kube-api-access-9rw9p\") pod \"neutron-970d-account-create-update-65w8k\" (UID: \"5af63911-2f28-40ca-9d02-721f5871f52d\") " pod="openstack/neutron-970d-account-create-update-65w8k" Dec 11 08:28:37 crc kubenswrapper[4860]: I1211 08:28:37.111006 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ppvdj\" (UniqueName: \"kubernetes.io/projected/54f31848-ee29-43be-8363-0242dbe36164-kube-api-access-ppvdj\") pod \"cinder-7567-account-create-update-5mzbx\" (UID: \"54f31848-ee29-43be-8363-0242dbe36164\") " pod="openstack/cinder-7567-account-create-update-5mzbx" Dec 11 08:28:37 crc kubenswrapper[4860]: I1211 08:28:37.111062 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gx2pt\" (UniqueName: \"kubernetes.io/projected/c602556f-fbc0-4e66-bd77-0d5b029e81bb-kube-api-access-gx2pt\") pod \"keystone-db-sync-wkrv5\" (UID: \"c602556f-fbc0-4e66-bd77-0d5b029e81bb\") " pod="openstack/keystone-db-sync-wkrv5" Dec 11 08:28:37 crc kubenswrapper[4860]: I1211 08:28:37.111107 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c602556f-fbc0-4e66-bd77-0d5b029e81bb-config-data\") pod \"keystone-db-sync-wkrv5\" (UID: \"c602556f-fbc0-4e66-bd77-0d5b029e81bb\") " pod="openstack/keystone-db-sync-wkrv5" Dec 11 08:28:37 crc kubenswrapper[4860]: I1211 08:28:37.111155 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/54f31848-ee29-43be-8363-0242dbe36164-operator-scripts\") pod \"cinder-7567-account-create-update-5mzbx\" (UID: \"54f31848-ee29-43be-8363-0242dbe36164\") " pod="openstack/cinder-7567-account-create-update-5mzbx" Dec 11 08:28:37 crc kubenswrapper[4860]: I1211 08:28:37.111234 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c602556f-fbc0-4e66-bd77-0d5b029e81bb-combined-ca-bundle\") pod \"keystone-db-sync-wkrv5\" (UID: \"c602556f-fbc0-4e66-bd77-0d5b029e81bb\") " pod="openstack/keystone-db-sync-wkrv5" Dec 11 08:28:37 crc kubenswrapper[4860]: I1211 08:28:37.111295 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/07fbfcf2-916b-4cd8-b603-48c55736a6b0-operator-scripts\") pod \"barbican-db-create-dqvkz\" (UID: \"07fbfcf2-916b-4cd8-b603-48c55736a6b0\") " pod="openstack/barbican-db-create-dqvkz" Dec 11 08:28:37 crc kubenswrapper[4860]: I1211 08:28:37.111323 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5g6b7\" (UniqueName: \"kubernetes.io/projected/07fbfcf2-916b-4cd8-b603-48c55736a6b0-kube-api-access-5g6b7\") pod \"barbican-db-create-dqvkz\" (UID: \"07fbfcf2-916b-4cd8-b603-48c55736a6b0\") " pod="openstack/barbican-db-create-dqvkz" Dec 11 08:28:37 crc kubenswrapper[4860]: I1211 08:28:37.112420 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/54f31848-ee29-43be-8363-0242dbe36164-operator-scripts\") pod \"cinder-7567-account-create-update-5mzbx\" (UID: \"54f31848-ee29-43be-8363-0242dbe36164\") " pod="openstack/cinder-7567-account-create-update-5mzbx" Dec 11 08:28:37 crc kubenswrapper[4860]: I1211 08:28:37.112979 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/07fbfcf2-916b-4cd8-b603-48c55736a6b0-operator-scripts\") pod \"barbican-db-create-dqvkz\" (UID: \"07fbfcf2-916b-4cd8-b603-48c55736a6b0\") " pod="openstack/barbican-db-create-dqvkz" Dec 11 08:28:37 crc kubenswrapper[4860]: I1211 08:28:37.157728 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5g6b7\" (UniqueName: \"kubernetes.io/projected/07fbfcf2-916b-4cd8-b603-48c55736a6b0-kube-api-access-5g6b7\") pod \"barbican-db-create-dqvkz\" (UID: \"07fbfcf2-916b-4cd8-b603-48c55736a6b0\") " pod="openstack/barbican-db-create-dqvkz" Dec 11 08:28:37 crc kubenswrapper[4860]: I1211 08:28:37.159541 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ppvdj\" (UniqueName: \"kubernetes.io/projected/54f31848-ee29-43be-8363-0242dbe36164-kube-api-access-ppvdj\") pod \"cinder-7567-account-create-update-5mzbx\" (UID: \"54f31848-ee29-43be-8363-0242dbe36164\") " pod="openstack/cinder-7567-account-create-update-5mzbx" Dec 11 08:28:37 crc kubenswrapper[4860]: I1211 08:28:37.163321 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-lrszm"] Dec 11 08:28:37 crc kubenswrapper[4860]: I1211 08:28:37.164945 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-lrszm" Dec 11 08:28:37 crc kubenswrapper[4860]: I1211 08:28:37.172402 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-lrszm"] Dec 11 08:28:37 crc kubenswrapper[4860]: I1211 08:28:37.213612 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5af63911-2f28-40ca-9d02-721f5871f52d-operator-scripts\") pod \"neutron-970d-account-create-update-65w8k\" (UID: \"5af63911-2f28-40ca-9d02-721f5871f52d\") " pod="openstack/neutron-970d-account-create-update-65w8k" Dec 11 08:28:37 crc kubenswrapper[4860]: I1211 08:28:37.213755 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9rw9p\" (UniqueName: \"kubernetes.io/projected/5af63911-2f28-40ca-9d02-721f5871f52d-kube-api-access-9rw9p\") pod \"neutron-970d-account-create-update-65w8k\" (UID: \"5af63911-2f28-40ca-9d02-721f5871f52d\") " pod="openstack/neutron-970d-account-create-update-65w8k" Dec 11 08:28:37 crc kubenswrapper[4860]: I1211 08:28:37.213828 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gx2pt\" (UniqueName: \"kubernetes.io/projected/c602556f-fbc0-4e66-bd77-0d5b029e81bb-kube-api-access-gx2pt\") pod \"keystone-db-sync-wkrv5\" (UID: \"c602556f-fbc0-4e66-bd77-0d5b029e81bb\") " pod="openstack/keystone-db-sync-wkrv5" Dec 11 08:28:37 crc kubenswrapper[4860]: I1211 08:28:37.213881 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c602556f-fbc0-4e66-bd77-0d5b029e81bb-config-data\") pod \"keystone-db-sync-wkrv5\" (UID: \"c602556f-fbc0-4e66-bd77-0d5b029e81bb\") " pod="openstack/keystone-db-sync-wkrv5" Dec 11 08:28:37 crc kubenswrapper[4860]: I1211 08:28:37.213931 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c602556f-fbc0-4e66-bd77-0d5b029e81bb-combined-ca-bundle\") pod \"keystone-db-sync-wkrv5\" (UID: \"c602556f-fbc0-4e66-bd77-0d5b029e81bb\") " pod="openstack/keystone-db-sync-wkrv5" Dec 11 08:28:37 crc kubenswrapper[4860]: I1211 08:28:37.215451 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5af63911-2f28-40ca-9d02-721f5871f52d-operator-scripts\") pod \"neutron-970d-account-create-update-65w8k\" (UID: \"5af63911-2f28-40ca-9d02-721f5871f52d\") " pod="openstack/neutron-970d-account-create-update-65w8k" Dec 11 08:28:37 crc kubenswrapper[4860]: I1211 08:28:37.220075 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c602556f-fbc0-4e66-bd77-0d5b029e81bb-combined-ca-bundle\") pod \"keystone-db-sync-wkrv5\" (UID: \"c602556f-fbc0-4e66-bd77-0d5b029e81bb\") " pod="openstack/keystone-db-sync-wkrv5" Dec 11 08:28:37 crc kubenswrapper[4860]: I1211 08:28:37.224537 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c602556f-fbc0-4e66-bd77-0d5b029e81bb-config-data\") pod \"keystone-db-sync-wkrv5\" (UID: \"c602556f-fbc0-4e66-bd77-0d5b029e81bb\") " pod="openstack/keystone-db-sync-wkrv5" Dec 11 08:28:37 crc kubenswrapper[4860]: I1211 08:28:37.238468 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9rw9p\" (UniqueName: \"kubernetes.io/projected/5af63911-2f28-40ca-9d02-721f5871f52d-kube-api-access-9rw9p\") pod \"neutron-970d-account-create-update-65w8k\" (UID: \"5af63911-2f28-40ca-9d02-721f5871f52d\") " pod="openstack/neutron-970d-account-create-update-65w8k" Dec 11 08:28:37 crc kubenswrapper[4860]: I1211 08:28:37.246882 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-dqvkz" Dec 11 08:28:37 crc kubenswrapper[4860]: I1211 08:28:37.249158 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gx2pt\" (UniqueName: \"kubernetes.io/projected/c602556f-fbc0-4e66-bd77-0d5b029e81bb-kube-api-access-gx2pt\") pod \"keystone-db-sync-wkrv5\" (UID: \"c602556f-fbc0-4e66-bd77-0d5b029e81bb\") " pod="openstack/keystone-db-sync-wkrv5" Dec 11 08:28:37 crc kubenswrapper[4860]: I1211 08:28:37.261426 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-7567-account-create-update-5mzbx" Dec 11 08:28:37 crc kubenswrapper[4860]: I1211 08:28:37.325300 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ks4ln\" (UniqueName: \"kubernetes.io/projected/989e314b-f91d-4ceb-ba16-79d180225afe-kube-api-access-ks4ln\") pod \"neutron-db-create-lrszm\" (UID: \"989e314b-f91d-4ceb-ba16-79d180225afe\") " pod="openstack/neutron-db-create-lrszm" Dec 11 08:28:37 crc kubenswrapper[4860]: I1211 08:28:37.325452 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/989e314b-f91d-4ceb-ba16-79d180225afe-operator-scripts\") pod \"neutron-db-create-lrszm\" (UID: \"989e314b-f91d-4ceb-ba16-79d180225afe\") " pod="openstack/neutron-db-create-lrszm" Dec 11 08:28:37 crc kubenswrapper[4860]: I1211 08:28:37.339280 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-wkrv5" Dec 11 08:28:37 crc kubenswrapper[4860]: I1211 08:28:37.419337 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-970d-account-create-update-65w8k" Dec 11 08:28:37 crc kubenswrapper[4860]: I1211 08:28:37.427256 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ks4ln\" (UniqueName: \"kubernetes.io/projected/989e314b-f91d-4ceb-ba16-79d180225afe-kube-api-access-ks4ln\") pod \"neutron-db-create-lrszm\" (UID: \"989e314b-f91d-4ceb-ba16-79d180225afe\") " pod="openstack/neutron-db-create-lrszm" Dec 11 08:28:37 crc kubenswrapper[4860]: I1211 08:28:37.427361 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/989e314b-f91d-4ceb-ba16-79d180225afe-operator-scripts\") pod \"neutron-db-create-lrszm\" (UID: \"989e314b-f91d-4ceb-ba16-79d180225afe\") " pod="openstack/neutron-db-create-lrszm" Dec 11 08:28:37 crc kubenswrapper[4860]: W1211 08:28:37.647825 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7bea23d9_3be0_4a68_8ec1_8f21bd7feb80.slice/crio-ab608a8a6e34f7731f78a9efaac355fb0691043bb294a9980b57271b4a79aa35 WatchSource:0}: Error finding container ab608a8a6e34f7731f78a9efaac355fb0691043bb294a9980b57271b4a79aa35: Status 404 returned error can't find the container with id ab608a8a6e34f7731f78a9efaac355fb0691043bb294a9980b57271b4a79aa35 Dec 11 08:28:37 crc kubenswrapper[4860]: I1211 08:28:37.662396 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-c4mc4"] Dec 11 08:28:37 crc kubenswrapper[4860]: I1211 08:28:37.812043 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-7567-account-create-update-5mzbx"] Dec 11 08:28:37 crc kubenswrapper[4860]: W1211 08:28:37.820625 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod54f31848_ee29_43be_8363_0242dbe36164.slice/crio-19228c8b2300ccad4375e6b3832ce28cfe988a9c7be9cda6f79c1cbf4259351c WatchSource:0}: Error finding container 19228c8b2300ccad4375e6b3832ce28cfe988a9c7be9cda6f79c1cbf4259351c: Status 404 returned error can't find the container with id 19228c8b2300ccad4375e6b3832ce28cfe988a9c7be9cda6f79c1cbf4259351c Dec 11 08:28:37 crc kubenswrapper[4860]: I1211 08:28:37.864624 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-dqvkz"] Dec 11 08:28:37 crc kubenswrapper[4860]: I1211 08:28:37.982383 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-970d-account-create-update-65w8k"] Dec 11 08:28:38 crc kubenswrapper[4860]: I1211 08:28:38.003426 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-wkrv5"] Dec 11 08:28:38 crc kubenswrapper[4860]: W1211 08:28:38.041200 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5af63911_2f28_40ca_9d02_721f5871f52d.slice/crio-5ce211da1331a0b7ee47c2e2d9f00f5c0da183761829855e6b02b964905c7623 WatchSource:0}: Error finding container 5ce211da1331a0b7ee47c2e2d9f00f5c0da183761829855e6b02b964905c7623: Status 404 returned error can't find the container with id 5ce211da1331a0b7ee47c2e2d9f00f5c0da183761829855e6b02b964905c7623 Dec 11 08:28:38 crc kubenswrapper[4860]: I1211 08:28:38.206843 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/989e314b-f91d-4ceb-ba16-79d180225afe-operator-scripts\") pod \"neutron-db-create-lrszm\" (UID: \"989e314b-f91d-4ceb-ba16-79d180225afe\") " pod="openstack/neutron-db-create-lrszm" Dec 11 08:28:38 crc kubenswrapper[4860]: I1211 08:28:38.212400 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sl594\" (UniqueName: \"kubernetes.io/projected/e8941d7b-8d06-4ec6-8da3-101e001f824f-kube-api-access-sl594\") pod \"barbican-6822-account-create-update-tjqd7\" (UID: \"e8941d7b-8d06-4ec6-8da3-101e001f824f\") " pod="openstack/barbican-6822-account-create-update-tjqd7" Dec 11 08:28:38 crc kubenswrapper[4860]: I1211 08:28:38.221392 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ks4ln\" (UniqueName: \"kubernetes.io/projected/989e314b-f91d-4ceb-ba16-79d180225afe-kube-api-access-ks4ln\") pod \"neutron-db-create-lrszm\" (UID: \"989e314b-f91d-4ceb-ba16-79d180225afe\") " pod="openstack/neutron-db-create-lrszm" Dec 11 08:28:38 crc kubenswrapper[4860]: I1211 08:28:38.233086 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-dqvkz" event={"ID":"07fbfcf2-916b-4cd8-b603-48c55736a6b0","Type":"ContainerStarted","Data":"7ae9e34206ea54a1ca4f2e34155874a9a8d9379c0d3845a5efbefd454615d72e"} Dec 11 08:28:38 crc kubenswrapper[4860]: I1211 08:28:38.263170 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-7567-account-create-update-5mzbx" event={"ID":"54f31848-ee29-43be-8363-0242dbe36164","Type":"ContainerStarted","Data":"19228c8b2300ccad4375e6b3832ce28cfe988a9c7be9cda6f79c1cbf4259351c"} Dec 11 08:28:38 crc kubenswrapper[4860]: I1211 08:28:38.268830 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-c4mc4" event={"ID":"7bea23d9-3be0-4a68-8ec1-8f21bd7feb80","Type":"ContainerStarted","Data":"ab608a8a6e34f7731f78a9efaac355fb0691043bb294a9980b57271b4a79aa35"} Dec 11 08:28:38 crc kubenswrapper[4860]: I1211 08:28:38.279273 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-970d-account-create-update-65w8k" event={"ID":"5af63911-2f28-40ca-9d02-721f5871f52d","Type":"ContainerStarted","Data":"5ce211da1331a0b7ee47c2e2d9f00f5c0da183761829855e6b02b964905c7623"} Dec 11 08:28:38 crc kubenswrapper[4860]: I1211 08:28:38.282585 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-wkrv5" event={"ID":"c602556f-fbc0-4e66-bd77-0d5b029e81bb","Type":"ContainerStarted","Data":"2495d53ebd4963320e8049c2dfa6ebcbefbff1951a71652a285f1b7aca5fd57a"} Dec 11 08:28:38 crc kubenswrapper[4860]: I1211 08:28:38.329875 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-6822-account-create-update-tjqd7" Dec 11 08:28:38 crc kubenswrapper[4860]: I1211 08:28:38.511881 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-lrszm" Dec 11 08:28:38 crc kubenswrapper[4860]: I1211 08:28:38.797317 4860 patch_prober.go:28] interesting pod/machine-config-daemon-99qgp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 08:28:38 crc kubenswrapper[4860]: I1211 08:28:38.797784 4860 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 08:28:38 crc kubenswrapper[4860]: I1211 08:28:38.797837 4860 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" Dec 11 08:28:38 crc kubenswrapper[4860]: I1211 08:28:38.798594 4860 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"565fbddca3ff6a011767df936f9e699e4560197af6e486d467234b4599b2d2f6"} pod="openshift-machine-config-operator/machine-config-daemon-99qgp" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 11 08:28:38 crc kubenswrapper[4860]: I1211 08:28:38.798666 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" containerName="machine-config-daemon" containerID="cri-o://565fbddca3ff6a011767df936f9e699e4560197af6e486d467234b4599b2d2f6" gracePeriod=600 Dec 11 08:28:38 crc kubenswrapper[4860]: I1211 08:28:38.946335 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-6822-account-create-update-tjqd7"] Dec 11 08:28:39 crc kubenswrapper[4860]: I1211 08:28:39.209784 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-lrszm"] Dec 11 08:28:39 crc kubenswrapper[4860]: I1211 08:28:39.297377 4860 generic.go:334] "Generic (PLEG): container finished" podID="5af63911-2f28-40ca-9d02-721f5871f52d" containerID="0b11a85c964986dfd28dd78240fb9ced04ee192ff15828c7ed754be53dcb5116" exitCode=0 Dec 11 08:28:39 crc kubenswrapper[4860]: I1211 08:28:39.297461 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-970d-account-create-update-65w8k" event={"ID":"5af63911-2f28-40ca-9d02-721f5871f52d","Type":"ContainerDied","Data":"0b11a85c964986dfd28dd78240fb9ced04ee192ff15828c7ed754be53dcb5116"} Dec 11 08:28:39 crc kubenswrapper[4860]: I1211 08:28:39.319348 4860 generic.go:334] "Generic (PLEG): container finished" podID="31c30642-6e60-41b4-a477-0d802424e0aa" containerID="565fbddca3ff6a011767df936f9e699e4560197af6e486d467234b4599b2d2f6" exitCode=0 Dec 11 08:28:39 crc kubenswrapper[4860]: I1211 08:28:39.319433 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" event={"ID":"31c30642-6e60-41b4-a477-0d802424e0aa","Type":"ContainerDied","Data":"565fbddca3ff6a011767df936f9e699e4560197af6e486d467234b4599b2d2f6"} Dec 11 08:28:39 crc kubenswrapper[4860]: I1211 08:28:39.319538 4860 scope.go:117] "RemoveContainer" containerID="7abbc21a5acc8c61b19323dc53f4bb2e28d39cd9e4964ba8a2b38c1bd09ed9bf" Dec 11 08:28:39 crc kubenswrapper[4860]: I1211 08:28:39.322931 4860 generic.go:334] "Generic (PLEG): container finished" podID="07fbfcf2-916b-4cd8-b603-48c55736a6b0" containerID="da10db9abaea4b7eac10a44bfd1ba1fec17869d0afab3a371521004f840f6199" exitCode=0 Dec 11 08:28:39 crc kubenswrapper[4860]: I1211 08:28:39.323010 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-dqvkz" event={"ID":"07fbfcf2-916b-4cd8-b603-48c55736a6b0","Type":"ContainerDied","Data":"da10db9abaea4b7eac10a44bfd1ba1fec17869d0afab3a371521004f840f6199"} Dec 11 08:28:39 crc kubenswrapper[4860]: I1211 08:28:39.332303 4860 generic.go:334] "Generic (PLEG): container finished" podID="54f31848-ee29-43be-8363-0242dbe36164" containerID="e5a0736b99cd2f37cf621cf0decadfdb56873bb6ff930b8c31cd4a8e299cb90f" exitCode=0 Dec 11 08:28:39 crc kubenswrapper[4860]: I1211 08:28:39.332382 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-7567-account-create-update-5mzbx" event={"ID":"54f31848-ee29-43be-8363-0242dbe36164","Type":"ContainerDied","Data":"e5a0736b99cd2f37cf621cf0decadfdb56873bb6ff930b8c31cd4a8e299cb90f"} Dec 11 08:28:39 crc kubenswrapper[4860]: I1211 08:28:39.338391 4860 generic.go:334] "Generic (PLEG): container finished" podID="7bea23d9-3be0-4a68-8ec1-8f21bd7feb80" containerID="1d267bbbb0fe1655be49f1107ace8ca93c9627c3c5d44ca882c0d994a7f8e050" exitCode=0 Dec 11 08:28:39 crc kubenswrapper[4860]: I1211 08:28:39.338459 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-c4mc4" event={"ID":"7bea23d9-3be0-4a68-8ec1-8f21bd7feb80","Type":"ContainerDied","Data":"1d267bbbb0fe1655be49f1107ace8ca93c9627c3c5d44ca882c0d994a7f8e050"} Dec 11 08:28:40 crc kubenswrapper[4860]: I1211 08:28:40.848969 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6d5b6d6b67-pxxrf" Dec 11 08:28:40 crc kubenswrapper[4860]: I1211 08:28:40.926138 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-l7vfs"] Dec 11 08:28:40 crc kubenswrapper[4860]: I1211 08:28:40.926514 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-b8fbc5445-l7vfs" podUID="77f658e2-b468-45dd-8204-691cdfe9bf61" containerName="dnsmasq-dns" containerID="cri-o://aca4cdc244eaa6aa25f8078b98879c5acf09174e9e804d8ee9f36118d2bd6cdb" gracePeriod=10 Dec 11 08:28:41 crc kubenswrapper[4860]: I1211 08:28:41.369219 4860 generic.go:334] "Generic (PLEG): container finished" podID="77f658e2-b468-45dd-8204-691cdfe9bf61" containerID="aca4cdc244eaa6aa25f8078b98879c5acf09174e9e804d8ee9f36118d2bd6cdb" exitCode=0 Dec 11 08:28:41 crc kubenswrapper[4860]: I1211 08:28:41.369276 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-l7vfs" event={"ID":"77f658e2-b468-45dd-8204-691cdfe9bf61","Type":"ContainerDied","Data":"aca4cdc244eaa6aa25f8078b98879c5acf09174e9e804d8ee9f36118d2bd6cdb"} Dec 11 08:28:41 crc kubenswrapper[4860]: I1211 08:28:41.649549 4860 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-b8fbc5445-l7vfs" podUID="77f658e2-b468-45dd-8204-691cdfe9bf61" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.111:5353: connect: connection refused" Dec 11 08:28:46 crc kubenswrapper[4860]: I1211 08:28:46.649112 4860 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-b8fbc5445-l7vfs" podUID="77f658e2-b468-45dd-8204-691cdfe9bf61" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.111:5353: connect: connection refused" Dec 11 08:28:47 crc kubenswrapper[4860]: W1211 08:28:47.420835 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode8941d7b_8d06_4ec6_8da3_101e001f824f.slice/crio-6b719eb1082e2b08eb928bb0c8558065855e353ce4b9285483147626e7de2b87 WatchSource:0}: Error finding container 6b719eb1082e2b08eb928bb0c8558065855e353ce4b9285483147626e7de2b87: Status 404 returned error can't find the container with id 6b719eb1082e2b08eb928bb0c8558065855e353ce4b9285483147626e7de2b87 Dec 11 08:28:47 crc kubenswrapper[4860]: I1211 08:28:47.427130 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-dqvkz" event={"ID":"07fbfcf2-916b-4cd8-b603-48c55736a6b0","Type":"ContainerDied","Data":"7ae9e34206ea54a1ca4f2e34155874a9a8d9379c0d3845a5efbefd454615d72e"} Dec 11 08:28:47 crc kubenswrapper[4860]: I1211 08:28:47.427196 4860 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7ae9e34206ea54a1ca4f2e34155874a9a8d9379c0d3845a5efbefd454615d72e" Dec 11 08:28:47 crc kubenswrapper[4860]: W1211 08:28:47.436976 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod989e314b_f91d_4ceb_ba16_79d180225afe.slice/crio-ab91588af10cd77e03efd539d649d16014e97eaac5b312ab801d44a2949bf664 WatchSource:0}: Error finding container ab91588af10cd77e03efd539d649d16014e97eaac5b312ab801d44a2949bf664: Status 404 returned error can't find the container with id ab91588af10cd77e03efd539d649d16014e97eaac5b312ab801d44a2949bf664 Dec 11 08:28:47 crc kubenswrapper[4860]: E1211 08:28:47.485594 4860 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-glance-api:current-podified" Dec 11 08:28:47 crc kubenswrapper[4860]: E1211 08:28:47.486097 4860 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:glance-db-sync,Image:quay.io/podified-antelope-centos9/openstack-glance-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/glance/glance.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-cmvl7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42415,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42415,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-db-sync-4fsqr_openstack(cf433d5d-8c95-4d72-864c-3b74f315dd2b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 11 08:28:47 crc kubenswrapper[4860]: E1211 08:28:47.487545 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/glance-db-sync-4fsqr" podUID="cf433d5d-8c95-4d72-864c-3b74f315dd2b" Dec 11 08:28:47 crc kubenswrapper[4860]: I1211 08:28:47.655452 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-dqvkz" Dec 11 08:28:47 crc kubenswrapper[4860]: I1211 08:28:47.701105 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-7567-account-create-update-5mzbx" Dec 11 08:28:47 crc kubenswrapper[4860]: I1211 08:28:47.713947 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-970d-account-create-update-65w8k" Dec 11 08:28:47 crc kubenswrapper[4860]: I1211 08:28:47.736829 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-c4mc4" Dec 11 08:28:47 crc kubenswrapper[4860]: I1211 08:28:47.811063 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/54f31848-ee29-43be-8363-0242dbe36164-operator-scripts\") pod \"54f31848-ee29-43be-8363-0242dbe36164\" (UID: \"54f31848-ee29-43be-8363-0242dbe36164\") " Dec 11 08:28:47 crc kubenswrapper[4860]: I1211 08:28:47.811113 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ppvdj\" (UniqueName: \"kubernetes.io/projected/54f31848-ee29-43be-8363-0242dbe36164-kube-api-access-ppvdj\") pod \"54f31848-ee29-43be-8363-0242dbe36164\" (UID: \"54f31848-ee29-43be-8363-0242dbe36164\") " Dec 11 08:28:47 crc kubenswrapper[4860]: I1211 08:28:47.811243 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/07fbfcf2-916b-4cd8-b603-48c55736a6b0-operator-scripts\") pod \"07fbfcf2-916b-4cd8-b603-48c55736a6b0\" (UID: \"07fbfcf2-916b-4cd8-b603-48c55736a6b0\") " Dec 11 08:28:47 crc kubenswrapper[4860]: I1211 08:28:47.811300 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5af63911-2f28-40ca-9d02-721f5871f52d-operator-scripts\") pod \"5af63911-2f28-40ca-9d02-721f5871f52d\" (UID: \"5af63911-2f28-40ca-9d02-721f5871f52d\") " Dec 11 08:28:47 crc kubenswrapper[4860]: I1211 08:28:47.811385 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngc56\" (UniqueName: \"kubernetes.io/projected/7bea23d9-3be0-4a68-8ec1-8f21bd7feb80-kube-api-access-ngc56\") pod \"7bea23d9-3be0-4a68-8ec1-8f21bd7feb80\" (UID: \"7bea23d9-3be0-4a68-8ec1-8f21bd7feb80\") " Dec 11 08:28:47 crc kubenswrapper[4860]: I1211 08:28:47.811454 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7bea23d9-3be0-4a68-8ec1-8f21bd7feb80-operator-scripts\") pod \"7bea23d9-3be0-4a68-8ec1-8f21bd7feb80\" (UID: \"7bea23d9-3be0-4a68-8ec1-8f21bd7feb80\") " Dec 11 08:28:47 crc kubenswrapper[4860]: I1211 08:28:47.811482 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9rw9p\" (UniqueName: \"kubernetes.io/projected/5af63911-2f28-40ca-9d02-721f5871f52d-kube-api-access-9rw9p\") pod \"5af63911-2f28-40ca-9d02-721f5871f52d\" (UID: \"5af63911-2f28-40ca-9d02-721f5871f52d\") " Dec 11 08:28:47 crc kubenswrapper[4860]: I1211 08:28:47.811509 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5g6b7\" (UniqueName: \"kubernetes.io/projected/07fbfcf2-916b-4cd8-b603-48c55736a6b0-kube-api-access-5g6b7\") pod \"07fbfcf2-916b-4cd8-b603-48c55736a6b0\" (UID: \"07fbfcf2-916b-4cd8-b603-48c55736a6b0\") " Dec 11 08:28:47 crc kubenswrapper[4860]: I1211 08:28:47.812480 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5af63911-2f28-40ca-9d02-721f5871f52d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5af63911-2f28-40ca-9d02-721f5871f52d" (UID: "5af63911-2f28-40ca-9d02-721f5871f52d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:28:47 crc kubenswrapper[4860]: I1211 08:28:47.812758 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bea23d9-3be0-4a68-8ec1-8f21bd7feb80-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7bea23d9-3be0-4a68-8ec1-8f21bd7feb80" (UID: "7bea23d9-3be0-4a68-8ec1-8f21bd7feb80"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:28:47 crc kubenswrapper[4860]: I1211 08:28:47.813170 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/54f31848-ee29-43be-8363-0242dbe36164-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "54f31848-ee29-43be-8363-0242dbe36164" (UID: "54f31848-ee29-43be-8363-0242dbe36164"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:28:47 crc kubenswrapper[4860]: I1211 08:28:47.816489 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/07fbfcf2-916b-4cd8-b603-48c55736a6b0-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "07fbfcf2-916b-4cd8-b603-48c55736a6b0" (UID: "07fbfcf2-916b-4cd8-b603-48c55736a6b0"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:28:47 crc kubenswrapper[4860]: I1211 08:28:47.819805 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/07fbfcf2-916b-4cd8-b603-48c55736a6b0-kube-api-access-5g6b7" (OuterVolumeSpecName: "kube-api-access-5g6b7") pod "07fbfcf2-916b-4cd8-b603-48c55736a6b0" (UID: "07fbfcf2-916b-4cd8-b603-48c55736a6b0"). InnerVolumeSpecName "kube-api-access-5g6b7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:28:47 crc kubenswrapper[4860]: I1211 08:28:47.821829 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bea23d9-3be0-4a68-8ec1-8f21bd7feb80-kube-api-access-ngc56" (OuterVolumeSpecName: "kube-api-access-ngc56") pod "7bea23d9-3be0-4a68-8ec1-8f21bd7feb80" (UID: "7bea23d9-3be0-4a68-8ec1-8f21bd7feb80"). InnerVolumeSpecName "kube-api-access-ngc56". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:28:47 crc kubenswrapper[4860]: I1211 08:28:47.821967 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5af63911-2f28-40ca-9d02-721f5871f52d-kube-api-access-9rw9p" (OuterVolumeSpecName: "kube-api-access-9rw9p") pod "5af63911-2f28-40ca-9d02-721f5871f52d" (UID: "5af63911-2f28-40ca-9d02-721f5871f52d"). InnerVolumeSpecName "kube-api-access-9rw9p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:28:47 crc kubenswrapper[4860]: I1211 08:28:47.823497 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/54f31848-ee29-43be-8363-0242dbe36164-kube-api-access-ppvdj" (OuterVolumeSpecName: "kube-api-access-ppvdj") pod "54f31848-ee29-43be-8363-0242dbe36164" (UID: "54f31848-ee29-43be-8363-0242dbe36164"). InnerVolumeSpecName "kube-api-access-ppvdj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:28:47 crc kubenswrapper[4860]: I1211 08:28:47.902115 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-l7vfs" Dec 11 08:28:47 crc kubenswrapper[4860]: I1211 08:28:47.913499 4860 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/07fbfcf2-916b-4cd8-b603-48c55736a6b0-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 08:28:47 crc kubenswrapper[4860]: I1211 08:28:47.913546 4860 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5af63911-2f28-40ca-9d02-721f5871f52d-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 08:28:47 crc kubenswrapper[4860]: I1211 08:28:47.913561 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngc56\" (UniqueName: \"kubernetes.io/projected/7bea23d9-3be0-4a68-8ec1-8f21bd7feb80-kube-api-access-ngc56\") on node \"crc\" DevicePath \"\"" Dec 11 08:28:47 crc kubenswrapper[4860]: I1211 08:28:47.913577 4860 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7bea23d9-3be0-4a68-8ec1-8f21bd7feb80-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 08:28:47 crc kubenswrapper[4860]: I1211 08:28:47.913594 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9rw9p\" (UniqueName: \"kubernetes.io/projected/5af63911-2f28-40ca-9d02-721f5871f52d-kube-api-access-9rw9p\") on node \"crc\" DevicePath \"\"" Dec 11 08:28:47 crc kubenswrapper[4860]: I1211 08:28:47.913609 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5g6b7\" (UniqueName: \"kubernetes.io/projected/07fbfcf2-916b-4cd8-b603-48c55736a6b0-kube-api-access-5g6b7\") on node \"crc\" DevicePath \"\"" Dec 11 08:28:47 crc kubenswrapper[4860]: I1211 08:28:47.913621 4860 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/54f31848-ee29-43be-8363-0242dbe36164-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 08:28:47 crc kubenswrapper[4860]: I1211 08:28:47.913633 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ppvdj\" (UniqueName: \"kubernetes.io/projected/54f31848-ee29-43be-8363-0242dbe36164-kube-api-access-ppvdj\") on node \"crc\" DevicePath \"\"" Dec 11 08:28:48 crc kubenswrapper[4860]: I1211 08:28:48.015107 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/77f658e2-b468-45dd-8204-691cdfe9bf61-dns-svc\") pod \"77f658e2-b468-45dd-8204-691cdfe9bf61\" (UID: \"77f658e2-b468-45dd-8204-691cdfe9bf61\") " Dec 11 08:28:48 crc kubenswrapper[4860]: I1211 08:28:48.015513 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/77f658e2-b468-45dd-8204-691cdfe9bf61-ovsdbserver-nb\") pod \"77f658e2-b468-45dd-8204-691cdfe9bf61\" (UID: \"77f658e2-b468-45dd-8204-691cdfe9bf61\") " Dec 11 08:28:48 crc kubenswrapper[4860]: I1211 08:28:48.015596 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/77f658e2-b468-45dd-8204-691cdfe9bf61-config\") pod \"77f658e2-b468-45dd-8204-691cdfe9bf61\" (UID: \"77f658e2-b468-45dd-8204-691cdfe9bf61\") " Dec 11 08:28:48 crc kubenswrapper[4860]: I1211 08:28:48.015636 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dxtnk\" (UniqueName: \"kubernetes.io/projected/77f658e2-b468-45dd-8204-691cdfe9bf61-kube-api-access-dxtnk\") pod \"77f658e2-b468-45dd-8204-691cdfe9bf61\" (UID: \"77f658e2-b468-45dd-8204-691cdfe9bf61\") " Dec 11 08:28:48 crc kubenswrapper[4860]: I1211 08:28:48.015676 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/77f658e2-b468-45dd-8204-691cdfe9bf61-ovsdbserver-sb\") pod \"77f658e2-b468-45dd-8204-691cdfe9bf61\" (UID: \"77f658e2-b468-45dd-8204-691cdfe9bf61\") " Dec 11 08:28:48 crc kubenswrapper[4860]: I1211 08:28:48.021408 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/77f658e2-b468-45dd-8204-691cdfe9bf61-kube-api-access-dxtnk" (OuterVolumeSpecName: "kube-api-access-dxtnk") pod "77f658e2-b468-45dd-8204-691cdfe9bf61" (UID: "77f658e2-b468-45dd-8204-691cdfe9bf61"). InnerVolumeSpecName "kube-api-access-dxtnk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:28:48 crc kubenswrapper[4860]: I1211 08:28:48.073194 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/77f658e2-b468-45dd-8204-691cdfe9bf61-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "77f658e2-b468-45dd-8204-691cdfe9bf61" (UID: "77f658e2-b468-45dd-8204-691cdfe9bf61"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:28:48 crc kubenswrapper[4860]: I1211 08:28:48.078818 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/77f658e2-b468-45dd-8204-691cdfe9bf61-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "77f658e2-b468-45dd-8204-691cdfe9bf61" (UID: "77f658e2-b468-45dd-8204-691cdfe9bf61"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:28:48 crc kubenswrapper[4860]: I1211 08:28:48.087680 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/77f658e2-b468-45dd-8204-691cdfe9bf61-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "77f658e2-b468-45dd-8204-691cdfe9bf61" (UID: "77f658e2-b468-45dd-8204-691cdfe9bf61"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:28:48 crc kubenswrapper[4860]: I1211 08:28:48.100313 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/77f658e2-b468-45dd-8204-691cdfe9bf61-config" (OuterVolumeSpecName: "config") pod "77f658e2-b468-45dd-8204-691cdfe9bf61" (UID: "77f658e2-b468-45dd-8204-691cdfe9bf61"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:28:48 crc kubenswrapper[4860]: I1211 08:28:48.118533 4860 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/77f658e2-b468-45dd-8204-691cdfe9bf61-config\") on node \"crc\" DevicePath \"\"" Dec 11 08:28:48 crc kubenswrapper[4860]: I1211 08:28:48.118575 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dxtnk\" (UniqueName: \"kubernetes.io/projected/77f658e2-b468-45dd-8204-691cdfe9bf61-kube-api-access-dxtnk\") on node \"crc\" DevicePath \"\"" Dec 11 08:28:48 crc kubenswrapper[4860]: I1211 08:28:48.118596 4860 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/77f658e2-b468-45dd-8204-691cdfe9bf61-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 11 08:28:48 crc kubenswrapper[4860]: I1211 08:28:48.118610 4860 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/77f658e2-b468-45dd-8204-691cdfe9bf61-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 11 08:28:48 crc kubenswrapper[4860]: I1211 08:28:48.118625 4860 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/77f658e2-b468-45dd-8204-691cdfe9bf61-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 11 08:28:48 crc kubenswrapper[4860]: I1211 08:28:48.440516 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-l7vfs" Dec 11 08:28:48 crc kubenswrapper[4860]: I1211 08:28:48.440510 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-l7vfs" event={"ID":"77f658e2-b468-45dd-8204-691cdfe9bf61","Type":"ContainerDied","Data":"4696d8e509df5e9af91eb56115f4eaf5bb108a19f8ef126b9099787d80ca2d77"} Dec 11 08:28:48 crc kubenswrapper[4860]: I1211 08:28:48.440747 4860 scope.go:117] "RemoveContainer" containerID="aca4cdc244eaa6aa25f8078b98879c5acf09174e9e804d8ee9f36118d2bd6cdb" Dec 11 08:28:48 crc kubenswrapper[4860]: I1211 08:28:48.450167 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-970d-account-create-update-65w8k" Dec 11 08:28:48 crc kubenswrapper[4860]: I1211 08:28:48.450199 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-970d-account-create-update-65w8k" event={"ID":"5af63911-2f28-40ca-9d02-721f5871f52d","Type":"ContainerDied","Data":"5ce211da1331a0b7ee47c2e2d9f00f5c0da183761829855e6b02b964905c7623"} Dec 11 08:28:48 crc kubenswrapper[4860]: I1211 08:28:48.450259 4860 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5ce211da1331a0b7ee47c2e2d9f00f5c0da183761829855e6b02b964905c7623" Dec 11 08:28:48 crc kubenswrapper[4860]: I1211 08:28:48.453987 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" event={"ID":"31c30642-6e60-41b4-a477-0d802424e0aa","Type":"ContainerStarted","Data":"6be7e2592a74fa92891c9f53dda9c3fdad0d78e647aef9f0b578adb34ea14caa"} Dec 11 08:28:48 crc kubenswrapper[4860]: I1211 08:28:48.457243 4860 generic.go:334] "Generic (PLEG): container finished" podID="e8941d7b-8d06-4ec6-8da3-101e001f824f" containerID="cf5a7e627cf219a9c4504179ba2642a4118fe67030941daa18f9ca7063564d27" exitCode=0 Dec 11 08:28:48 crc kubenswrapper[4860]: I1211 08:28:48.457356 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-6822-account-create-update-tjqd7" event={"ID":"e8941d7b-8d06-4ec6-8da3-101e001f824f","Type":"ContainerDied","Data":"cf5a7e627cf219a9c4504179ba2642a4118fe67030941daa18f9ca7063564d27"} Dec 11 08:28:48 crc kubenswrapper[4860]: I1211 08:28:48.457406 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-6822-account-create-update-tjqd7" event={"ID":"e8941d7b-8d06-4ec6-8da3-101e001f824f","Type":"ContainerStarted","Data":"6b719eb1082e2b08eb928bb0c8558065855e353ce4b9285483147626e7de2b87"} Dec 11 08:28:48 crc kubenswrapper[4860]: I1211 08:28:48.460988 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-c4mc4" Dec 11 08:28:48 crc kubenswrapper[4860]: I1211 08:28:48.460991 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-c4mc4" event={"ID":"7bea23d9-3be0-4a68-8ec1-8f21bd7feb80","Type":"ContainerDied","Data":"ab608a8a6e34f7731f78a9efaac355fb0691043bb294a9980b57271b4a79aa35"} Dec 11 08:28:48 crc kubenswrapper[4860]: I1211 08:28:48.461132 4860 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ab608a8a6e34f7731f78a9efaac355fb0691043bb294a9980b57271b4a79aa35" Dec 11 08:28:48 crc kubenswrapper[4860]: I1211 08:28:48.463293 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-7567-account-create-update-5mzbx" event={"ID":"54f31848-ee29-43be-8363-0242dbe36164","Type":"ContainerDied","Data":"19228c8b2300ccad4375e6b3832ce28cfe988a9c7be9cda6f79c1cbf4259351c"} Dec 11 08:28:48 crc kubenswrapper[4860]: I1211 08:28:48.463368 4860 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="19228c8b2300ccad4375e6b3832ce28cfe988a9c7be9cda6f79c1cbf4259351c" Dec 11 08:28:48 crc kubenswrapper[4860]: I1211 08:28:48.463480 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-7567-account-create-update-5mzbx" Dec 11 08:28:48 crc kubenswrapper[4860]: I1211 08:28:48.481261 4860 generic.go:334] "Generic (PLEG): container finished" podID="989e314b-f91d-4ceb-ba16-79d180225afe" containerID="732cd87f4f33757f7f25275c3c1bafe341bdbdc36956a4339205959a3b10edbd" exitCode=0 Dec 11 08:28:48 crc kubenswrapper[4860]: I1211 08:28:48.482034 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-dqvkz" Dec 11 08:28:48 crc kubenswrapper[4860]: I1211 08:28:48.483151 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-lrszm" event={"ID":"989e314b-f91d-4ceb-ba16-79d180225afe","Type":"ContainerDied","Data":"732cd87f4f33757f7f25275c3c1bafe341bdbdc36956a4339205959a3b10edbd"} Dec 11 08:28:48 crc kubenswrapper[4860]: I1211 08:28:48.483202 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-lrszm" event={"ID":"989e314b-f91d-4ceb-ba16-79d180225afe","Type":"ContainerStarted","Data":"ab91588af10cd77e03efd539d649d16014e97eaac5b312ab801d44a2949bf664"} Dec 11 08:28:48 crc kubenswrapper[4860]: I1211 08:28:48.485269 4860 scope.go:117] "RemoveContainer" containerID="49c890dc4873c5f72a9b92e00eabd501e06c2c566ccf2eeab6288d19dddd536b" Dec 11 08:28:48 crc kubenswrapper[4860]: E1211 08:28:48.490966 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-glance-api:current-podified\\\"\"" pod="openstack/glance-db-sync-4fsqr" podUID="cf433d5d-8c95-4d72-864c-3b74f315dd2b" Dec 11 08:28:48 crc kubenswrapper[4860]: I1211 08:28:48.555093 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-l7vfs"] Dec 11 08:28:48 crc kubenswrapper[4860]: I1211 08:28:48.564856 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-l7vfs"] Dec 11 08:28:49 crc kubenswrapper[4860]: I1211 08:28:49.612203 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="77f658e2-b468-45dd-8204-691cdfe9bf61" path="/var/lib/kubelet/pods/77f658e2-b468-45dd-8204-691cdfe9bf61/volumes" Dec 11 08:28:52 crc kubenswrapper[4860]: I1211 08:28:52.190667 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-6822-account-create-update-tjqd7" Dec 11 08:28:52 crc kubenswrapper[4860]: I1211 08:28:52.222200 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-lrszm" Dec 11 08:28:52 crc kubenswrapper[4860]: I1211 08:28:52.318278 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/989e314b-f91d-4ceb-ba16-79d180225afe-operator-scripts\") pod \"989e314b-f91d-4ceb-ba16-79d180225afe\" (UID: \"989e314b-f91d-4ceb-ba16-79d180225afe\") " Dec 11 08:28:52 crc kubenswrapper[4860]: I1211 08:28:52.318877 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ks4ln\" (UniqueName: \"kubernetes.io/projected/989e314b-f91d-4ceb-ba16-79d180225afe-kube-api-access-ks4ln\") pod \"989e314b-f91d-4ceb-ba16-79d180225afe\" (UID: \"989e314b-f91d-4ceb-ba16-79d180225afe\") " Dec 11 08:28:52 crc kubenswrapper[4860]: I1211 08:28:52.319108 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e8941d7b-8d06-4ec6-8da3-101e001f824f-operator-scripts\") pod \"e8941d7b-8d06-4ec6-8da3-101e001f824f\" (UID: \"e8941d7b-8d06-4ec6-8da3-101e001f824f\") " Dec 11 08:28:52 crc kubenswrapper[4860]: I1211 08:28:52.319301 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sl594\" (UniqueName: \"kubernetes.io/projected/e8941d7b-8d06-4ec6-8da3-101e001f824f-kube-api-access-sl594\") pod \"e8941d7b-8d06-4ec6-8da3-101e001f824f\" (UID: \"e8941d7b-8d06-4ec6-8da3-101e001f824f\") " Dec 11 08:28:52 crc kubenswrapper[4860]: I1211 08:28:52.319681 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/989e314b-f91d-4ceb-ba16-79d180225afe-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "989e314b-f91d-4ceb-ba16-79d180225afe" (UID: "989e314b-f91d-4ceb-ba16-79d180225afe"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:28:52 crc kubenswrapper[4860]: I1211 08:28:52.320314 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e8941d7b-8d06-4ec6-8da3-101e001f824f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e8941d7b-8d06-4ec6-8da3-101e001f824f" (UID: "e8941d7b-8d06-4ec6-8da3-101e001f824f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:28:52 crc kubenswrapper[4860]: I1211 08:28:52.320480 4860 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/989e314b-f91d-4ceb-ba16-79d180225afe-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 08:28:52 crc kubenswrapper[4860]: I1211 08:28:52.324864 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/989e314b-f91d-4ceb-ba16-79d180225afe-kube-api-access-ks4ln" (OuterVolumeSpecName: "kube-api-access-ks4ln") pod "989e314b-f91d-4ceb-ba16-79d180225afe" (UID: "989e314b-f91d-4ceb-ba16-79d180225afe"). InnerVolumeSpecName "kube-api-access-ks4ln". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:28:52 crc kubenswrapper[4860]: I1211 08:28:52.324983 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e8941d7b-8d06-4ec6-8da3-101e001f824f-kube-api-access-sl594" (OuterVolumeSpecName: "kube-api-access-sl594") pod "e8941d7b-8d06-4ec6-8da3-101e001f824f" (UID: "e8941d7b-8d06-4ec6-8da3-101e001f824f"). InnerVolumeSpecName "kube-api-access-sl594". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:28:52 crc kubenswrapper[4860]: I1211 08:28:52.423169 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ks4ln\" (UniqueName: \"kubernetes.io/projected/989e314b-f91d-4ceb-ba16-79d180225afe-kube-api-access-ks4ln\") on node \"crc\" DevicePath \"\"" Dec 11 08:28:52 crc kubenswrapper[4860]: I1211 08:28:52.423257 4860 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e8941d7b-8d06-4ec6-8da3-101e001f824f-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 08:28:52 crc kubenswrapper[4860]: I1211 08:28:52.423273 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sl594\" (UniqueName: \"kubernetes.io/projected/e8941d7b-8d06-4ec6-8da3-101e001f824f-kube-api-access-sl594\") on node \"crc\" DevicePath \"\"" Dec 11 08:28:52 crc kubenswrapper[4860]: I1211 08:28:52.526292 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-6822-account-create-update-tjqd7" event={"ID":"e8941d7b-8d06-4ec6-8da3-101e001f824f","Type":"ContainerDied","Data":"6b719eb1082e2b08eb928bb0c8558065855e353ce4b9285483147626e7de2b87"} Dec 11 08:28:52 crc kubenswrapper[4860]: I1211 08:28:52.526932 4860 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6b719eb1082e2b08eb928bb0c8558065855e353ce4b9285483147626e7de2b87" Dec 11 08:28:52 crc kubenswrapper[4860]: I1211 08:28:52.526383 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-6822-account-create-update-tjqd7" Dec 11 08:28:52 crc kubenswrapper[4860]: I1211 08:28:52.527745 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-wkrv5" event={"ID":"c602556f-fbc0-4e66-bd77-0d5b029e81bb","Type":"ContainerStarted","Data":"452366f5a3dcc497c0b8da7d4128cab636c136a87b2ae2638147a11767c43545"} Dec 11 08:28:52 crc kubenswrapper[4860]: I1211 08:28:52.531235 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-lrszm" event={"ID":"989e314b-f91d-4ceb-ba16-79d180225afe","Type":"ContainerDied","Data":"ab91588af10cd77e03efd539d649d16014e97eaac5b312ab801d44a2949bf664"} Dec 11 08:28:52 crc kubenswrapper[4860]: I1211 08:28:52.531261 4860 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ab91588af10cd77e03efd539d649d16014e97eaac5b312ab801d44a2949bf664" Dec 11 08:28:52 crc kubenswrapper[4860]: I1211 08:28:52.531322 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-lrszm" Dec 11 08:28:52 crc kubenswrapper[4860]: I1211 08:28:52.553851 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-wkrv5" podStartSLOduration=2.627064977 podStartE2EDuration="16.553828449s" podCreationTimestamp="2025-12-11 08:28:36 +0000 UTC" firstStartedPulling="2025-12-11 08:28:38.139771451 +0000 UTC m=+1050.868290496" lastFinishedPulling="2025-12-11 08:28:52.066534913 +0000 UTC m=+1064.795053968" observedRunningTime="2025-12-11 08:28:52.547014204 +0000 UTC m=+1065.275533259" watchObservedRunningTime="2025-12-11 08:28:52.553828449 +0000 UTC m=+1065.282347504" Dec 11 08:28:55 crc kubenswrapper[4860]: I1211 08:28:55.575085 4860 generic.go:334] "Generic (PLEG): container finished" podID="c602556f-fbc0-4e66-bd77-0d5b029e81bb" containerID="452366f5a3dcc497c0b8da7d4128cab636c136a87b2ae2638147a11767c43545" exitCode=0 Dec 11 08:28:55 crc kubenswrapper[4860]: I1211 08:28:55.575183 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-wkrv5" event={"ID":"c602556f-fbc0-4e66-bd77-0d5b029e81bb","Type":"ContainerDied","Data":"452366f5a3dcc497c0b8da7d4128cab636c136a87b2ae2638147a11767c43545"} Dec 11 08:28:56 crc kubenswrapper[4860]: I1211 08:28:56.941183 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-wkrv5" Dec 11 08:28:57 crc kubenswrapper[4860]: I1211 08:28:57.053059 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c602556f-fbc0-4e66-bd77-0d5b029e81bb-combined-ca-bundle\") pod \"c602556f-fbc0-4e66-bd77-0d5b029e81bb\" (UID: \"c602556f-fbc0-4e66-bd77-0d5b029e81bb\") " Dec 11 08:28:57 crc kubenswrapper[4860]: I1211 08:28:57.053169 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c602556f-fbc0-4e66-bd77-0d5b029e81bb-config-data\") pod \"c602556f-fbc0-4e66-bd77-0d5b029e81bb\" (UID: \"c602556f-fbc0-4e66-bd77-0d5b029e81bb\") " Dec 11 08:28:57 crc kubenswrapper[4860]: I1211 08:28:57.053467 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gx2pt\" (UniqueName: \"kubernetes.io/projected/c602556f-fbc0-4e66-bd77-0d5b029e81bb-kube-api-access-gx2pt\") pod \"c602556f-fbc0-4e66-bd77-0d5b029e81bb\" (UID: \"c602556f-fbc0-4e66-bd77-0d5b029e81bb\") " Dec 11 08:28:57 crc kubenswrapper[4860]: I1211 08:28:57.063928 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c602556f-fbc0-4e66-bd77-0d5b029e81bb-kube-api-access-gx2pt" (OuterVolumeSpecName: "kube-api-access-gx2pt") pod "c602556f-fbc0-4e66-bd77-0d5b029e81bb" (UID: "c602556f-fbc0-4e66-bd77-0d5b029e81bb"). InnerVolumeSpecName "kube-api-access-gx2pt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:28:57 crc kubenswrapper[4860]: I1211 08:28:57.122218 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c602556f-fbc0-4e66-bd77-0d5b029e81bb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c602556f-fbc0-4e66-bd77-0d5b029e81bb" (UID: "c602556f-fbc0-4e66-bd77-0d5b029e81bb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:28:57 crc kubenswrapper[4860]: I1211 08:28:57.128883 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c602556f-fbc0-4e66-bd77-0d5b029e81bb-config-data" (OuterVolumeSpecName: "config-data") pod "c602556f-fbc0-4e66-bd77-0d5b029e81bb" (UID: "c602556f-fbc0-4e66-bd77-0d5b029e81bb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:28:57 crc kubenswrapper[4860]: I1211 08:28:57.156337 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gx2pt\" (UniqueName: \"kubernetes.io/projected/c602556f-fbc0-4e66-bd77-0d5b029e81bb-kube-api-access-gx2pt\") on node \"crc\" DevicePath \"\"" Dec 11 08:28:57 crc kubenswrapper[4860]: I1211 08:28:57.156394 4860 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c602556f-fbc0-4e66-bd77-0d5b029e81bb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 08:28:57 crc kubenswrapper[4860]: I1211 08:28:57.156408 4860 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c602556f-fbc0-4e66-bd77-0d5b029e81bb-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 08:28:57 crc kubenswrapper[4860]: I1211 08:28:57.598447 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-wkrv5" event={"ID":"c602556f-fbc0-4e66-bd77-0d5b029e81bb","Type":"ContainerDied","Data":"2495d53ebd4963320e8049c2dfa6ebcbefbff1951a71652a285f1b7aca5fd57a"} Dec 11 08:28:57 crc kubenswrapper[4860]: I1211 08:28:57.598500 4860 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2495d53ebd4963320e8049c2dfa6ebcbefbff1951a71652a285f1b7aca5fd57a" Dec 11 08:28:57 crc kubenswrapper[4860]: I1211 08:28:57.598510 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-wkrv5" Dec 11 08:28:57 crc kubenswrapper[4860]: I1211 08:28:57.842270 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6f8c45789f-zdxj7"] Dec 11 08:28:57 crc kubenswrapper[4860]: E1211 08:28:57.842687 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77f658e2-b468-45dd-8204-691cdfe9bf61" containerName="init" Dec 11 08:28:57 crc kubenswrapper[4860]: I1211 08:28:57.842705 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="77f658e2-b468-45dd-8204-691cdfe9bf61" containerName="init" Dec 11 08:28:57 crc kubenswrapper[4860]: E1211 08:28:57.842724 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7bea23d9-3be0-4a68-8ec1-8f21bd7feb80" containerName="mariadb-database-create" Dec 11 08:28:57 crc kubenswrapper[4860]: I1211 08:28:57.842731 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="7bea23d9-3be0-4a68-8ec1-8f21bd7feb80" containerName="mariadb-database-create" Dec 11 08:28:57 crc kubenswrapper[4860]: E1211 08:28:57.842747 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8941d7b-8d06-4ec6-8da3-101e001f824f" containerName="mariadb-account-create-update" Dec 11 08:28:57 crc kubenswrapper[4860]: I1211 08:28:57.842754 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8941d7b-8d06-4ec6-8da3-101e001f824f" containerName="mariadb-account-create-update" Dec 11 08:28:57 crc kubenswrapper[4860]: E1211 08:28:57.842766 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c602556f-fbc0-4e66-bd77-0d5b029e81bb" containerName="keystone-db-sync" Dec 11 08:28:57 crc kubenswrapper[4860]: I1211 08:28:57.842772 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="c602556f-fbc0-4e66-bd77-0d5b029e81bb" containerName="keystone-db-sync" Dec 11 08:28:57 crc kubenswrapper[4860]: E1211 08:28:57.842783 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77f658e2-b468-45dd-8204-691cdfe9bf61" containerName="dnsmasq-dns" Dec 11 08:28:57 crc kubenswrapper[4860]: I1211 08:28:57.842789 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="77f658e2-b468-45dd-8204-691cdfe9bf61" containerName="dnsmasq-dns" Dec 11 08:28:57 crc kubenswrapper[4860]: E1211 08:28:57.842803 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5af63911-2f28-40ca-9d02-721f5871f52d" containerName="mariadb-account-create-update" Dec 11 08:28:57 crc kubenswrapper[4860]: I1211 08:28:57.842810 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="5af63911-2f28-40ca-9d02-721f5871f52d" containerName="mariadb-account-create-update" Dec 11 08:28:57 crc kubenswrapper[4860]: E1211 08:28:57.842824 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54f31848-ee29-43be-8363-0242dbe36164" containerName="mariadb-account-create-update" Dec 11 08:28:57 crc kubenswrapper[4860]: I1211 08:28:57.842829 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="54f31848-ee29-43be-8363-0242dbe36164" containerName="mariadb-account-create-update" Dec 11 08:28:57 crc kubenswrapper[4860]: E1211 08:28:57.842838 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07fbfcf2-916b-4cd8-b603-48c55736a6b0" containerName="mariadb-database-create" Dec 11 08:28:57 crc kubenswrapper[4860]: I1211 08:28:57.842846 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="07fbfcf2-916b-4cd8-b603-48c55736a6b0" containerName="mariadb-database-create" Dec 11 08:28:57 crc kubenswrapper[4860]: E1211 08:28:57.842863 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="989e314b-f91d-4ceb-ba16-79d180225afe" containerName="mariadb-database-create" Dec 11 08:28:57 crc kubenswrapper[4860]: I1211 08:28:57.842870 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="989e314b-f91d-4ceb-ba16-79d180225afe" containerName="mariadb-database-create" Dec 11 08:28:57 crc kubenswrapper[4860]: I1211 08:28:57.843053 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8941d7b-8d06-4ec6-8da3-101e001f824f" containerName="mariadb-account-create-update" Dec 11 08:28:57 crc kubenswrapper[4860]: I1211 08:28:57.843067 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="77f658e2-b468-45dd-8204-691cdfe9bf61" containerName="dnsmasq-dns" Dec 11 08:28:57 crc kubenswrapper[4860]: I1211 08:28:57.843073 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="c602556f-fbc0-4e66-bd77-0d5b029e81bb" containerName="keystone-db-sync" Dec 11 08:28:57 crc kubenswrapper[4860]: I1211 08:28:57.843083 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="54f31848-ee29-43be-8363-0242dbe36164" containerName="mariadb-account-create-update" Dec 11 08:28:57 crc kubenswrapper[4860]: I1211 08:28:57.843093 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="5af63911-2f28-40ca-9d02-721f5871f52d" containerName="mariadb-account-create-update" Dec 11 08:28:57 crc kubenswrapper[4860]: I1211 08:28:57.843105 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="07fbfcf2-916b-4cd8-b603-48c55736a6b0" containerName="mariadb-database-create" Dec 11 08:28:57 crc kubenswrapper[4860]: I1211 08:28:57.843117 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="989e314b-f91d-4ceb-ba16-79d180225afe" containerName="mariadb-database-create" Dec 11 08:28:57 crc kubenswrapper[4860]: I1211 08:28:57.843127 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="7bea23d9-3be0-4a68-8ec1-8f21bd7feb80" containerName="mariadb-database-create" Dec 11 08:28:57 crc kubenswrapper[4860]: I1211 08:28:57.844094 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f8c45789f-zdxj7" Dec 11 08:28:57 crc kubenswrapper[4860]: I1211 08:28:57.861824 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6f8c45789f-zdxj7"] Dec 11 08:28:57 crc kubenswrapper[4860]: I1211 08:28:57.884324 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-xq969"] Dec 11 08:28:57 crc kubenswrapper[4860]: I1211 08:28:57.886395 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-xq969" Dec 11 08:28:57 crc kubenswrapper[4860]: I1211 08:28:57.891070 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-qjrrt" Dec 11 08:28:57 crc kubenswrapper[4860]: I1211 08:28:57.891244 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 11 08:28:57 crc kubenswrapper[4860]: I1211 08:28:57.891253 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 11 08:28:57 crc kubenswrapper[4860]: I1211 08:28:57.891389 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 11 08:28:57 crc kubenswrapper[4860]: I1211 08:28:57.891989 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 11 08:28:57 crc kubenswrapper[4860]: I1211 08:28:57.901761 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-xq969"] Dec 11 08:28:57 crc kubenswrapper[4860]: I1211 08:28:57.997367 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/232401ac-edc1-41a7-a802-a744276304c8-config\") pod \"dnsmasq-dns-6f8c45789f-zdxj7\" (UID: \"232401ac-edc1-41a7-a802-a744276304c8\") " pod="openstack/dnsmasq-dns-6f8c45789f-zdxj7" Dec 11 08:28:57 crc kubenswrapper[4860]: I1211 08:28:57.998478 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a7bc1b0c-9f61-4bd1-8739-6219931987fa-credential-keys\") pod \"keystone-bootstrap-xq969\" (UID: \"a7bc1b0c-9f61-4bd1-8739-6219931987fa\") " pod="openstack/keystone-bootstrap-xq969" Dec 11 08:28:57 crc kubenswrapper[4860]: I1211 08:28:57.998684 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/232401ac-edc1-41a7-a802-a744276304c8-dns-swift-storage-0\") pod \"dnsmasq-dns-6f8c45789f-zdxj7\" (UID: \"232401ac-edc1-41a7-a802-a744276304c8\") " pod="openstack/dnsmasq-dns-6f8c45789f-zdxj7" Dec 11 08:28:57 crc kubenswrapper[4860]: I1211 08:28:57.998795 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/232401ac-edc1-41a7-a802-a744276304c8-ovsdbserver-nb\") pod \"dnsmasq-dns-6f8c45789f-zdxj7\" (UID: \"232401ac-edc1-41a7-a802-a744276304c8\") " pod="openstack/dnsmasq-dns-6f8c45789f-zdxj7" Dec 11 08:28:57 crc kubenswrapper[4860]: I1211 08:28:57.998907 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6lxlg\" (UniqueName: \"kubernetes.io/projected/232401ac-edc1-41a7-a802-a744276304c8-kube-api-access-6lxlg\") pod \"dnsmasq-dns-6f8c45789f-zdxj7\" (UID: \"232401ac-edc1-41a7-a802-a744276304c8\") " pod="openstack/dnsmasq-dns-6f8c45789f-zdxj7" Dec 11 08:28:57 crc kubenswrapper[4860]: I1211 08:28:57.999043 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-75456\" (UniqueName: \"kubernetes.io/projected/a7bc1b0c-9f61-4bd1-8739-6219931987fa-kube-api-access-75456\") pod \"keystone-bootstrap-xq969\" (UID: \"a7bc1b0c-9f61-4bd1-8739-6219931987fa\") " pod="openstack/keystone-bootstrap-xq969" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.000871 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a7bc1b0c-9f61-4bd1-8739-6219931987fa-scripts\") pod \"keystone-bootstrap-xq969\" (UID: \"a7bc1b0c-9f61-4bd1-8739-6219931987fa\") " pod="openstack/keystone-bootstrap-xq969" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.002756 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7bc1b0c-9f61-4bd1-8739-6219931987fa-combined-ca-bundle\") pod \"keystone-bootstrap-xq969\" (UID: \"a7bc1b0c-9f61-4bd1-8739-6219931987fa\") " pod="openstack/keystone-bootstrap-xq969" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.002913 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a7bc1b0c-9f61-4bd1-8739-6219931987fa-fernet-keys\") pod \"keystone-bootstrap-xq969\" (UID: \"a7bc1b0c-9f61-4bd1-8739-6219931987fa\") " pod="openstack/keystone-bootstrap-xq969" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.003155 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/232401ac-edc1-41a7-a802-a744276304c8-dns-svc\") pod \"dnsmasq-dns-6f8c45789f-zdxj7\" (UID: \"232401ac-edc1-41a7-a802-a744276304c8\") " pod="openstack/dnsmasq-dns-6f8c45789f-zdxj7" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.003256 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/232401ac-edc1-41a7-a802-a744276304c8-ovsdbserver-sb\") pod \"dnsmasq-dns-6f8c45789f-zdxj7\" (UID: \"232401ac-edc1-41a7-a802-a744276304c8\") " pod="openstack/dnsmasq-dns-6f8c45789f-zdxj7" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.003501 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7bc1b0c-9f61-4bd1-8739-6219931987fa-config-data\") pod \"keystone-bootstrap-xq969\" (UID: \"a7bc1b0c-9f61-4bd1-8739-6219931987fa\") " pod="openstack/keystone-bootstrap-xq969" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.094975 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-hccts"] Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.096403 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-hccts" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.105280 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.105544 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-8nqnh" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.105707 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.106985 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-75456\" (UniqueName: \"kubernetes.io/projected/a7bc1b0c-9f61-4bd1-8739-6219931987fa-kube-api-access-75456\") pod \"keystone-bootstrap-xq969\" (UID: \"a7bc1b0c-9f61-4bd1-8739-6219931987fa\") " pod="openstack/keystone-bootstrap-xq969" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.107024 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d472d27e-9bc9-4d99-8fd9-697f7f44f57d-config-data\") pod \"cinder-db-sync-hccts\" (UID: \"d472d27e-9bc9-4d99-8fd9-697f7f44f57d\") " pod="openstack/cinder-db-sync-hccts" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.107050 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a7bc1b0c-9f61-4bd1-8739-6219931987fa-scripts\") pod \"keystone-bootstrap-xq969\" (UID: \"a7bc1b0c-9f61-4bd1-8739-6219931987fa\") " pod="openstack/keystone-bootstrap-xq969" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.107075 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d472d27e-9bc9-4d99-8fd9-697f7f44f57d-combined-ca-bundle\") pod \"cinder-db-sync-hccts\" (UID: \"d472d27e-9bc9-4d99-8fd9-697f7f44f57d\") " pod="openstack/cinder-db-sync-hccts" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.107115 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7bc1b0c-9f61-4bd1-8739-6219931987fa-combined-ca-bundle\") pod \"keystone-bootstrap-xq969\" (UID: \"a7bc1b0c-9f61-4bd1-8739-6219931987fa\") " pod="openstack/keystone-bootstrap-xq969" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.107138 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a7bc1b0c-9f61-4bd1-8739-6219931987fa-fernet-keys\") pod \"keystone-bootstrap-xq969\" (UID: \"a7bc1b0c-9f61-4bd1-8739-6219931987fa\") " pod="openstack/keystone-bootstrap-xq969" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.107166 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-84jp8\" (UniqueName: \"kubernetes.io/projected/d472d27e-9bc9-4d99-8fd9-697f7f44f57d-kube-api-access-84jp8\") pod \"cinder-db-sync-hccts\" (UID: \"d472d27e-9bc9-4d99-8fd9-697f7f44f57d\") " pod="openstack/cinder-db-sync-hccts" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.107218 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/232401ac-edc1-41a7-a802-a744276304c8-dns-svc\") pod \"dnsmasq-dns-6f8c45789f-zdxj7\" (UID: \"232401ac-edc1-41a7-a802-a744276304c8\") " pod="openstack/dnsmasq-dns-6f8c45789f-zdxj7" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.107241 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/232401ac-edc1-41a7-a802-a744276304c8-ovsdbserver-sb\") pod \"dnsmasq-dns-6f8c45789f-zdxj7\" (UID: \"232401ac-edc1-41a7-a802-a744276304c8\") " pod="openstack/dnsmasq-dns-6f8c45789f-zdxj7" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.107281 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7bc1b0c-9f61-4bd1-8739-6219931987fa-config-data\") pod \"keystone-bootstrap-xq969\" (UID: \"a7bc1b0c-9f61-4bd1-8739-6219931987fa\") " pod="openstack/keystone-bootstrap-xq969" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.107307 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d472d27e-9bc9-4d99-8fd9-697f7f44f57d-scripts\") pod \"cinder-db-sync-hccts\" (UID: \"d472d27e-9bc9-4d99-8fd9-697f7f44f57d\") " pod="openstack/cinder-db-sync-hccts" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.107328 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d472d27e-9bc9-4d99-8fd9-697f7f44f57d-db-sync-config-data\") pod \"cinder-db-sync-hccts\" (UID: \"d472d27e-9bc9-4d99-8fd9-697f7f44f57d\") " pod="openstack/cinder-db-sync-hccts" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.107360 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/232401ac-edc1-41a7-a802-a744276304c8-config\") pod \"dnsmasq-dns-6f8c45789f-zdxj7\" (UID: \"232401ac-edc1-41a7-a802-a744276304c8\") " pod="openstack/dnsmasq-dns-6f8c45789f-zdxj7" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.107386 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a7bc1b0c-9f61-4bd1-8739-6219931987fa-credential-keys\") pod \"keystone-bootstrap-xq969\" (UID: \"a7bc1b0c-9f61-4bd1-8739-6219931987fa\") " pod="openstack/keystone-bootstrap-xq969" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.107406 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d472d27e-9bc9-4d99-8fd9-697f7f44f57d-etc-machine-id\") pod \"cinder-db-sync-hccts\" (UID: \"d472d27e-9bc9-4d99-8fd9-697f7f44f57d\") " pod="openstack/cinder-db-sync-hccts" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.107434 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/232401ac-edc1-41a7-a802-a744276304c8-dns-swift-storage-0\") pod \"dnsmasq-dns-6f8c45789f-zdxj7\" (UID: \"232401ac-edc1-41a7-a802-a744276304c8\") " pod="openstack/dnsmasq-dns-6f8c45789f-zdxj7" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.107454 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/232401ac-edc1-41a7-a802-a744276304c8-ovsdbserver-nb\") pod \"dnsmasq-dns-6f8c45789f-zdxj7\" (UID: \"232401ac-edc1-41a7-a802-a744276304c8\") " pod="openstack/dnsmasq-dns-6f8c45789f-zdxj7" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.107480 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6lxlg\" (UniqueName: \"kubernetes.io/projected/232401ac-edc1-41a7-a802-a744276304c8-kube-api-access-6lxlg\") pod \"dnsmasq-dns-6f8c45789f-zdxj7\" (UID: \"232401ac-edc1-41a7-a802-a744276304c8\") " pod="openstack/dnsmasq-dns-6f8c45789f-zdxj7" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.122877 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/232401ac-edc1-41a7-a802-a744276304c8-config\") pod \"dnsmasq-dns-6f8c45789f-zdxj7\" (UID: \"232401ac-edc1-41a7-a802-a744276304c8\") " pod="openstack/dnsmasq-dns-6f8c45789f-zdxj7" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.123290 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/232401ac-edc1-41a7-a802-a744276304c8-dns-svc\") pod \"dnsmasq-dns-6f8c45789f-zdxj7\" (UID: \"232401ac-edc1-41a7-a802-a744276304c8\") " pod="openstack/dnsmasq-dns-6f8c45789f-zdxj7" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.123478 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/232401ac-edc1-41a7-a802-a744276304c8-ovsdbserver-sb\") pod \"dnsmasq-dns-6f8c45789f-zdxj7\" (UID: \"232401ac-edc1-41a7-a802-a744276304c8\") " pod="openstack/dnsmasq-dns-6f8c45789f-zdxj7" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.124052 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/232401ac-edc1-41a7-a802-a744276304c8-ovsdbserver-nb\") pod \"dnsmasq-dns-6f8c45789f-zdxj7\" (UID: \"232401ac-edc1-41a7-a802-a744276304c8\") " pod="openstack/dnsmasq-dns-6f8c45789f-zdxj7" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.125187 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/232401ac-edc1-41a7-a802-a744276304c8-dns-swift-storage-0\") pod \"dnsmasq-dns-6f8c45789f-zdxj7\" (UID: \"232401ac-edc1-41a7-a802-a744276304c8\") " pod="openstack/dnsmasq-dns-6f8c45789f-zdxj7" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.137727 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-hccts"] Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.154279 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.156566 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a7bc1b0c-9f61-4bd1-8739-6219931987fa-scripts\") pod \"keystone-bootstrap-xq969\" (UID: \"a7bc1b0c-9f61-4bd1-8739-6219931987fa\") " pod="openstack/keystone-bootstrap-xq969" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.158357 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a7bc1b0c-9f61-4bd1-8739-6219931987fa-credential-keys\") pod \"keystone-bootstrap-xq969\" (UID: \"a7bc1b0c-9f61-4bd1-8739-6219931987fa\") " pod="openstack/keystone-bootstrap-xq969" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.161146 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a7bc1b0c-9f61-4bd1-8739-6219931987fa-fernet-keys\") pod \"keystone-bootstrap-xq969\" (UID: \"a7bc1b0c-9f61-4bd1-8739-6219931987fa\") " pod="openstack/keystone-bootstrap-xq969" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.181616 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7bc1b0c-9f61-4bd1-8739-6219931987fa-combined-ca-bundle\") pod \"keystone-bootstrap-xq969\" (UID: \"a7bc1b0c-9f61-4bd1-8739-6219931987fa\") " pod="openstack/keystone-bootstrap-xq969" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.184878 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7bc1b0c-9f61-4bd1-8739-6219931987fa-config-data\") pod \"keystone-bootstrap-xq969\" (UID: \"a7bc1b0c-9f61-4bd1-8739-6219931987fa\") " pod="openstack/keystone-bootstrap-xq969" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.187707 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.196034 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-75456\" (UniqueName: \"kubernetes.io/projected/a7bc1b0c-9f61-4bd1-8739-6219931987fa-kube-api-access-75456\") pod \"keystone-bootstrap-xq969\" (UID: \"a7bc1b0c-9f61-4bd1-8739-6219931987fa\") " pod="openstack/keystone-bootstrap-xq969" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.201825 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.202152 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.209487 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-84jp8\" (UniqueName: \"kubernetes.io/projected/d472d27e-9bc9-4d99-8fd9-697f7f44f57d-kube-api-access-84jp8\") pod \"cinder-db-sync-hccts\" (UID: \"d472d27e-9bc9-4d99-8fd9-697f7f44f57d\") " pod="openstack/cinder-db-sync-hccts" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.209578 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vdbx8\" (UniqueName: \"kubernetes.io/projected/040a0e8c-efd3-4e3b-be51-59e311ae4406-kube-api-access-vdbx8\") pod \"ceilometer-0\" (UID: \"040a0e8c-efd3-4e3b-be51-59e311ae4406\") " pod="openstack/ceilometer-0" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.209617 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/040a0e8c-efd3-4e3b-be51-59e311ae4406-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"040a0e8c-efd3-4e3b-be51-59e311ae4406\") " pod="openstack/ceilometer-0" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.209697 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d472d27e-9bc9-4d99-8fd9-697f7f44f57d-scripts\") pod \"cinder-db-sync-hccts\" (UID: \"d472d27e-9bc9-4d99-8fd9-697f7f44f57d\") " pod="openstack/cinder-db-sync-hccts" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.209723 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d472d27e-9bc9-4d99-8fd9-697f7f44f57d-db-sync-config-data\") pod \"cinder-db-sync-hccts\" (UID: \"d472d27e-9bc9-4d99-8fd9-697f7f44f57d\") " pod="openstack/cinder-db-sync-hccts" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.209784 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d472d27e-9bc9-4d99-8fd9-697f7f44f57d-etc-machine-id\") pod \"cinder-db-sync-hccts\" (UID: \"d472d27e-9bc9-4d99-8fd9-697f7f44f57d\") " pod="openstack/cinder-db-sync-hccts" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.209817 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/040a0e8c-efd3-4e3b-be51-59e311ae4406-run-httpd\") pod \"ceilometer-0\" (UID: \"040a0e8c-efd3-4e3b-be51-59e311ae4406\") " pod="openstack/ceilometer-0" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.209865 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/040a0e8c-efd3-4e3b-be51-59e311ae4406-log-httpd\") pod \"ceilometer-0\" (UID: \"040a0e8c-efd3-4e3b-be51-59e311ae4406\") " pod="openstack/ceilometer-0" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.209945 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d472d27e-9bc9-4d99-8fd9-697f7f44f57d-config-data\") pod \"cinder-db-sync-hccts\" (UID: \"d472d27e-9bc9-4d99-8fd9-697f7f44f57d\") " pod="openstack/cinder-db-sync-hccts" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.209977 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/040a0e8c-efd3-4e3b-be51-59e311ae4406-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"040a0e8c-efd3-4e3b-be51-59e311ae4406\") " pod="openstack/ceilometer-0" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.210016 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d472d27e-9bc9-4d99-8fd9-697f7f44f57d-combined-ca-bundle\") pod \"cinder-db-sync-hccts\" (UID: \"d472d27e-9bc9-4d99-8fd9-697f7f44f57d\") " pod="openstack/cinder-db-sync-hccts" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.210041 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/040a0e8c-efd3-4e3b-be51-59e311ae4406-scripts\") pod \"ceilometer-0\" (UID: \"040a0e8c-efd3-4e3b-be51-59e311ae4406\") " pod="openstack/ceilometer-0" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.210072 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/040a0e8c-efd3-4e3b-be51-59e311ae4406-config-data\") pod \"ceilometer-0\" (UID: \"040a0e8c-efd3-4e3b-be51-59e311ae4406\") " pod="openstack/ceilometer-0" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.211179 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d472d27e-9bc9-4d99-8fd9-697f7f44f57d-etc-machine-id\") pod \"cinder-db-sync-hccts\" (UID: \"d472d27e-9bc9-4d99-8fd9-697f7f44f57d\") " pod="openstack/cinder-db-sync-hccts" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.215514 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d472d27e-9bc9-4d99-8fd9-697f7f44f57d-scripts\") pod \"cinder-db-sync-hccts\" (UID: \"d472d27e-9bc9-4d99-8fd9-697f7f44f57d\") " pod="openstack/cinder-db-sync-hccts" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.215741 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-768465cc7-f7gwz"] Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.222671 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6lxlg\" (UniqueName: \"kubernetes.io/projected/232401ac-edc1-41a7-a802-a744276304c8-kube-api-access-6lxlg\") pod \"dnsmasq-dns-6f8c45789f-zdxj7\" (UID: \"232401ac-edc1-41a7-a802-a744276304c8\") " pod="openstack/dnsmasq-dns-6f8c45789f-zdxj7" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.225441 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-768465cc7-f7gwz" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.228925 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.230397 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.231677 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-45n6w" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.231724 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.231904 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-xq969" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.236224 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-84jp8\" (UniqueName: \"kubernetes.io/projected/d472d27e-9bc9-4d99-8fd9-697f7f44f57d-kube-api-access-84jp8\") pod \"cinder-db-sync-hccts\" (UID: \"d472d27e-9bc9-4d99-8fd9-697f7f44f57d\") " pod="openstack/cinder-db-sync-hccts" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.243043 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d472d27e-9bc9-4d99-8fd9-697f7f44f57d-combined-ca-bundle\") pod \"cinder-db-sync-hccts\" (UID: \"d472d27e-9bc9-4d99-8fd9-697f7f44f57d\") " pod="openstack/cinder-db-sync-hccts" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.244153 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d472d27e-9bc9-4d99-8fd9-697f7f44f57d-db-sync-config-data\") pod \"cinder-db-sync-hccts\" (UID: \"d472d27e-9bc9-4d99-8fd9-697f7f44f57d\") " pod="openstack/cinder-db-sync-hccts" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.262730 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.266273 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d472d27e-9bc9-4d99-8fd9-697f7f44f57d-config-data\") pod \"cinder-db-sync-hccts\" (UID: \"d472d27e-9bc9-4d99-8fd9-697f7f44f57d\") " pod="openstack/cinder-db-sync-hccts" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.286359 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-wgk5t"] Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.287950 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-wgk5t" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.294228 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.294769 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.294899 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-dxbrv" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.306734 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-768465cc7-f7gwz"] Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.311719 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1ee42a0c-5029-44cf-ba8d-720165ac6138-logs\") pod \"horizon-768465cc7-f7gwz\" (UID: \"1ee42a0c-5029-44cf-ba8d-720165ac6138\") " pod="openstack/horizon-768465cc7-f7gwz" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.311779 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/040a0e8c-efd3-4e3b-be51-59e311ae4406-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"040a0e8c-efd3-4e3b-be51-59e311ae4406\") " pod="openstack/ceilometer-0" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.311802 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1ee42a0c-5029-44cf-ba8d-720165ac6138-scripts\") pod \"horizon-768465cc7-f7gwz\" (UID: \"1ee42a0c-5029-44cf-ba8d-720165ac6138\") " pod="openstack/horizon-768465cc7-f7gwz" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.311821 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/040a0e8c-efd3-4e3b-be51-59e311ae4406-scripts\") pod \"ceilometer-0\" (UID: \"040a0e8c-efd3-4e3b-be51-59e311ae4406\") " pod="openstack/ceilometer-0" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.311845 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/040a0e8c-efd3-4e3b-be51-59e311ae4406-config-data\") pod \"ceilometer-0\" (UID: \"040a0e8c-efd3-4e3b-be51-59e311ae4406\") " pod="openstack/ceilometer-0" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.311919 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vdbx8\" (UniqueName: \"kubernetes.io/projected/040a0e8c-efd3-4e3b-be51-59e311ae4406-kube-api-access-vdbx8\") pod \"ceilometer-0\" (UID: \"040a0e8c-efd3-4e3b-be51-59e311ae4406\") " pod="openstack/ceilometer-0" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.311943 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/040a0e8c-efd3-4e3b-be51-59e311ae4406-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"040a0e8c-efd3-4e3b-be51-59e311ae4406\") " pod="openstack/ceilometer-0" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.311967 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rdsd6\" (UniqueName: \"kubernetes.io/projected/e3822ab5-d1a7-46da-9533-21b67bd240eb-kube-api-access-rdsd6\") pod \"neutron-db-sync-wgk5t\" (UID: \"e3822ab5-d1a7-46da-9533-21b67bd240eb\") " pod="openstack/neutron-db-sync-wgk5t" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.312001 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/1ee42a0c-5029-44cf-ba8d-720165ac6138-horizon-secret-key\") pod \"horizon-768465cc7-f7gwz\" (UID: \"1ee42a0c-5029-44cf-ba8d-720165ac6138\") " pod="openstack/horizon-768465cc7-f7gwz" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.312020 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1ee42a0c-5029-44cf-ba8d-720165ac6138-config-data\") pod \"horizon-768465cc7-f7gwz\" (UID: \"1ee42a0c-5029-44cf-ba8d-720165ac6138\") " pod="openstack/horizon-768465cc7-f7gwz" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.312046 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3822ab5-d1a7-46da-9533-21b67bd240eb-combined-ca-bundle\") pod \"neutron-db-sync-wgk5t\" (UID: \"e3822ab5-d1a7-46da-9533-21b67bd240eb\") " pod="openstack/neutron-db-sync-wgk5t" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.312069 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/040a0e8c-efd3-4e3b-be51-59e311ae4406-run-httpd\") pod \"ceilometer-0\" (UID: \"040a0e8c-efd3-4e3b-be51-59e311ae4406\") " pod="openstack/ceilometer-0" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.312094 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/040a0e8c-efd3-4e3b-be51-59e311ae4406-log-httpd\") pod \"ceilometer-0\" (UID: \"040a0e8c-efd3-4e3b-be51-59e311ae4406\") " pod="openstack/ceilometer-0" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.312112 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sgw5k\" (UniqueName: \"kubernetes.io/projected/1ee42a0c-5029-44cf-ba8d-720165ac6138-kube-api-access-sgw5k\") pod \"horizon-768465cc7-f7gwz\" (UID: \"1ee42a0c-5029-44cf-ba8d-720165ac6138\") " pod="openstack/horizon-768465cc7-f7gwz" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.312135 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e3822ab5-d1a7-46da-9533-21b67bd240eb-config\") pod \"neutron-db-sync-wgk5t\" (UID: \"e3822ab5-d1a7-46da-9533-21b67bd240eb\") " pod="openstack/neutron-db-sync-wgk5t" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.313080 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/040a0e8c-efd3-4e3b-be51-59e311ae4406-run-httpd\") pod \"ceilometer-0\" (UID: \"040a0e8c-efd3-4e3b-be51-59e311ae4406\") " pod="openstack/ceilometer-0" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.313334 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/040a0e8c-efd3-4e3b-be51-59e311ae4406-log-httpd\") pod \"ceilometer-0\" (UID: \"040a0e8c-efd3-4e3b-be51-59e311ae4406\") " pod="openstack/ceilometer-0" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.318736 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/040a0e8c-efd3-4e3b-be51-59e311ae4406-config-data\") pod \"ceilometer-0\" (UID: \"040a0e8c-efd3-4e3b-be51-59e311ae4406\") " pod="openstack/ceilometer-0" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.327181 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/040a0e8c-efd3-4e3b-be51-59e311ae4406-scripts\") pod \"ceilometer-0\" (UID: \"040a0e8c-efd3-4e3b-be51-59e311ae4406\") " pod="openstack/ceilometer-0" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.327247 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-wgk5t"] Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.332456 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/040a0e8c-efd3-4e3b-be51-59e311ae4406-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"040a0e8c-efd3-4e3b-be51-59e311ae4406\") " pod="openstack/ceilometer-0" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.345863 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/040a0e8c-efd3-4e3b-be51-59e311ae4406-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"040a0e8c-efd3-4e3b-be51-59e311ae4406\") " pod="openstack/ceilometer-0" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.365788 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vdbx8\" (UniqueName: \"kubernetes.io/projected/040a0e8c-efd3-4e3b-be51-59e311ae4406-kube-api-access-vdbx8\") pod \"ceilometer-0\" (UID: \"040a0e8c-efd3-4e3b-be51-59e311ae4406\") " pod="openstack/ceilometer-0" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.380232 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-hccts" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.415172 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e3822ab5-d1a7-46da-9533-21b67bd240eb-config\") pod \"neutron-db-sync-wgk5t\" (UID: \"e3822ab5-d1a7-46da-9533-21b67bd240eb\") " pod="openstack/neutron-db-sync-wgk5t" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.415365 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1ee42a0c-5029-44cf-ba8d-720165ac6138-logs\") pod \"horizon-768465cc7-f7gwz\" (UID: \"1ee42a0c-5029-44cf-ba8d-720165ac6138\") " pod="openstack/horizon-768465cc7-f7gwz" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.416178 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1ee42a0c-5029-44cf-ba8d-720165ac6138-logs\") pod \"horizon-768465cc7-f7gwz\" (UID: \"1ee42a0c-5029-44cf-ba8d-720165ac6138\") " pod="openstack/horizon-768465cc7-f7gwz" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.417778 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1ee42a0c-5029-44cf-ba8d-720165ac6138-scripts\") pod \"horizon-768465cc7-f7gwz\" (UID: \"1ee42a0c-5029-44cf-ba8d-720165ac6138\") " pod="openstack/horizon-768465cc7-f7gwz" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.419445 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/e3822ab5-d1a7-46da-9533-21b67bd240eb-config\") pod \"neutron-db-sync-wgk5t\" (UID: \"e3822ab5-d1a7-46da-9533-21b67bd240eb\") " pod="openstack/neutron-db-sync-wgk5t" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.420413 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1ee42a0c-5029-44cf-ba8d-720165ac6138-scripts\") pod \"horizon-768465cc7-f7gwz\" (UID: \"1ee42a0c-5029-44cf-ba8d-720165ac6138\") " pod="openstack/horizon-768465cc7-f7gwz" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.427671 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.428705 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdsd6\" (UniqueName: \"kubernetes.io/projected/e3822ab5-d1a7-46da-9533-21b67bd240eb-kube-api-access-rdsd6\") pod \"neutron-db-sync-wgk5t\" (UID: \"e3822ab5-d1a7-46da-9533-21b67bd240eb\") " pod="openstack/neutron-db-sync-wgk5t" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.428762 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/1ee42a0c-5029-44cf-ba8d-720165ac6138-horizon-secret-key\") pod \"horizon-768465cc7-f7gwz\" (UID: \"1ee42a0c-5029-44cf-ba8d-720165ac6138\") " pod="openstack/horizon-768465cc7-f7gwz" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.428814 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1ee42a0c-5029-44cf-ba8d-720165ac6138-config-data\") pod \"horizon-768465cc7-f7gwz\" (UID: \"1ee42a0c-5029-44cf-ba8d-720165ac6138\") " pod="openstack/horizon-768465cc7-f7gwz" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.428854 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3822ab5-d1a7-46da-9533-21b67bd240eb-combined-ca-bundle\") pod \"neutron-db-sync-wgk5t\" (UID: \"e3822ab5-d1a7-46da-9533-21b67bd240eb\") " pod="openstack/neutron-db-sync-wgk5t" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.428908 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sgw5k\" (UniqueName: \"kubernetes.io/projected/1ee42a0c-5029-44cf-ba8d-720165ac6138-kube-api-access-sgw5k\") pod \"horizon-768465cc7-f7gwz\" (UID: \"1ee42a0c-5029-44cf-ba8d-720165ac6138\") " pod="openstack/horizon-768465cc7-f7gwz" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.432922 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1ee42a0c-5029-44cf-ba8d-720165ac6138-config-data\") pod \"horizon-768465cc7-f7gwz\" (UID: \"1ee42a0c-5029-44cf-ba8d-720165ac6138\") " pod="openstack/horizon-768465cc7-f7gwz" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.437097 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/1ee42a0c-5029-44cf-ba8d-720165ac6138-horizon-secret-key\") pod \"horizon-768465cc7-f7gwz\" (UID: \"1ee42a0c-5029-44cf-ba8d-720165ac6138\") " pod="openstack/horizon-768465cc7-f7gwz" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.443548 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3822ab5-d1a7-46da-9533-21b67bd240eb-combined-ca-bundle\") pod \"neutron-db-sync-wgk5t\" (UID: \"e3822ab5-d1a7-46da-9533-21b67bd240eb\") " pod="openstack/neutron-db-sync-wgk5t" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.463595 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdsd6\" (UniqueName: \"kubernetes.io/projected/e3822ab5-d1a7-46da-9533-21b67bd240eb-kube-api-access-rdsd6\") pod \"neutron-db-sync-wgk5t\" (UID: \"e3822ab5-d1a7-46da-9533-21b67bd240eb\") " pod="openstack/neutron-db-sync-wgk5t" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.471214 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f8c45789f-zdxj7" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.493710 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-pfz4t"] Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.495298 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-pfz4t" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.495371 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-wgk5t" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.500655 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.510481 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-pfz4t"] Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.510686 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-xpbmn" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.527390 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7cbbd9698f-ft2xt"] Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.532088 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7cbbd9698f-ft2xt" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.538597 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sgw5k\" (UniqueName: \"kubernetes.io/projected/1ee42a0c-5029-44cf-ba8d-720165ac6138-kube-api-access-sgw5k\") pod \"horizon-768465cc7-f7gwz\" (UID: \"1ee42a0c-5029-44cf-ba8d-720165ac6138\") " pod="openstack/horizon-768465cc7-f7gwz" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.545700 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4ff61ef1-6a67-4a8e-80e2-11f0154e54c6-db-sync-config-data\") pod \"barbican-db-sync-pfz4t\" (UID: \"4ff61ef1-6a67-4a8e-80e2-11f0154e54c6\") " pod="openstack/barbican-db-sync-pfz4t" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.545806 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ff61ef1-6a67-4a8e-80e2-11f0154e54c6-combined-ca-bundle\") pod \"barbican-db-sync-pfz4t\" (UID: \"4ff61ef1-6a67-4a8e-80e2-11f0154e54c6\") " pod="openstack/barbican-db-sync-pfz4t" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.545991 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x7pxv\" (UniqueName: \"kubernetes.io/projected/4ff61ef1-6a67-4a8e-80e2-11f0154e54c6-kube-api-access-x7pxv\") pod \"barbican-db-sync-pfz4t\" (UID: \"4ff61ef1-6a67-4a8e-80e2-11f0154e54c6\") " pod="openstack/barbican-db-sync-pfz4t" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.571918 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7cbbd9698f-ft2xt"] Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.637554 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6f8c45789f-zdxj7"] Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.659209 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b4a3ee06-7c42-4c4a-bce5-d75f9da8983d-scripts\") pod \"horizon-7cbbd9698f-ft2xt\" (UID: \"b4a3ee06-7c42-4c4a-bce5-d75f9da8983d\") " pod="openstack/horizon-7cbbd9698f-ft2xt" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.659301 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7wlsq\" (UniqueName: \"kubernetes.io/projected/b4a3ee06-7c42-4c4a-bce5-d75f9da8983d-kube-api-access-7wlsq\") pod \"horizon-7cbbd9698f-ft2xt\" (UID: \"b4a3ee06-7c42-4c4a-bce5-d75f9da8983d\") " pod="openstack/horizon-7cbbd9698f-ft2xt" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.659338 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x7pxv\" (UniqueName: \"kubernetes.io/projected/4ff61ef1-6a67-4a8e-80e2-11f0154e54c6-kube-api-access-x7pxv\") pod \"barbican-db-sync-pfz4t\" (UID: \"4ff61ef1-6a67-4a8e-80e2-11f0154e54c6\") " pod="openstack/barbican-db-sync-pfz4t" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.659408 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b4a3ee06-7c42-4c4a-bce5-d75f9da8983d-logs\") pod \"horizon-7cbbd9698f-ft2xt\" (UID: \"b4a3ee06-7c42-4c4a-bce5-d75f9da8983d\") " pod="openstack/horizon-7cbbd9698f-ft2xt" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.659473 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b4a3ee06-7c42-4c4a-bce5-d75f9da8983d-config-data\") pod \"horizon-7cbbd9698f-ft2xt\" (UID: \"b4a3ee06-7c42-4c4a-bce5-d75f9da8983d\") " pod="openstack/horizon-7cbbd9698f-ft2xt" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.659534 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/b4a3ee06-7c42-4c4a-bce5-d75f9da8983d-horizon-secret-key\") pod \"horizon-7cbbd9698f-ft2xt\" (UID: \"b4a3ee06-7c42-4c4a-bce5-d75f9da8983d\") " pod="openstack/horizon-7cbbd9698f-ft2xt" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.659583 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4ff61ef1-6a67-4a8e-80e2-11f0154e54c6-db-sync-config-data\") pod \"barbican-db-sync-pfz4t\" (UID: \"4ff61ef1-6a67-4a8e-80e2-11f0154e54c6\") " pod="openstack/barbican-db-sync-pfz4t" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.659627 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ff61ef1-6a67-4a8e-80e2-11f0154e54c6-combined-ca-bundle\") pod \"barbican-db-sync-pfz4t\" (UID: \"4ff61ef1-6a67-4a8e-80e2-11f0154e54c6\") " pod="openstack/barbican-db-sync-pfz4t" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.683414 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4ff61ef1-6a67-4a8e-80e2-11f0154e54c6-db-sync-config-data\") pod \"barbican-db-sync-pfz4t\" (UID: \"4ff61ef1-6a67-4a8e-80e2-11f0154e54c6\") " pod="openstack/barbican-db-sync-pfz4t" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.695145 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x7pxv\" (UniqueName: \"kubernetes.io/projected/4ff61ef1-6a67-4a8e-80e2-11f0154e54c6-kube-api-access-x7pxv\") pod \"barbican-db-sync-pfz4t\" (UID: \"4ff61ef1-6a67-4a8e-80e2-11f0154e54c6\") " pod="openstack/barbican-db-sync-pfz4t" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.718820 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-fcfdd6f9f-9vj6f"] Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.720730 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-fcfdd6f9f-9vj6f" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.730897 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ff61ef1-6a67-4a8e-80e2-11f0154e54c6-combined-ca-bundle\") pod \"barbican-db-sync-pfz4t\" (UID: \"4ff61ef1-6a67-4a8e-80e2-11f0154e54c6\") " pod="openstack/barbican-db-sync-pfz4t" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.751031 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-fcfdd6f9f-9vj6f"] Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.762822 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b4a3ee06-7c42-4c4a-bce5-d75f9da8983d-logs\") pod \"horizon-7cbbd9698f-ft2xt\" (UID: \"b4a3ee06-7c42-4c4a-bce5-d75f9da8983d\") " pod="openstack/horizon-7cbbd9698f-ft2xt" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.763419 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7wlsq\" (UniqueName: \"kubernetes.io/projected/b4a3ee06-7c42-4c4a-bce5-d75f9da8983d-kube-api-access-7wlsq\") pod \"horizon-7cbbd9698f-ft2xt\" (UID: \"b4a3ee06-7c42-4c4a-bce5-d75f9da8983d\") " pod="openstack/horizon-7cbbd9698f-ft2xt" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.763467 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b469f4ba-a16a-4425-b2d6-6d95be8dcc0d-dns-swift-storage-0\") pod \"dnsmasq-dns-fcfdd6f9f-9vj6f\" (UID: \"b469f4ba-a16a-4425-b2d6-6d95be8dcc0d\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-9vj6f" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.763502 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b469f4ba-a16a-4425-b2d6-6d95be8dcc0d-ovsdbserver-sb\") pod \"dnsmasq-dns-fcfdd6f9f-9vj6f\" (UID: \"b469f4ba-a16a-4425-b2d6-6d95be8dcc0d\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-9vj6f" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.763548 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b4a3ee06-7c42-4c4a-bce5-d75f9da8983d-config-data\") pod \"horizon-7cbbd9698f-ft2xt\" (UID: \"b4a3ee06-7c42-4c4a-bce5-d75f9da8983d\") " pod="openstack/horizon-7cbbd9698f-ft2xt" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.763601 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b469f4ba-a16a-4425-b2d6-6d95be8dcc0d-config\") pod \"dnsmasq-dns-fcfdd6f9f-9vj6f\" (UID: \"b469f4ba-a16a-4425-b2d6-6d95be8dcc0d\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-9vj6f" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.763635 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/b4a3ee06-7c42-4c4a-bce5-d75f9da8983d-horizon-secret-key\") pod \"horizon-7cbbd9698f-ft2xt\" (UID: \"b4a3ee06-7c42-4c4a-bce5-d75f9da8983d\") " pod="openstack/horizon-7cbbd9698f-ft2xt" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.763672 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b469f4ba-a16a-4425-b2d6-6d95be8dcc0d-dns-svc\") pod \"dnsmasq-dns-fcfdd6f9f-9vj6f\" (UID: \"b469f4ba-a16a-4425-b2d6-6d95be8dcc0d\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-9vj6f" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.763701 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w86nj\" (UniqueName: \"kubernetes.io/projected/b469f4ba-a16a-4425-b2d6-6d95be8dcc0d-kube-api-access-w86nj\") pod \"dnsmasq-dns-fcfdd6f9f-9vj6f\" (UID: \"b469f4ba-a16a-4425-b2d6-6d95be8dcc0d\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-9vj6f" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.763765 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b4a3ee06-7c42-4c4a-bce5-d75f9da8983d-scripts\") pod \"horizon-7cbbd9698f-ft2xt\" (UID: \"b4a3ee06-7c42-4c4a-bce5-d75f9da8983d\") " pod="openstack/horizon-7cbbd9698f-ft2xt" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.763805 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b469f4ba-a16a-4425-b2d6-6d95be8dcc0d-ovsdbserver-nb\") pod \"dnsmasq-dns-fcfdd6f9f-9vj6f\" (UID: \"b469f4ba-a16a-4425-b2d6-6d95be8dcc0d\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-9vj6f" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.765450 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b4a3ee06-7c42-4c4a-bce5-d75f9da8983d-config-data\") pod \"horizon-7cbbd9698f-ft2xt\" (UID: \"b4a3ee06-7c42-4c4a-bce5-d75f9da8983d\") " pod="openstack/horizon-7cbbd9698f-ft2xt" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.766894 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b4a3ee06-7c42-4c4a-bce5-d75f9da8983d-logs\") pod \"horizon-7cbbd9698f-ft2xt\" (UID: \"b4a3ee06-7c42-4c4a-bce5-d75f9da8983d\") " pod="openstack/horizon-7cbbd9698f-ft2xt" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.767447 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b4a3ee06-7c42-4c4a-bce5-d75f9da8983d-scripts\") pod \"horizon-7cbbd9698f-ft2xt\" (UID: \"b4a3ee06-7c42-4c4a-bce5-d75f9da8983d\") " pod="openstack/horizon-7cbbd9698f-ft2xt" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.790585 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-768465cc7-f7gwz" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.788203 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-jqgbc"] Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.819397 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/b4a3ee06-7c42-4c4a-bce5-d75f9da8983d-horizon-secret-key\") pod \"horizon-7cbbd9698f-ft2xt\" (UID: \"b4a3ee06-7c42-4c4a-bce5-d75f9da8983d\") " pod="openstack/horizon-7cbbd9698f-ft2xt" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.820279 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7wlsq\" (UniqueName: \"kubernetes.io/projected/b4a3ee06-7c42-4c4a-bce5-d75f9da8983d-kube-api-access-7wlsq\") pod \"horizon-7cbbd9698f-ft2xt\" (UID: \"b4a3ee06-7c42-4c4a-bce5-d75f9da8983d\") " pod="openstack/horizon-7cbbd9698f-ft2xt" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.830054 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-jqgbc" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.834860 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-jqgbc"] Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.837880 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.838044 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-9dlgs" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.838905 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.873166 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b469f4ba-a16a-4425-b2d6-6d95be8dcc0d-config\") pod \"dnsmasq-dns-fcfdd6f9f-9vj6f\" (UID: \"b469f4ba-a16a-4425-b2d6-6d95be8dcc0d\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-9vj6f" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.873272 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b469f4ba-a16a-4425-b2d6-6d95be8dcc0d-dns-svc\") pod \"dnsmasq-dns-fcfdd6f9f-9vj6f\" (UID: \"b469f4ba-a16a-4425-b2d6-6d95be8dcc0d\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-9vj6f" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.873319 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eafaa4e1-5187-4406-b55a-9d9349e3a744-combined-ca-bundle\") pod \"placement-db-sync-jqgbc\" (UID: \"eafaa4e1-5187-4406-b55a-9d9349e3a744\") " pod="openstack/placement-db-sync-jqgbc" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.873361 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w86nj\" (UniqueName: \"kubernetes.io/projected/b469f4ba-a16a-4425-b2d6-6d95be8dcc0d-kube-api-access-w86nj\") pod \"dnsmasq-dns-fcfdd6f9f-9vj6f\" (UID: \"b469f4ba-a16a-4425-b2d6-6d95be8dcc0d\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-9vj6f" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.873478 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eafaa4e1-5187-4406-b55a-9d9349e3a744-scripts\") pod \"placement-db-sync-jqgbc\" (UID: \"eafaa4e1-5187-4406-b55a-9d9349e3a744\") " pod="openstack/placement-db-sync-jqgbc" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.873546 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7cxwk\" (UniqueName: \"kubernetes.io/projected/eafaa4e1-5187-4406-b55a-9d9349e3a744-kube-api-access-7cxwk\") pod \"placement-db-sync-jqgbc\" (UID: \"eafaa4e1-5187-4406-b55a-9d9349e3a744\") " pod="openstack/placement-db-sync-jqgbc" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.873672 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b469f4ba-a16a-4425-b2d6-6d95be8dcc0d-ovsdbserver-nb\") pod \"dnsmasq-dns-fcfdd6f9f-9vj6f\" (UID: \"b469f4ba-a16a-4425-b2d6-6d95be8dcc0d\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-9vj6f" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.873741 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b469f4ba-a16a-4425-b2d6-6d95be8dcc0d-dns-swift-storage-0\") pod \"dnsmasq-dns-fcfdd6f9f-9vj6f\" (UID: \"b469f4ba-a16a-4425-b2d6-6d95be8dcc0d\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-9vj6f" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.873787 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b469f4ba-a16a-4425-b2d6-6d95be8dcc0d-ovsdbserver-sb\") pod \"dnsmasq-dns-fcfdd6f9f-9vj6f\" (UID: \"b469f4ba-a16a-4425-b2d6-6d95be8dcc0d\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-9vj6f" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.873875 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eafaa4e1-5187-4406-b55a-9d9349e3a744-logs\") pod \"placement-db-sync-jqgbc\" (UID: \"eafaa4e1-5187-4406-b55a-9d9349e3a744\") " pod="openstack/placement-db-sync-jqgbc" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.873929 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eafaa4e1-5187-4406-b55a-9d9349e3a744-config-data\") pod \"placement-db-sync-jqgbc\" (UID: \"eafaa4e1-5187-4406-b55a-9d9349e3a744\") " pod="openstack/placement-db-sync-jqgbc" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.875866 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b469f4ba-a16a-4425-b2d6-6d95be8dcc0d-dns-svc\") pod \"dnsmasq-dns-fcfdd6f9f-9vj6f\" (UID: \"b469f4ba-a16a-4425-b2d6-6d95be8dcc0d\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-9vj6f" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.876320 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b469f4ba-a16a-4425-b2d6-6d95be8dcc0d-config\") pod \"dnsmasq-dns-fcfdd6f9f-9vj6f\" (UID: \"b469f4ba-a16a-4425-b2d6-6d95be8dcc0d\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-9vj6f" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.876539 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b469f4ba-a16a-4425-b2d6-6d95be8dcc0d-ovsdbserver-nb\") pod \"dnsmasq-dns-fcfdd6f9f-9vj6f\" (UID: \"b469f4ba-a16a-4425-b2d6-6d95be8dcc0d\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-9vj6f" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.877025 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b469f4ba-a16a-4425-b2d6-6d95be8dcc0d-ovsdbserver-sb\") pod \"dnsmasq-dns-fcfdd6f9f-9vj6f\" (UID: \"b469f4ba-a16a-4425-b2d6-6d95be8dcc0d\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-9vj6f" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.877667 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b469f4ba-a16a-4425-b2d6-6d95be8dcc0d-dns-swift-storage-0\") pod \"dnsmasq-dns-fcfdd6f9f-9vj6f\" (UID: \"b469f4ba-a16a-4425-b2d6-6d95be8dcc0d\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-9vj6f" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.908755 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w86nj\" (UniqueName: \"kubernetes.io/projected/b469f4ba-a16a-4425-b2d6-6d95be8dcc0d-kube-api-access-w86nj\") pod \"dnsmasq-dns-fcfdd6f9f-9vj6f\" (UID: \"b469f4ba-a16a-4425-b2d6-6d95be8dcc0d\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-9vj6f" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.911088 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-pfz4t" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.943415 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7cbbd9698f-ft2xt" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.977917 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eafaa4e1-5187-4406-b55a-9d9349e3a744-logs\") pod \"placement-db-sync-jqgbc\" (UID: \"eafaa4e1-5187-4406-b55a-9d9349e3a744\") " pod="openstack/placement-db-sync-jqgbc" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.978498 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eafaa4e1-5187-4406-b55a-9d9349e3a744-config-data\") pod \"placement-db-sync-jqgbc\" (UID: \"eafaa4e1-5187-4406-b55a-9d9349e3a744\") " pod="openstack/placement-db-sync-jqgbc" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.978717 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eafaa4e1-5187-4406-b55a-9d9349e3a744-logs\") pod \"placement-db-sync-jqgbc\" (UID: \"eafaa4e1-5187-4406-b55a-9d9349e3a744\") " pod="openstack/placement-db-sync-jqgbc" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.978833 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eafaa4e1-5187-4406-b55a-9d9349e3a744-combined-ca-bundle\") pod \"placement-db-sync-jqgbc\" (UID: \"eafaa4e1-5187-4406-b55a-9d9349e3a744\") " pod="openstack/placement-db-sync-jqgbc" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.978915 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eafaa4e1-5187-4406-b55a-9d9349e3a744-scripts\") pod \"placement-db-sync-jqgbc\" (UID: \"eafaa4e1-5187-4406-b55a-9d9349e3a744\") " pod="openstack/placement-db-sync-jqgbc" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.978966 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7cxwk\" (UniqueName: \"kubernetes.io/projected/eafaa4e1-5187-4406-b55a-9d9349e3a744-kube-api-access-7cxwk\") pod \"placement-db-sync-jqgbc\" (UID: \"eafaa4e1-5187-4406-b55a-9d9349e3a744\") " pod="openstack/placement-db-sync-jqgbc" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.983548 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eafaa4e1-5187-4406-b55a-9d9349e3a744-scripts\") pod \"placement-db-sync-jqgbc\" (UID: \"eafaa4e1-5187-4406-b55a-9d9349e3a744\") " pod="openstack/placement-db-sync-jqgbc" Dec 11 08:28:58 crc kubenswrapper[4860]: I1211 08:28:58.985000 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eafaa4e1-5187-4406-b55a-9d9349e3a744-combined-ca-bundle\") pod \"placement-db-sync-jqgbc\" (UID: \"eafaa4e1-5187-4406-b55a-9d9349e3a744\") " pod="openstack/placement-db-sync-jqgbc" Dec 11 08:28:59 crc kubenswrapper[4860]: I1211 08:28:59.001291 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eafaa4e1-5187-4406-b55a-9d9349e3a744-config-data\") pod \"placement-db-sync-jqgbc\" (UID: \"eafaa4e1-5187-4406-b55a-9d9349e3a744\") " pod="openstack/placement-db-sync-jqgbc" Dec 11 08:28:59 crc kubenswrapper[4860]: I1211 08:28:59.007609 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7cxwk\" (UniqueName: \"kubernetes.io/projected/eafaa4e1-5187-4406-b55a-9d9349e3a744-kube-api-access-7cxwk\") pod \"placement-db-sync-jqgbc\" (UID: \"eafaa4e1-5187-4406-b55a-9d9349e3a744\") " pod="openstack/placement-db-sync-jqgbc" Dec 11 08:28:59 crc kubenswrapper[4860]: I1211 08:28:59.055242 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-fcfdd6f9f-9vj6f" Dec 11 08:28:59 crc kubenswrapper[4860]: I1211 08:28:59.156181 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-jqgbc" Dec 11 08:28:59 crc kubenswrapper[4860]: I1211 08:28:59.263513 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-xq969"] Dec 11 08:28:59 crc kubenswrapper[4860]: I1211 08:28:59.503827 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-wgk5t"] Dec 11 08:28:59 crc kubenswrapper[4860]: I1211 08:28:59.515155 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 11 08:28:59 crc kubenswrapper[4860]: I1211 08:28:59.533967 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-hccts"] Dec 11 08:28:59 crc kubenswrapper[4860]: W1211 08:28:59.701430 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode3822ab5_d1a7_46da_9533_21b67bd240eb.slice/crio-3c8a25c0e0a663af7cce9b56a57ec9372e8c0a04ad5fc9e5c618b575b544d25a WatchSource:0}: Error finding container 3c8a25c0e0a663af7cce9b56a57ec9372e8c0a04ad5fc9e5c618b575b544d25a: Status 404 returned error can't find the container with id 3c8a25c0e0a663af7cce9b56a57ec9372e8c0a04ad5fc9e5c618b575b544d25a Dec 11 08:28:59 crc kubenswrapper[4860]: I1211 08:28:59.735792 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-xq969" event={"ID":"a7bc1b0c-9f61-4bd1-8739-6219931987fa","Type":"ContainerStarted","Data":"72d58e20b23e3fe5f3bb76949c37ffecf67efe0c996bffb7e8ef8a0db75b9fa6"} Dec 11 08:28:59 crc kubenswrapper[4860]: I1211 08:28:59.919582 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6f8c45789f-zdxj7"] Dec 11 08:28:59 crc kubenswrapper[4860]: W1211 08:28:59.928180 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod232401ac_edc1_41a7_a802_a744276304c8.slice/crio-0bd6c433bba293f3386a7367239e76234d62247a2c59a2166e79c169dca72a01 WatchSource:0}: Error finding container 0bd6c433bba293f3386a7367239e76234d62247a2c59a2166e79c169dca72a01: Status 404 returned error can't find the container with id 0bd6c433bba293f3386a7367239e76234d62247a2c59a2166e79c169dca72a01 Dec 11 08:29:00 crc kubenswrapper[4860]: I1211 08:29:00.054764 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-pfz4t"] Dec 11 08:29:00 crc kubenswrapper[4860]: I1211 08:29:00.065554 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-768465cc7-f7gwz"] Dec 11 08:29:00 crc kubenswrapper[4860]: I1211 08:29:00.077389 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-fcfdd6f9f-9vj6f"] Dec 11 08:29:00 crc kubenswrapper[4860]: I1211 08:29:00.086325 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7cbbd9698f-ft2xt"] Dec 11 08:29:00 crc kubenswrapper[4860]: W1211 08:29:00.087515 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb469f4ba_a16a_4425_b2d6_6d95be8dcc0d.slice/crio-c3e348dd7e58f1be91412928b9c09c6522ff3d83d2b6500bdd0733b2cc707285 WatchSource:0}: Error finding container c3e348dd7e58f1be91412928b9c09c6522ff3d83d2b6500bdd0733b2cc707285: Status 404 returned error can't find the container with id c3e348dd7e58f1be91412928b9c09c6522ff3d83d2b6500bdd0733b2cc707285 Dec 11 08:29:00 crc kubenswrapper[4860]: W1211 08:29:00.109352 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb4a3ee06_7c42_4c4a_bce5_d75f9da8983d.slice/crio-8b1e36928dc90198ac3111a6937367dae82532daf1167906a97b1c1979bd0a00 WatchSource:0}: Error finding container 8b1e36928dc90198ac3111a6937367dae82532daf1167906a97b1c1979bd0a00: Status 404 returned error can't find the container with id 8b1e36928dc90198ac3111a6937367dae82532daf1167906a97b1c1979bd0a00 Dec 11 08:29:00 crc kubenswrapper[4860]: I1211 08:29:00.229984 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-jqgbc"] Dec 11 08:29:00 crc kubenswrapper[4860]: I1211 08:29:00.453439 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 11 08:29:00 crc kubenswrapper[4860]: I1211 08:29:00.474459 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-768465cc7-f7gwz"] Dec 11 08:29:00 crc kubenswrapper[4860]: I1211 08:29:00.523875 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-6b45499b8f-8xhjk"] Dec 11 08:29:00 crc kubenswrapper[4860]: I1211 08:29:00.526168 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6b45499b8f-8xhjk" Dec 11 08:29:00 crc kubenswrapper[4860]: I1211 08:29:00.544781 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6b45499b8f-8xhjk"] Dec 11 08:29:00 crc kubenswrapper[4860]: I1211 08:29:00.653903 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5c9ab8b4-ed9e-46a5-b553-3d085ff20722-scripts\") pod \"horizon-6b45499b8f-8xhjk\" (UID: \"5c9ab8b4-ed9e-46a5-b553-3d085ff20722\") " pod="openstack/horizon-6b45499b8f-8xhjk" Dec 11 08:29:00 crc kubenswrapper[4860]: I1211 08:29:00.654030 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fjxs5\" (UniqueName: \"kubernetes.io/projected/5c9ab8b4-ed9e-46a5-b553-3d085ff20722-kube-api-access-fjxs5\") pod \"horizon-6b45499b8f-8xhjk\" (UID: \"5c9ab8b4-ed9e-46a5-b553-3d085ff20722\") " pod="openstack/horizon-6b45499b8f-8xhjk" Dec 11 08:29:00 crc kubenswrapper[4860]: I1211 08:29:00.654077 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/5c9ab8b4-ed9e-46a5-b553-3d085ff20722-horizon-secret-key\") pod \"horizon-6b45499b8f-8xhjk\" (UID: \"5c9ab8b4-ed9e-46a5-b553-3d085ff20722\") " pod="openstack/horizon-6b45499b8f-8xhjk" Dec 11 08:29:00 crc kubenswrapper[4860]: I1211 08:29:00.654133 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5c9ab8b4-ed9e-46a5-b553-3d085ff20722-logs\") pod \"horizon-6b45499b8f-8xhjk\" (UID: \"5c9ab8b4-ed9e-46a5-b553-3d085ff20722\") " pod="openstack/horizon-6b45499b8f-8xhjk" Dec 11 08:29:00 crc kubenswrapper[4860]: I1211 08:29:00.654152 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5c9ab8b4-ed9e-46a5-b553-3d085ff20722-config-data\") pod \"horizon-6b45499b8f-8xhjk\" (UID: \"5c9ab8b4-ed9e-46a5-b553-3d085ff20722\") " pod="openstack/horizon-6b45499b8f-8xhjk" Dec 11 08:29:00 crc kubenswrapper[4860]: I1211 08:29:00.758939 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fjxs5\" (UniqueName: \"kubernetes.io/projected/5c9ab8b4-ed9e-46a5-b553-3d085ff20722-kube-api-access-fjxs5\") pod \"horizon-6b45499b8f-8xhjk\" (UID: \"5c9ab8b4-ed9e-46a5-b553-3d085ff20722\") " pod="openstack/horizon-6b45499b8f-8xhjk" Dec 11 08:29:00 crc kubenswrapper[4860]: I1211 08:29:00.759041 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/5c9ab8b4-ed9e-46a5-b553-3d085ff20722-horizon-secret-key\") pod \"horizon-6b45499b8f-8xhjk\" (UID: \"5c9ab8b4-ed9e-46a5-b553-3d085ff20722\") " pod="openstack/horizon-6b45499b8f-8xhjk" Dec 11 08:29:00 crc kubenswrapper[4860]: I1211 08:29:00.759087 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5c9ab8b4-ed9e-46a5-b553-3d085ff20722-logs\") pod \"horizon-6b45499b8f-8xhjk\" (UID: \"5c9ab8b4-ed9e-46a5-b553-3d085ff20722\") " pod="openstack/horizon-6b45499b8f-8xhjk" Dec 11 08:29:00 crc kubenswrapper[4860]: I1211 08:29:00.759107 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5c9ab8b4-ed9e-46a5-b553-3d085ff20722-config-data\") pod \"horizon-6b45499b8f-8xhjk\" (UID: \"5c9ab8b4-ed9e-46a5-b553-3d085ff20722\") " pod="openstack/horizon-6b45499b8f-8xhjk" Dec 11 08:29:00 crc kubenswrapper[4860]: I1211 08:29:00.759166 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5c9ab8b4-ed9e-46a5-b553-3d085ff20722-scripts\") pod \"horizon-6b45499b8f-8xhjk\" (UID: \"5c9ab8b4-ed9e-46a5-b553-3d085ff20722\") " pod="openstack/horizon-6b45499b8f-8xhjk" Dec 11 08:29:00 crc kubenswrapper[4860]: I1211 08:29:00.760516 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5c9ab8b4-ed9e-46a5-b553-3d085ff20722-scripts\") pod \"horizon-6b45499b8f-8xhjk\" (UID: \"5c9ab8b4-ed9e-46a5-b553-3d085ff20722\") " pod="openstack/horizon-6b45499b8f-8xhjk" Dec 11 08:29:00 crc kubenswrapper[4860]: I1211 08:29:00.761329 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5c9ab8b4-ed9e-46a5-b553-3d085ff20722-logs\") pod \"horizon-6b45499b8f-8xhjk\" (UID: \"5c9ab8b4-ed9e-46a5-b553-3d085ff20722\") " pod="openstack/horizon-6b45499b8f-8xhjk" Dec 11 08:29:00 crc kubenswrapper[4860]: I1211 08:29:00.762165 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5c9ab8b4-ed9e-46a5-b553-3d085ff20722-config-data\") pod \"horizon-6b45499b8f-8xhjk\" (UID: \"5c9ab8b4-ed9e-46a5-b553-3d085ff20722\") " pod="openstack/horizon-6b45499b8f-8xhjk" Dec 11 08:29:00 crc kubenswrapper[4860]: I1211 08:29:00.768128 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/5c9ab8b4-ed9e-46a5-b553-3d085ff20722-horizon-secret-key\") pod \"horizon-6b45499b8f-8xhjk\" (UID: \"5c9ab8b4-ed9e-46a5-b553-3d085ff20722\") " pod="openstack/horizon-6b45499b8f-8xhjk" Dec 11 08:29:00 crc kubenswrapper[4860]: I1211 08:29:00.768875 4860 generic.go:334] "Generic (PLEG): container finished" podID="b469f4ba-a16a-4425-b2d6-6d95be8dcc0d" containerID="2812c3cdf32d7175891d726a1e9edd0bf65a84cf2104bcd2d114a5b27c8145ed" exitCode=0 Dec 11 08:29:00 crc kubenswrapper[4860]: I1211 08:29:00.768951 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-fcfdd6f9f-9vj6f" event={"ID":"b469f4ba-a16a-4425-b2d6-6d95be8dcc0d","Type":"ContainerDied","Data":"2812c3cdf32d7175891d726a1e9edd0bf65a84cf2104bcd2d114a5b27c8145ed"} Dec 11 08:29:00 crc kubenswrapper[4860]: I1211 08:29:00.768984 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-fcfdd6f9f-9vj6f" event={"ID":"b469f4ba-a16a-4425-b2d6-6d95be8dcc0d","Type":"ContainerStarted","Data":"c3e348dd7e58f1be91412928b9c09c6522ff3d83d2b6500bdd0733b2cc707285"} Dec 11 08:29:00 crc kubenswrapper[4860]: I1211 08:29:00.775450 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7cbbd9698f-ft2xt" event={"ID":"b4a3ee06-7c42-4c4a-bce5-d75f9da8983d","Type":"ContainerStarted","Data":"8b1e36928dc90198ac3111a6937367dae82532daf1167906a97b1c1979bd0a00"} Dec 11 08:29:00 crc kubenswrapper[4860]: I1211 08:29:00.782004 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fjxs5\" (UniqueName: \"kubernetes.io/projected/5c9ab8b4-ed9e-46a5-b553-3d085ff20722-kube-api-access-fjxs5\") pod \"horizon-6b45499b8f-8xhjk\" (UID: \"5c9ab8b4-ed9e-46a5-b553-3d085ff20722\") " pod="openstack/horizon-6b45499b8f-8xhjk" Dec 11 08:29:00 crc kubenswrapper[4860]: I1211 08:29:00.784873 4860 generic.go:334] "Generic (PLEG): container finished" podID="232401ac-edc1-41a7-a802-a744276304c8" containerID="d8c429932a455336233785313d79e64462655a9f936c09e6491ab90133b53c46" exitCode=0 Dec 11 08:29:00 crc kubenswrapper[4860]: I1211 08:29:00.785147 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f8c45789f-zdxj7" event={"ID":"232401ac-edc1-41a7-a802-a744276304c8","Type":"ContainerDied","Data":"d8c429932a455336233785313d79e64462655a9f936c09e6491ab90133b53c46"} Dec 11 08:29:00 crc kubenswrapper[4860]: I1211 08:29:00.785182 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f8c45789f-zdxj7" event={"ID":"232401ac-edc1-41a7-a802-a744276304c8","Type":"ContainerStarted","Data":"0bd6c433bba293f3386a7367239e76234d62247a2c59a2166e79c169dca72a01"} Dec 11 08:29:00 crc kubenswrapper[4860]: I1211 08:29:00.800706 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"040a0e8c-efd3-4e3b-be51-59e311ae4406","Type":"ContainerStarted","Data":"445eca66a37a53e77187da28ab3dd3c55b55729539d52878dfbe12fabc64bdd3"} Dec 11 08:29:00 crc kubenswrapper[4860]: I1211 08:29:00.802690 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-hccts" event={"ID":"d472d27e-9bc9-4d99-8fd9-697f7f44f57d","Type":"ContainerStarted","Data":"b3e71423675304467fc90781a8bf9efd1d88d3d2db17dc44e313cfabdf925503"} Dec 11 08:29:00 crc kubenswrapper[4860]: I1211 08:29:00.808180 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-xq969" event={"ID":"a7bc1b0c-9f61-4bd1-8739-6219931987fa","Type":"ContainerStarted","Data":"3038d16b28f6b130939667f3daca0bc372de9ac7d60de3e167f8e56da8339937"} Dec 11 08:29:00 crc kubenswrapper[4860]: I1211 08:29:00.812372 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-pfz4t" event={"ID":"4ff61ef1-6a67-4a8e-80e2-11f0154e54c6","Type":"ContainerStarted","Data":"6ada87093bbc32624bf5ed6b56c88f8ae8afb66ca49ea4598f7c4115649d070a"} Dec 11 08:29:00 crc kubenswrapper[4860]: I1211 08:29:00.815798 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-wgk5t" event={"ID":"e3822ab5-d1a7-46da-9533-21b67bd240eb","Type":"ContainerStarted","Data":"650883b438680b9c1ab516bfe7c40142cef24b094b49f32d2439af5c7e7508d6"} Dec 11 08:29:00 crc kubenswrapper[4860]: I1211 08:29:00.815846 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-wgk5t" event={"ID":"e3822ab5-d1a7-46da-9533-21b67bd240eb","Type":"ContainerStarted","Data":"3c8a25c0e0a663af7cce9b56a57ec9372e8c0a04ad5fc9e5c618b575b544d25a"} Dec 11 08:29:00 crc kubenswrapper[4860]: I1211 08:29:00.835675 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-768465cc7-f7gwz" event={"ID":"1ee42a0c-5029-44cf-ba8d-720165ac6138","Type":"ContainerStarted","Data":"dabe4e3751a6cae5d7034bf68fdb500a6600bbf56561b35df1730fca3dbb7ba6"} Dec 11 08:29:00 crc kubenswrapper[4860]: I1211 08:29:00.837312 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-jqgbc" event={"ID":"eafaa4e1-5187-4406-b55a-9d9349e3a744","Type":"ContainerStarted","Data":"180aa31559929b074de4c1faefdc518264286cf7496070de82b469144a62a978"} Dec 11 08:29:00 crc kubenswrapper[4860]: I1211 08:29:00.853968 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6b45499b8f-8xhjk" Dec 11 08:29:00 crc kubenswrapper[4860]: I1211 08:29:00.865448 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-wgk5t" podStartSLOduration=2.86541612 podStartE2EDuration="2.86541612s" podCreationTimestamp="2025-12-11 08:28:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:29:00.854213308 +0000 UTC m=+1073.582732383" watchObservedRunningTime="2025-12-11 08:29:00.86541612 +0000 UTC m=+1073.593935185" Dec 11 08:29:00 crc kubenswrapper[4860]: I1211 08:29:00.883450 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-xq969" podStartSLOduration=3.883424566 podStartE2EDuration="3.883424566s" podCreationTimestamp="2025-12-11 08:28:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:29:00.881398042 +0000 UTC m=+1073.609917097" watchObservedRunningTime="2025-12-11 08:29:00.883424566 +0000 UTC m=+1073.611943621" Dec 11 08:29:01 crc kubenswrapper[4860]: I1211 08:29:01.283511 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f8c45789f-zdxj7" Dec 11 08:29:01 crc kubenswrapper[4860]: I1211 08:29:01.398662 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/232401ac-edc1-41a7-a802-a744276304c8-ovsdbserver-nb\") pod \"232401ac-edc1-41a7-a802-a744276304c8\" (UID: \"232401ac-edc1-41a7-a802-a744276304c8\") " Dec 11 08:29:01 crc kubenswrapper[4860]: I1211 08:29:01.399949 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6lxlg\" (UniqueName: \"kubernetes.io/projected/232401ac-edc1-41a7-a802-a744276304c8-kube-api-access-6lxlg\") pod \"232401ac-edc1-41a7-a802-a744276304c8\" (UID: \"232401ac-edc1-41a7-a802-a744276304c8\") " Dec 11 08:29:01 crc kubenswrapper[4860]: I1211 08:29:01.400062 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/232401ac-edc1-41a7-a802-a744276304c8-ovsdbserver-sb\") pod \"232401ac-edc1-41a7-a802-a744276304c8\" (UID: \"232401ac-edc1-41a7-a802-a744276304c8\") " Dec 11 08:29:01 crc kubenswrapper[4860]: I1211 08:29:01.400151 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/232401ac-edc1-41a7-a802-a744276304c8-dns-svc\") pod \"232401ac-edc1-41a7-a802-a744276304c8\" (UID: \"232401ac-edc1-41a7-a802-a744276304c8\") " Dec 11 08:29:01 crc kubenswrapper[4860]: I1211 08:29:01.400193 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/232401ac-edc1-41a7-a802-a744276304c8-dns-swift-storage-0\") pod \"232401ac-edc1-41a7-a802-a744276304c8\" (UID: \"232401ac-edc1-41a7-a802-a744276304c8\") " Dec 11 08:29:01 crc kubenswrapper[4860]: I1211 08:29:01.400222 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/232401ac-edc1-41a7-a802-a744276304c8-config\") pod \"232401ac-edc1-41a7-a802-a744276304c8\" (UID: \"232401ac-edc1-41a7-a802-a744276304c8\") " Dec 11 08:29:01 crc kubenswrapper[4860]: I1211 08:29:01.438191 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/232401ac-edc1-41a7-a802-a744276304c8-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "232401ac-edc1-41a7-a802-a744276304c8" (UID: "232401ac-edc1-41a7-a802-a744276304c8"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:29:01 crc kubenswrapper[4860]: I1211 08:29:01.439309 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/232401ac-edc1-41a7-a802-a744276304c8-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "232401ac-edc1-41a7-a802-a744276304c8" (UID: "232401ac-edc1-41a7-a802-a744276304c8"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:29:01 crc kubenswrapper[4860]: I1211 08:29:01.446088 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/232401ac-edc1-41a7-a802-a744276304c8-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "232401ac-edc1-41a7-a802-a744276304c8" (UID: "232401ac-edc1-41a7-a802-a744276304c8"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:29:01 crc kubenswrapper[4860]: I1211 08:29:01.463047 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/232401ac-edc1-41a7-a802-a744276304c8-kube-api-access-6lxlg" (OuterVolumeSpecName: "kube-api-access-6lxlg") pod "232401ac-edc1-41a7-a802-a744276304c8" (UID: "232401ac-edc1-41a7-a802-a744276304c8"). InnerVolumeSpecName "kube-api-access-6lxlg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:29:01 crc kubenswrapper[4860]: I1211 08:29:01.466137 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/232401ac-edc1-41a7-a802-a744276304c8-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "232401ac-edc1-41a7-a802-a744276304c8" (UID: "232401ac-edc1-41a7-a802-a744276304c8"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:29:01 crc kubenswrapper[4860]: I1211 08:29:01.492594 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/232401ac-edc1-41a7-a802-a744276304c8-config" (OuterVolumeSpecName: "config") pod "232401ac-edc1-41a7-a802-a744276304c8" (UID: "232401ac-edc1-41a7-a802-a744276304c8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:29:01 crc kubenswrapper[4860]: I1211 08:29:01.504197 4860 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/232401ac-edc1-41a7-a802-a744276304c8-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 11 08:29:01 crc kubenswrapper[4860]: I1211 08:29:01.504249 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6lxlg\" (UniqueName: \"kubernetes.io/projected/232401ac-edc1-41a7-a802-a744276304c8-kube-api-access-6lxlg\") on node \"crc\" DevicePath \"\"" Dec 11 08:29:01 crc kubenswrapper[4860]: I1211 08:29:01.504268 4860 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/232401ac-edc1-41a7-a802-a744276304c8-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 11 08:29:01 crc kubenswrapper[4860]: I1211 08:29:01.504283 4860 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/232401ac-edc1-41a7-a802-a744276304c8-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 11 08:29:01 crc kubenswrapper[4860]: I1211 08:29:01.504294 4860 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/232401ac-edc1-41a7-a802-a744276304c8-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 11 08:29:01 crc kubenswrapper[4860]: I1211 08:29:01.504305 4860 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/232401ac-edc1-41a7-a802-a744276304c8-config\") on node \"crc\" DevicePath \"\"" Dec 11 08:29:01 crc kubenswrapper[4860]: I1211 08:29:01.529828 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6b45499b8f-8xhjk"] Dec 11 08:29:01 crc kubenswrapper[4860]: I1211 08:29:01.860656 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f8c45789f-zdxj7" event={"ID":"232401ac-edc1-41a7-a802-a744276304c8","Type":"ContainerDied","Data":"0bd6c433bba293f3386a7367239e76234d62247a2c59a2166e79c169dca72a01"} Dec 11 08:29:01 crc kubenswrapper[4860]: I1211 08:29:01.860729 4860 scope.go:117] "RemoveContainer" containerID="d8c429932a455336233785313d79e64462655a9f936c09e6491ab90133b53c46" Dec 11 08:29:01 crc kubenswrapper[4860]: I1211 08:29:01.860892 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f8c45789f-zdxj7" Dec 11 08:29:01 crc kubenswrapper[4860]: I1211 08:29:01.864105 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6b45499b8f-8xhjk" event={"ID":"5c9ab8b4-ed9e-46a5-b553-3d085ff20722","Type":"ContainerStarted","Data":"6dff7cdae86985e8baa6bad4ec0cd2b4da501d0c3494a18976771c194d5c5217"} Dec 11 08:29:01 crc kubenswrapper[4860]: I1211 08:29:01.872225 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-fcfdd6f9f-9vj6f" event={"ID":"b469f4ba-a16a-4425-b2d6-6d95be8dcc0d","Type":"ContainerStarted","Data":"8e4b1bc15563056317722c20ddb943a307f432b3ab351a7aa0200777cd3ea1c5"} Dec 11 08:29:01 crc kubenswrapper[4860]: I1211 08:29:01.872271 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-fcfdd6f9f-9vj6f" Dec 11 08:29:01 crc kubenswrapper[4860]: I1211 08:29:01.953759 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6f8c45789f-zdxj7"] Dec 11 08:29:01 crc kubenswrapper[4860]: I1211 08:29:01.979899 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6f8c45789f-zdxj7"] Dec 11 08:29:01 crc kubenswrapper[4860]: I1211 08:29:01.983919 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-fcfdd6f9f-9vj6f" podStartSLOduration=3.983887514 podStartE2EDuration="3.983887514s" podCreationTimestamp="2025-12-11 08:28:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:29:01.956337419 +0000 UTC m=+1074.684856474" watchObservedRunningTime="2025-12-11 08:29:01.983887514 +0000 UTC m=+1074.712406569" Dec 11 08:29:03 crc kubenswrapper[4860]: I1211 08:29:03.595511 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="232401ac-edc1-41a7-a802-a744276304c8" path="/var/lib/kubelet/pods/232401ac-edc1-41a7-a802-a744276304c8/volumes" Dec 11 08:29:04 crc kubenswrapper[4860]: I1211 08:29:04.919953 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-4fsqr" event={"ID":"cf433d5d-8c95-4d72-864c-3b74f315dd2b","Type":"ContainerStarted","Data":"d737396b95b4dbb30ba8c85e823e3e330e74286479aa940ead93372b72ec06f3"} Dec 11 08:29:04 crc kubenswrapper[4860]: I1211 08:29:04.948592 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-4fsqr" podStartSLOduration=2.326239496 podStartE2EDuration="34.948568237s" podCreationTimestamp="2025-12-11 08:28:30 +0000 UTC" firstStartedPulling="2025-12-11 08:28:30.978515586 +0000 UTC m=+1043.707034661" lastFinishedPulling="2025-12-11 08:29:03.600844347 +0000 UTC m=+1076.329363402" observedRunningTime="2025-12-11 08:29:04.943547699 +0000 UTC m=+1077.672066764" watchObservedRunningTime="2025-12-11 08:29:04.948568237 +0000 UTC m=+1077.677087292" Dec 11 08:29:06 crc kubenswrapper[4860]: I1211 08:29:06.306024 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7cbbd9698f-ft2xt"] Dec 11 08:29:06 crc kubenswrapper[4860]: I1211 08:29:06.357798 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-776755b658-9fwr9"] Dec 11 08:29:06 crc kubenswrapper[4860]: E1211 08:29:06.358338 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="232401ac-edc1-41a7-a802-a744276304c8" containerName="init" Dec 11 08:29:06 crc kubenswrapper[4860]: I1211 08:29:06.358357 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="232401ac-edc1-41a7-a802-a744276304c8" containerName="init" Dec 11 08:29:06 crc kubenswrapper[4860]: I1211 08:29:06.358553 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="232401ac-edc1-41a7-a802-a744276304c8" containerName="init" Dec 11 08:29:06 crc kubenswrapper[4860]: I1211 08:29:06.359555 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-776755b658-9fwr9" Dec 11 08:29:06 crc kubenswrapper[4860]: I1211 08:29:06.365317 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Dec 11 08:29:06 crc kubenswrapper[4860]: I1211 08:29:06.369503 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-776755b658-9fwr9"] Dec 11 08:29:06 crc kubenswrapper[4860]: I1211 08:29:06.449982 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nx4wx\" (UniqueName: \"kubernetes.io/projected/8f38e033-5243-4e65-a7e0-29f1282b063f-kube-api-access-nx4wx\") pod \"horizon-776755b658-9fwr9\" (UID: \"8f38e033-5243-4e65-a7e0-29f1282b063f\") " pod="openstack/horizon-776755b658-9fwr9" Dec 11 08:29:06 crc kubenswrapper[4860]: I1211 08:29:06.450053 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8f38e033-5243-4e65-a7e0-29f1282b063f-logs\") pod \"horizon-776755b658-9fwr9\" (UID: \"8f38e033-5243-4e65-a7e0-29f1282b063f\") " pod="openstack/horizon-776755b658-9fwr9" Dec 11 08:29:06 crc kubenswrapper[4860]: I1211 08:29:06.450081 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/8f38e033-5243-4e65-a7e0-29f1282b063f-horizon-tls-certs\") pod \"horizon-776755b658-9fwr9\" (UID: \"8f38e033-5243-4e65-a7e0-29f1282b063f\") " pod="openstack/horizon-776755b658-9fwr9" Dec 11 08:29:06 crc kubenswrapper[4860]: I1211 08:29:06.450187 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f38e033-5243-4e65-a7e0-29f1282b063f-combined-ca-bundle\") pod \"horizon-776755b658-9fwr9\" (UID: \"8f38e033-5243-4e65-a7e0-29f1282b063f\") " pod="openstack/horizon-776755b658-9fwr9" Dec 11 08:29:06 crc kubenswrapper[4860]: I1211 08:29:06.450253 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8f38e033-5243-4e65-a7e0-29f1282b063f-scripts\") pod \"horizon-776755b658-9fwr9\" (UID: \"8f38e033-5243-4e65-a7e0-29f1282b063f\") " pod="openstack/horizon-776755b658-9fwr9" Dec 11 08:29:06 crc kubenswrapper[4860]: I1211 08:29:06.450295 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8f38e033-5243-4e65-a7e0-29f1282b063f-config-data\") pod \"horizon-776755b658-9fwr9\" (UID: \"8f38e033-5243-4e65-a7e0-29f1282b063f\") " pod="openstack/horizon-776755b658-9fwr9" Dec 11 08:29:06 crc kubenswrapper[4860]: I1211 08:29:06.450325 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/8f38e033-5243-4e65-a7e0-29f1282b063f-horizon-secret-key\") pod \"horizon-776755b658-9fwr9\" (UID: \"8f38e033-5243-4e65-a7e0-29f1282b063f\") " pod="openstack/horizon-776755b658-9fwr9" Dec 11 08:29:06 crc kubenswrapper[4860]: I1211 08:29:06.465710 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-6b45499b8f-8xhjk"] Dec 11 08:29:06 crc kubenswrapper[4860]: I1211 08:29:06.492476 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-6b4697fc68-vnd26"] Dec 11 08:29:06 crc kubenswrapper[4860]: I1211 08:29:06.494042 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6b4697fc68-vnd26" Dec 11 08:29:06 crc kubenswrapper[4860]: I1211 08:29:06.508946 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6b4697fc68-vnd26"] Dec 11 08:29:06 crc kubenswrapper[4860]: I1211 08:29:06.552386 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8f38e033-5243-4e65-a7e0-29f1282b063f-logs\") pod \"horizon-776755b658-9fwr9\" (UID: \"8f38e033-5243-4e65-a7e0-29f1282b063f\") " pod="openstack/horizon-776755b658-9fwr9" Dec 11 08:29:06 crc kubenswrapper[4860]: I1211 08:29:06.552452 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/8f38e033-5243-4e65-a7e0-29f1282b063f-horizon-tls-certs\") pod \"horizon-776755b658-9fwr9\" (UID: \"8f38e033-5243-4e65-a7e0-29f1282b063f\") " pod="openstack/horizon-776755b658-9fwr9" Dec 11 08:29:06 crc kubenswrapper[4860]: I1211 08:29:06.552517 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f38e033-5243-4e65-a7e0-29f1282b063f-combined-ca-bundle\") pod \"horizon-776755b658-9fwr9\" (UID: \"8f38e033-5243-4e65-a7e0-29f1282b063f\") " pod="openstack/horizon-776755b658-9fwr9" Dec 11 08:29:06 crc kubenswrapper[4860]: I1211 08:29:06.552607 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8f38e033-5243-4e65-a7e0-29f1282b063f-scripts\") pod \"horizon-776755b658-9fwr9\" (UID: \"8f38e033-5243-4e65-a7e0-29f1282b063f\") " pod="openstack/horizon-776755b658-9fwr9" Dec 11 08:29:06 crc kubenswrapper[4860]: I1211 08:29:06.552683 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8f38e033-5243-4e65-a7e0-29f1282b063f-config-data\") pod \"horizon-776755b658-9fwr9\" (UID: \"8f38e033-5243-4e65-a7e0-29f1282b063f\") " pod="openstack/horizon-776755b658-9fwr9" Dec 11 08:29:06 crc kubenswrapper[4860]: I1211 08:29:06.552729 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/8f38e033-5243-4e65-a7e0-29f1282b063f-horizon-secret-key\") pod \"horizon-776755b658-9fwr9\" (UID: \"8f38e033-5243-4e65-a7e0-29f1282b063f\") " pod="openstack/horizon-776755b658-9fwr9" Dec 11 08:29:06 crc kubenswrapper[4860]: I1211 08:29:06.552790 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nx4wx\" (UniqueName: \"kubernetes.io/projected/8f38e033-5243-4e65-a7e0-29f1282b063f-kube-api-access-nx4wx\") pod \"horizon-776755b658-9fwr9\" (UID: \"8f38e033-5243-4e65-a7e0-29f1282b063f\") " pod="openstack/horizon-776755b658-9fwr9" Dec 11 08:29:06 crc kubenswrapper[4860]: I1211 08:29:06.553097 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8f38e033-5243-4e65-a7e0-29f1282b063f-logs\") pod \"horizon-776755b658-9fwr9\" (UID: \"8f38e033-5243-4e65-a7e0-29f1282b063f\") " pod="openstack/horizon-776755b658-9fwr9" Dec 11 08:29:06 crc kubenswrapper[4860]: I1211 08:29:06.557413 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8f38e033-5243-4e65-a7e0-29f1282b063f-scripts\") pod \"horizon-776755b658-9fwr9\" (UID: \"8f38e033-5243-4e65-a7e0-29f1282b063f\") " pod="openstack/horizon-776755b658-9fwr9" Dec 11 08:29:06 crc kubenswrapper[4860]: I1211 08:29:06.561107 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/8f38e033-5243-4e65-a7e0-29f1282b063f-horizon-secret-key\") pod \"horizon-776755b658-9fwr9\" (UID: \"8f38e033-5243-4e65-a7e0-29f1282b063f\") " pod="openstack/horizon-776755b658-9fwr9" Dec 11 08:29:06 crc kubenswrapper[4860]: I1211 08:29:06.561871 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f38e033-5243-4e65-a7e0-29f1282b063f-combined-ca-bundle\") pod \"horizon-776755b658-9fwr9\" (UID: \"8f38e033-5243-4e65-a7e0-29f1282b063f\") " pod="openstack/horizon-776755b658-9fwr9" Dec 11 08:29:06 crc kubenswrapper[4860]: I1211 08:29:06.562133 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/8f38e033-5243-4e65-a7e0-29f1282b063f-horizon-tls-certs\") pod \"horizon-776755b658-9fwr9\" (UID: \"8f38e033-5243-4e65-a7e0-29f1282b063f\") " pod="openstack/horizon-776755b658-9fwr9" Dec 11 08:29:06 crc kubenswrapper[4860]: I1211 08:29:06.562290 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8f38e033-5243-4e65-a7e0-29f1282b063f-config-data\") pod \"horizon-776755b658-9fwr9\" (UID: \"8f38e033-5243-4e65-a7e0-29f1282b063f\") " pod="openstack/horizon-776755b658-9fwr9" Dec 11 08:29:06 crc kubenswrapper[4860]: I1211 08:29:06.572234 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nx4wx\" (UniqueName: \"kubernetes.io/projected/8f38e033-5243-4e65-a7e0-29f1282b063f-kube-api-access-nx4wx\") pod \"horizon-776755b658-9fwr9\" (UID: \"8f38e033-5243-4e65-a7e0-29f1282b063f\") " pod="openstack/horizon-776755b658-9fwr9" Dec 11 08:29:06 crc kubenswrapper[4860]: I1211 08:29:06.654558 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4pzkr\" (UniqueName: \"kubernetes.io/projected/78b0f75e-d447-4541-be11-9afe4549d2e1-kube-api-access-4pzkr\") pod \"horizon-6b4697fc68-vnd26\" (UID: \"78b0f75e-d447-4541-be11-9afe4549d2e1\") " pod="openstack/horizon-6b4697fc68-vnd26" Dec 11 08:29:06 crc kubenswrapper[4860]: I1211 08:29:06.654668 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/78b0f75e-d447-4541-be11-9afe4549d2e1-horizon-tls-certs\") pod \"horizon-6b4697fc68-vnd26\" (UID: \"78b0f75e-d447-4541-be11-9afe4549d2e1\") " pod="openstack/horizon-6b4697fc68-vnd26" Dec 11 08:29:06 crc kubenswrapper[4860]: I1211 08:29:06.654704 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/78b0f75e-d447-4541-be11-9afe4549d2e1-horizon-secret-key\") pod \"horizon-6b4697fc68-vnd26\" (UID: \"78b0f75e-d447-4541-be11-9afe4549d2e1\") " pod="openstack/horizon-6b4697fc68-vnd26" Dec 11 08:29:06 crc kubenswrapper[4860]: I1211 08:29:06.654733 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/78b0f75e-d447-4541-be11-9afe4549d2e1-scripts\") pod \"horizon-6b4697fc68-vnd26\" (UID: \"78b0f75e-d447-4541-be11-9afe4549d2e1\") " pod="openstack/horizon-6b4697fc68-vnd26" Dec 11 08:29:06 crc kubenswrapper[4860]: I1211 08:29:06.654757 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/78b0f75e-d447-4541-be11-9afe4549d2e1-logs\") pod \"horizon-6b4697fc68-vnd26\" (UID: \"78b0f75e-d447-4541-be11-9afe4549d2e1\") " pod="openstack/horizon-6b4697fc68-vnd26" Dec 11 08:29:06 crc kubenswrapper[4860]: I1211 08:29:06.654848 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78b0f75e-d447-4541-be11-9afe4549d2e1-combined-ca-bundle\") pod \"horizon-6b4697fc68-vnd26\" (UID: \"78b0f75e-d447-4541-be11-9afe4549d2e1\") " pod="openstack/horizon-6b4697fc68-vnd26" Dec 11 08:29:06 crc kubenswrapper[4860]: I1211 08:29:06.654897 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/78b0f75e-d447-4541-be11-9afe4549d2e1-config-data\") pod \"horizon-6b4697fc68-vnd26\" (UID: \"78b0f75e-d447-4541-be11-9afe4549d2e1\") " pod="openstack/horizon-6b4697fc68-vnd26" Dec 11 08:29:06 crc kubenswrapper[4860]: I1211 08:29:06.695089 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-776755b658-9fwr9" Dec 11 08:29:06 crc kubenswrapper[4860]: I1211 08:29:06.756453 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/78b0f75e-d447-4541-be11-9afe4549d2e1-horizon-secret-key\") pod \"horizon-6b4697fc68-vnd26\" (UID: \"78b0f75e-d447-4541-be11-9afe4549d2e1\") " pod="openstack/horizon-6b4697fc68-vnd26" Dec 11 08:29:06 crc kubenswrapper[4860]: I1211 08:29:06.756542 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/78b0f75e-d447-4541-be11-9afe4549d2e1-scripts\") pod \"horizon-6b4697fc68-vnd26\" (UID: \"78b0f75e-d447-4541-be11-9afe4549d2e1\") " pod="openstack/horizon-6b4697fc68-vnd26" Dec 11 08:29:06 crc kubenswrapper[4860]: I1211 08:29:06.756564 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/78b0f75e-d447-4541-be11-9afe4549d2e1-logs\") pod \"horizon-6b4697fc68-vnd26\" (UID: \"78b0f75e-d447-4541-be11-9afe4549d2e1\") " pod="openstack/horizon-6b4697fc68-vnd26" Dec 11 08:29:06 crc kubenswrapper[4860]: I1211 08:29:06.756623 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78b0f75e-d447-4541-be11-9afe4549d2e1-combined-ca-bundle\") pod \"horizon-6b4697fc68-vnd26\" (UID: \"78b0f75e-d447-4541-be11-9afe4549d2e1\") " pod="openstack/horizon-6b4697fc68-vnd26" Dec 11 08:29:06 crc kubenswrapper[4860]: I1211 08:29:06.756834 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/78b0f75e-d447-4541-be11-9afe4549d2e1-config-data\") pod \"horizon-6b4697fc68-vnd26\" (UID: \"78b0f75e-d447-4541-be11-9afe4549d2e1\") " pod="openstack/horizon-6b4697fc68-vnd26" Dec 11 08:29:06 crc kubenswrapper[4860]: I1211 08:29:06.756951 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4pzkr\" (UniqueName: \"kubernetes.io/projected/78b0f75e-d447-4541-be11-9afe4549d2e1-kube-api-access-4pzkr\") pod \"horizon-6b4697fc68-vnd26\" (UID: \"78b0f75e-d447-4541-be11-9afe4549d2e1\") " pod="openstack/horizon-6b4697fc68-vnd26" Dec 11 08:29:06 crc kubenswrapper[4860]: I1211 08:29:06.757541 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/78b0f75e-d447-4541-be11-9afe4549d2e1-logs\") pod \"horizon-6b4697fc68-vnd26\" (UID: \"78b0f75e-d447-4541-be11-9afe4549d2e1\") " pod="openstack/horizon-6b4697fc68-vnd26" Dec 11 08:29:06 crc kubenswrapper[4860]: I1211 08:29:06.758203 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/78b0f75e-d447-4541-be11-9afe4549d2e1-scripts\") pod \"horizon-6b4697fc68-vnd26\" (UID: \"78b0f75e-d447-4541-be11-9afe4549d2e1\") " pod="openstack/horizon-6b4697fc68-vnd26" Dec 11 08:29:06 crc kubenswrapper[4860]: I1211 08:29:06.758481 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/78b0f75e-d447-4541-be11-9afe4549d2e1-horizon-tls-certs\") pod \"horizon-6b4697fc68-vnd26\" (UID: \"78b0f75e-d447-4541-be11-9afe4549d2e1\") " pod="openstack/horizon-6b4697fc68-vnd26" Dec 11 08:29:06 crc kubenswrapper[4860]: I1211 08:29:06.759384 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/78b0f75e-d447-4541-be11-9afe4549d2e1-config-data\") pod \"horizon-6b4697fc68-vnd26\" (UID: \"78b0f75e-d447-4541-be11-9afe4549d2e1\") " pod="openstack/horizon-6b4697fc68-vnd26" Dec 11 08:29:06 crc kubenswrapper[4860]: I1211 08:29:06.762928 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/78b0f75e-d447-4541-be11-9afe4549d2e1-horizon-tls-certs\") pod \"horizon-6b4697fc68-vnd26\" (UID: \"78b0f75e-d447-4541-be11-9afe4549d2e1\") " pod="openstack/horizon-6b4697fc68-vnd26" Dec 11 08:29:06 crc kubenswrapper[4860]: I1211 08:29:06.774494 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78b0f75e-d447-4541-be11-9afe4549d2e1-combined-ca-bundle\") pod \"horizon-6b4697fc68-vnd26\" (UID: \"78b0f75e-d447-4541-be11-9afe4549d2e1\") " pod="openstack/horizon-6b4697fc68-vnd26" Dec 11 08:29:06 crc kubenswrapper[4860]: I1211 08:29:06.774549 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/78b0f75e-d447-4541-be11-9afe4549d2e1-horizon-secret-key\") pod \"horizon-6b4697fc68-vnd26\" (UID: \"78b0f75e-d447-4541-be11-9afe4549d2e1\") " pod="openstack/horizon-6b4697fc68-vnd26" Dec 11 08:29:06 crc kubenswrapper[4860]: I1211 08:29:06.780441 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4pzkr\" (UniqueName: \"kubernetes.io/projected/78b0f75e-d447-4541-be11-9afe4549d2e1-kube-api-access-4pzkr\") pod \"horizon-6b4697fc68-vnd26\" (UID: \"78b0f75e-d447-4541-be11-9afe4549d2e1\") " pod="openstack/horizon-6b4697fc68-vnd26" Dec 11 08:29:06 crc kubenswrapper[4860]: I1211 08:29:06.833247 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6b4697fc68-vnd26" Dec 11 08:29:06 crc kubenswrapper[4860]: I1211 08:29:06.948897 4860 generic.go:334] "Generic (PLEG): container finished" podID="a7bc1b0c-9f61-4bd1-8739-6219931987fa" containerID="3038d16b28f6b130939667f3daca0bc372de9ac7d60de3e167f8e56da8339937" exitCode=0 Dec 11 08:29:06 crc kubenswrapper[4860]: I1211 08:29:06.948969 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-xq969" event={"ID":"a7bc1b0c-9f61-4bd1-8739-6219931987fa","Type":"ContainerDied","Data":"3038d16b28f6b130939667f3daca0bc372de9ac7d60de3e167f8e56da8339937"} Dec 11 08:29:09 crc kubenswrapper[4860]: I1211 08:29:09.056892 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-fcfdd6f9f-9vj6f" Dec 11 08:29:09 crc kubenswrapper[4860]: I1211 08:29:09.163119 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d5b6d6b67-pxxrf"] Dec 11 08:29:09 crc kubenswrapper[4860]: I1211 08:29:09.163603 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6d5b6d6b67-pxxrf" podUID="d59cd3a6-bd2d-4c74-97d0-265b3fc56dfd" containerName="dnsmasq-dns" containerID="cri-o://8458e72beef31344067e7052e8cb6285a95889bd51a203b4e3d4a97d4c4e0d2c" gracePeriod=10 Dec 11 08:29:09 crc kubenswrapper[4860]: I1211 08:29:09.985907 4860 generic.go:334] "Generic (PLEG): container finished" podID="d59cd3a6-bd2d-4c74-97d0-265b3fc56dfd" containerID="8458e72beef31344067e7052e8cb6285a95889bd51a203b4e3d4a97d4c4e0d2c" exitCode=0 Dec 11 08:29:09 crc kubenswrapper[4860]: I1211 08:29:09.985978 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d5b6d6b67-pxxrf" event={"ID":"d59cd3a6-bd2d-4c74-97d0-265b3fc56dfd","Type":"ContainerDied","Data":"8458e72beef31344067e7052e8cb6285a95889bd51a203b4e3d4a97d4c4e0d2c"} Dec 11 08:29:10 crc kubenswrapper[4860]: I1211 08:29:10.847782 4860 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6d5b6d6b67-pxxrf" podUID="d59cd3a6-bd2d-4c74-97d0-265b3fc56dfd" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.123:5353: connect: connection refused" Dec 11 08:29:15 crc kubenswrapper[4860]: I1211 08:29:15.847704 4860 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6d5b6d6b67-pxxrf" podUID="d59cd3a6-bd2d-4c74-97d0-265b3fc56dfd" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.123:5353: connect: connection refused" Dec 11 08:29:16 crc kubenswrapper[4860]: E1211 08:29:16.924290 4860 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Dec 11 08:29:16 crc kubenswrapper[4860]: E1211 08:29:16.924707 4860 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n86h96hf5h57bh654h5c9h58h67bh5ch5cdh576h5fbhcbh5bbh5ch66h599h686h5dfhcch658hf9h66h585h97h59dh65ch679h578h597h679h57dq,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-7wlsq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-7cbbd9698f-ft2xt_openstack(b4a3ee06-7c42-4c4a-bce5-d75f9da8983d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 11 08:29:16 crc kubenswrapper[4860]: E1211 08:29:16.928442 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-7cbbd9698f-ft2xt" podUID="b4a3ee06-7c42-4c4a-bce5-d75f9da8983d" Dec 11 08:29:16 crc kubenswrapper[4860]: E1211 08:29:16.937354 4860 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Dec 11 08:29:16 crc kubenswrapper[4860]: E1211 08:29:16.937550 4860 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n59dh5b8hf8h578hf6h646h547h5b8hdh58bhfh584h699h5b8h5cdh657hd4hd8hcch679h7dh7fh7bh644h67h58dh646hb6h68h99h584h686q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-sgw5k,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-768465cc7-f7gwz_openstack(1ee42a0c-5029-44cf-ba8d-720165ac6138): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 11 08:29:16 crc kubenswrapper[4860]: E1211 08:29:16.939545 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-768465cc7-f7gwz" podUID="1ee42a0c-5029-44cf-ba8d-720165ac6138" Dec 11 08:29:17 crc kubenswrapper[4860]: I1211 08:29:17.010828 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-xq969" Dec 11 08:29:17 crc kubenswrapper[4860]: I1211 08:29:17.075080 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-xq969" Dec 11 08:29:17 crc kubenswrapper[4860]: I1211 08:29:17.075545 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-xq969" event={"ID":"a7bc1b0c-9f61-4bd1-8739-6219931987fa","Type":"ContainerDied","Data":"72d58e20b23e3fe5f3bb76949c37ffecf67efe0c996bffb7e8ef8a0db75b9fa6"} Dec 11 08:29:17 crc kubenswrapper[4860]: I1211 08:29:17.075583 4860 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="72d58e20b23e3fe5f3bb76949c37ffecf67efe0c996bffb7e8ef8a0db75b9fa6" Dec 11 08:29:17 crc kubenswrapper[4860]: I1211 08:29:17.131818 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7bc1b0c-9f61-4bd1-8739-6219931987fa-config-data\") pod \"a7bc1b0c-9f61-4bd1-8739-6219931987fa\" (UID: \"a7bc1b0c-9f61-4bd1-8739-6219931987fa\") " Dec 11 08:29:17 crc kubenswrapper[4860]: I1211 08:29:17.132065 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a7bc1b0c-9f61-4bd1-8739-6219931987fa-scripts\") pod \"a7bc1b0c-9f61-4bd1-8739-6219931987fa\" (UID: \"a7bc1b0c-9f61-4bd1-8739-6219931987fa\") " Dec 11 08:29:17 crc kubenswrapper[4860]: I1211 08:29:17.132095 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7bc1b0c-9f61-4bd1-8739-6219931987fa-combined-ca-bundle\") pod \"a7bc1b0c-9f61-4bd1-8739-6219931987fa\" (UID: \"a7bc1b0c-9f61-4bd1-8739-6219931987fa\") " Dec 11 08:29:17 crc kubenswrapper[4860]: I1211 08:29:17.132168 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a7bc1b0c-9f61-4bd1-8739-6219931987fa-credential-keys\") pod \"a7bc1b0c-9f61-4bd1-8739-6219931987fa\" (UID: \"a7bc1b0c-9f61-4bd1-8739-6219931987fa\") " Dec 11 08:29:17 crc kubenswrapper[4860]: I1211 08:29:17.132228 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-75456\" (UniqueName: \"kubernetes.io/projected/a7bc1b0c-9f61-4bd1-8739-6219931987fa-kube-api-access-75456\") pod \"a7bc1b0c-9f61-4bd1-8739-6219931987fa\" (UID: \"a7bc1b0c-9f61-4bd1-8739-6219931987fa\") " Dec 11 08:29:17 crc kubenswrapper[4860]: I1211 08:29:17.132301 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a7bc1b0c-9f61-4bd1-8739-6219931987fa-fernet-keys\") pod \"a7bc1b0c-9f61-4bd1-8739-6219931987fa\" (UID: \"a7bc1b0c-9f61-4bd1-8739-6219931987fa\") " Dec 11 08:29:17 crc kubenswrapper[4860]: I1211 08:29:17.150043 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7bc1b0c-9f61-4bd1-8739-6219931987fa-kube-api-access-75456" (OuterVolumeSpecName: "kube-api-access-75456") pod "a7bc1b0c-9f61-4bd1-8739-6219931987fa" (UID: "a7bc1b0c-9f61-4bd1-8739-6219931987fa"). InnerVolumeSpecName "kube-api-access-75456". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:29:17 crc kubenswrapper[4860]: I1211 08:29:17.153253 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7bc1b0c-9f61-4bd1-8739-6219931987fa-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "a7bc1b0c-9f61-4bd1-8739-6219931987fa" (UID: "a7bc1b0c-9f61-4bd1-8739-6219931987fa"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:29:17 crc kubenswrapper[4860]: I1211 08:29:17.157790 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7bc1b0c-9f61-4bd1-8739-6219931987fa-scripts" (OuterVolumeSpecName: "scripts") pod "a7bc1b0c-9f61-4bd1-8739-6219931987fa" (UID: "a7bc1b0c-9f61-4bd1-8739-6219931987fa"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:29:17 crc kubenswrapper[4860]: I1211 08:29:17.164115 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7bc1b0c-9f61-4bd1-8739-6219931987fa-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "a7bc1b0c-9f61-4bd1-8739-6219931987fa" (UID: "a7bc1b0c-9f61-4bd1-8739-6219931987fa"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:29:17 crc kubenswrapper[4860]: I1211 08:29:17.169676 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7bc1b0c-9f61-4bd1-8739-6219931987fa-config-data" (OuterVolumeSpecName: "config-data") pod "a7bc1b0c-9f61-4bd1-8739-6219931987fa" (UID: "a7bc1b0c-9f61-4bd1-8739-6219931987fa"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:29:17 crc kubenswrapper[4860]: I1211 08:29:17.210084 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7bc1b0c-9f61-4bd1-8739-6219931987fa-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a7bc1b0c-9f61-4bd1-8739-6219931987fa" (UID: "a7bc1b0c-9f61-4bd1-8739-6219931987fa"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:29:17 crc kubenswrapper[4860]: I1211 08:29:17.237598 4860 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a7bc1b0c-9f61-4bd1-8739-6219931987fa-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 11 08:29:17 crc kubenswrapper[4860]: I1211 08:29:17.237658 4860 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7bc1b0c-9f61-4bd1-8739-6219931987fa-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 08:29:17 crc kubenswrapper[4860]: I1211 08:29:17.237672 4860 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a7bc1b0c-9f61-4bd1-8739-6219931987fa-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 08:29:17 crc kubenswrapper[4860]: I1211 08:29:17.237684 4860 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7bc1b0c-9f61-4bd1-8739-6219931987fa-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 08:29:17 crc kubenswrapper[4860]: I1211 08:29:17.237704 4860 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a7bc1b0c-9f61-4bd1-8739-6219931987fa-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 11 08:29:17 crc kubenswrapper[4860]: I1211 08:29:17.237758 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-75456\" (UniqueName: \"kubernetes.io/projected/a7bc1b0c-9f61-4bd1-8739-6219931987fa-kube-api-access-75456\") on node \"crc\" DevicePath \"\"" Dec 11 08:29:18 crc kubenswrapper[4860]: I1211 08:29:18.102087 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-xq969"] Dec 11 08:29:18 crc kubenswrapper[4860]: I1211 08:29:18.111561 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-xq969"] Dec 11 08:29:18 crc kubenswrapper[4860]: I1211 08:29:18.200391 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-plxxc"] Dec 11 08:29:18 crc kubenswrapper[4860]: E1211 08:29:18.200936 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7bc1b0c-9f61-4bd1-8739-6219931987fa" containerName="keystone-bootstrap" Dec 11 08:29:18 crc kubenswrapper[4860]: I1211 08:29:18.200958 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7bc1b0c-9f61-4bd1-8739-6219931987fa" containerName="keystone-bootstrap" Dec 11 08:29:18 crc kubenswrapper[4860]: I1211 08:29:18.201220 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7bc1b0c-9f61-4bd1-8739-6219931987fa" containerName="keystone-bootstrap" Dec 11 08:29:18 crc kubenswrapper[4860]: I1211 08:29:18.202078 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-plxxc" Dec 11 08:29:18 crc kubenswrapper[4860]: I1211 08:29:18.206222 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 11 08:29:18 crc kubenswrapper[4860]: I1211 08:29:18.206741 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-qjrrt" Dec 11 08:29:18 crc kubenswrapper[4860]: I1211 08:29:18.206993 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 11 08:29:18 crc kubenswrapper[4860]: I1211 08:29:18.208100 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 11 08:29:18 crc kubenswrapper[4860]: I1211 08:29:18.209028 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 11 08:29:18 crc kubenswrapper[4860]: I1211 08:29:18.213778 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-plxxc"] Dec 11 08:29:18 crc kubenswrapper[4860]: I1211 08:29:18.359702 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/fecaee75-2dc0-4d48-9d40-11fc1ec8e972-fernet-keys\") pod \"keystone-bootstrap-plxxc\" (UID: \"fecaee75-2dc0-4d48-9d40-11fc1ec8e972\") " pod="openstack/keystone-bootstrap-plxxc" Dec 11 08:29:18 crc kubenswrapper[4860]: I1211 08:29:18.359760 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xzhjl\" (UniqueName: \"kubernetes.io/projected/fecaee75-2dc0-4d48-9d40-11fc1ec8e972-kube-api-access-xzhjl\") pod \"keystone-bootstrap-plxxc\" (UID: \"fecaee75-2dc0-4d48-9d40-11fc1ec8e972\") " pod="openstack/keystone-bootstrap-plxxc" Dec 11 08:29:18 crc kubenswrapper[4860]: I1211 08:29:18.359882 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fecaee75-2dc0-4d48-9d40-11fc1ec8e972-combined-ca-bundle\") pod \"keystone-bootstrap-plxxc\" (UID: \"fecaee75-2dc0-4d48-9d40-11fc1ec8e972\") " pod="openstack/keystone-bootstrap-plxxc" Dec 11 08:29:18 crc kubenswrapper[4860]: I1211 08:29:18.359914 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fecaee75-2dc0-4d48-9d40-11fc1ec8e972-scripts\") pod \"keystone-bootstrap-plxxc\" (UID: \"fecaee75-2dc0-4d48-9d40-11fc1ec8e972\") " pod="openstack/keystone-bootstrap-plxxc" Dec 11 08:29:18 crc kubenswrapper[4860]: I1211 08:29:18.359935 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fecaee75-2dc0-4d48-9d40-11fc1ec8e972-config-data\") pod \"keystone-bootstrap-plxxc\" (UID: \"fecaee75-2dc0-4d48-9d40-11fc1ec8e972\") " pod="openstack/keystone-bootstrap-plxxc" Dec 11 08:29:18 crc kubenswrapper[4860]: I1211 08:29:18.359964 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/fecaee75-2dc0-4d48-9d40-11fc1ec8e972-credential-keys\") pod \"keystone-bootstrap-plxxc\" (UID: \"fecaee75-2dc0-4d48-9d40-11fc1ec8e972\") " pod="openstack/keystone-bootstrap-plxxc" Dec 11 08:29:18 crc kubenswrapper[4860]: I1211 08:29:18.461989 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/fecaee75-2dc0-4d48-9d40-11fc1ec8e972-fernet-keys\") pod \"keystone-bootstrap-plxxc\" (UID: \"fecaee75-2dc0-4d48-9d40-11fc1ec8e972\") " pod="openstack/keystone-bootstrap-plxxc" Dec 11 08:29:18 crc kubenswrapper[4860]: I1211 08:29:18.462050 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xzhjl\" (UniqueName: \"kubernetes.io/projected/fecaee75-2dc0-4d48-9d40-11fc1ec8e972-kube-api-access-xzhjl\") pod \"keystone-bootstrap-plxxc\" (UID: \"fecaee75-2dc0-4d48-9d40-11fc1ec8e972\") " pod="openstack/keystone-bootstrap-plxxc" Dec 11 08:29:18 crc kubenswrapper[4860]: I1211 08:29:18.462103 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fecaee75-2dc0-4d48-9d40-11fc1ec8e972-combined-ca-bundle\") pod \"keystone-bootstrap-plxxc\" (UID: \"fecaee75-2dc0-4d48-9d40-11fc1ec8e972\") " pod="openstack/keystone-bootstrap-plxxc" Dec 11 08:29:18 crc kubenswrapper[4860]: I1211 08:29:18.462135 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fecaee75-2dc0-4d48-9d40-11fc1ec8e972-scripts\") pod \"keystone-bootstrap-plxxc\" (UID: \"fecaee75-2dc0-4d48-9d40-11fc1ec8e972\") " pod="openstack/keystone-bootstrap-plxxc" Dec 11 08:29:18 crc kubenswrapper[4860]: I1211 08:29:18.462160 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fecaee75-2dc0-4d48-9d40-11fc1ec8e972-config-data\") pod \"keystone-bootstrap-plxxc\" (UID: \"fecaee75-2dc0-4d48-9d40-11fc1ec8e972\") " pod="openstack/keystone-bootstrap-plxxc" Dec 11 08:29:18 crc kubenswrapper[4860]: I1211 08:29:18.462205 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/fecaee75-2dc0-4d48-9d40-11fc1ec8e972-credential-keys\") pod \"keystone-bootstrap-plxxc\" (UID: \"fecaee75-2dc0-4d48-9d40-11fc1ec8e972\") " pod="openstack/keystone-bootstrap-plxxc" Dec 11 08:29:18 crc kubenswrapper[4860]: I1211 08:29:18.469454 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fecaee75-2dc0-4d48-9d40-11fc1ec8e972-scripts\") pod \"keystone-bootstrap-plxxc\" (UID: \"fecaee75-2dc0-4d48-9d40-11fc1ec8e972\") " pod="openstack/keystone-bootstrap-plxxc" Dec 11 08:29:18 crc kubenswrapper[4860]: I1211 08:29:18.470443 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/fecaee75-2dc0-4d48-9d40-11fc1ec8e972-fernet-keys\") pod \"keystone-bootstrap-plxxc\" (UID: \"fecaee75-2dc0-4d48-9d40-11fc1ec8e972\") " pod="openstack/keystone-bootstrap-plxxc" Dec 11 08:29:18 crc kubenswrapper[4860]: I1211 08:29:18.475244 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/fecaee75-2dc0-4d48-9d40-11fc1ec8e972-credential-keys\") pod \"keystone-bootstrap-plxxc\" (UID: \"fecaee75-2dc0-4d48-9d40-11fc1ec8e972\") " pod="openstack/keystone-bootstrap-plxxc" Dec 11 08:29:18 crc kubenswrapper[4860]: I1211 08:29:18.479873 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fecaee75-2dc0-4d48-9d40-11fc1ec8e972-combined-ca-bundle\") pod \"keystone-bootstrap-plxxc\" (UID: \"fecaee75-2dc0-4d48-9d40-11fc1ec8e972\") " pod="openstack/keystone-bootstrap-plxxc" Dec 11 08:29:18 crc kubenswrapper[4860]: I1211 08:29:18.483565 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xzhjl\" (UniqueName: \"kubernetes.io/projected/fecaee75-2dc0-4d48-9d40-11fc1ec8e972-kube-api-access-xzhjl\") pod \"keystone-bootstrap-plxxc\" (UID: \"fecaee75-2dc0-4d48-9d40-11fc1ec8e972\") " pod="openstack/keystone-bootstrap-plxxc" Dec 11 08:29:18 crc kubenswrapper[4860]: I1211 08:29:18.484653 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fecaee75-2dc0-4d48-9d40-11fc1ec8e972-config-data\") pod \"keystone-bootstrap-plxxc\" (UID: \"fecaee75-2dc0-4d48-9d40-11fc1ec8e972\") " pod="openstack/keystone-bootstrap-plxxc" Dec 11 08:29:18 crc kubenswrapper[4860]: I1211 08:29:18.528299 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-plxxc" Dec 11 08:29:19 crc kubenswrapper[4860]: I1211 08:29:19.591096 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a7bc1b0c-9f61-4bd1-8739-6219931987fa" path="/var/lib/kubelet/pods/a7bc1b0c-9f61-4bd1-8739-6219931987fa/volumes" Dec 11 08:29:21 crc kubenswrapper[4860]: I1211 08:29:21.115392 4860 generic.go:334] "Generic (PLEG): container finished" podID="cf433d5d-8c95-4d72-864c-3b74f315dd2b" containerID="d737396b95b4dbb30ba8c85e823e3e330e74286479aa940ead93372b72ec06f3" exitCode=0 Dec 11 08:29:21 crc kubenswrapper[4860]: I1211 08:29:21.115541 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-4fsqr" event={"ID":"cf433d5d-8c95-4d72-864c-3b74f315dd2b","Type":"ContainerDied","Data":"d737396b95b4dbb30ba8c85e823e3e330e74286479aa940ead93372b72ec06f3"} Dec 11 08:29:25 crc kubenswrapper[4860]: I1211 08:29:25.847932 4860 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6d5b6d6b67-pxxrf" podUID="d59cd3a6-bd2d-4c74-97d0-265b3fc56dfd" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.123:5353: i/o timeout" Dec 11 08:29:25 crc kubenswrapper[4860]: I1211 08:29:25.848773 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6d5b6d6b67-pxxrf" Dec 11 08:29:27 crc kubenswrapper[4860]: E1211 08:29:27.175681 4860 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Dec 11 08:29:27 crc kubenswrapper[4860]: E1211 08:29:27.175908 4860 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n85h5dh5bbhd9h5f4h54bh568hfdhf6h58h549h554h8ch5c5h5cbh5d7h657h595h5ffh686hbfh67h5f8hdh7hd9h558h696h585h656h588h677q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-fjxs5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-6b45499b8f-8xhjk_openstack(5c9ab8b4-ed9e-46a5-b553-3d085ff20722): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 11 08:29:27 crc kubenswrapper[4860]: E1211 08:29:27.184830 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-6b45499b8f-8xhjk" podUID="5c9ab8b4-ed9e-46a5-b553-3d085ff20722" Dec 11 08:29:27 crc kubenswrapper[4860]: I1211 08:29:27.304701 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7cbbd9698f-ft2xt" Dec 11 08:29:27 crc kubenswrapper[4860]: I1211 08:29:27.316570 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-768465cc7-f7gwz" Dec 11 08:29:27 crc kubenswrapper[4860]: I1211 08:29:27.477507 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b4a3ee06-7c42-4c4a-bce5-d75f9da8983d-logs\") pod \"b4a3ee06-7c42-4c4a-bce5-d75f9da8983d\" (UID: \"b4a3ee06-7c42-4c4a-bce5-d75f9da8983d\") " Dec 11 08:29:27 crc kubenswrapper[4860]: I1211 08:29:27.477575 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b4a3ee06-7c42-4c4a-bce5-d75f9da8983d-scripts\") pod \"b4a3ee06-7c42-4c4a-bce5-d75f9da8983d\" (UID: \"b4a3ee06-7c42-4c4a-bce5-d75f9da8983d\") " Dec 11 08:29:27 crc kubenswrapper[4860]: I1211 08:29:27.477671 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sgw5k\" (UniqueName: \"kubernetes.io/projected/1ee42a0c-5029-44cf-ba8d-720165ac6138-kube-api-access-sgw5k\") pod \"1ee42a0c-5029-44cf-ba8d-720165ac6138\" (UID: \"1ee42a0c-5029-44cf-ba8d-720165ac6138\") " Dec 11 08:29:27 crc kubenswrapper[4860]: I1211 08:29:27.477719 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7wlsq\" (UniqueName: \"kubernetes.io/projected/b4a3ee06-7c42-4c4a-bce5-d75f9da8983d-kube-api-access-7wlsq\") pod \"b4a3ee06-7c42-4c4a-bce5-d75f9da8983d\" (UID: \"b4a3ee06-7c42-4c4a-bce5-d75f9da8983d\") " Dec 11 08:29:27 crc kubenswrapper[4860]: I1211 08:29:27.477766 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/1ee42a0c-5029-44cf-ba8d-720165ac6138-horizon-secret-key\") pod \"1ee42a0c-5029-44cf-ba8d-720165ac6138\" (UID: \"1ee42a0c-5029-44cf-ba8d-720165ac6138\") " Dec 11 08:29:27 crc kubenswrapper[4860]: I1211 08:29:27.477807 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1ee42a0c-5029-44cf-ba8d-720165ac6138-config-data\") pod \"1ee42a0c-5029-44cf-ba8d-720165ac6138\" (UID: \"1ee42a0c-5029-44cf-ba8d-720165ac6138\") " Dec 11 08:29:27 crc kubenswrapper[4860]: I1211 08:29:27.477850 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1ee42a0c-5029-44cf-ba8d-720165ac6138-logs\") pod \"1ee42a0c-5029-44cf-ba8d-720165ac6138\" (UID: \"1ee42a0c-5029-44cf-ba8d-720165ac6138\") " Dec 11 08:29:27 crc kubenswrapper[4860]: I1211 08:29:27.477950 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b4a3ee06-7c42-4c4a-bce5-d75f9da8983d-config-data\") pod \"b4a3ee06-7c42-4c4a-bce5-d75f9da8983d\" (UID: \"b4a3ee06-7c42-4c4a-bce5-d75f9da8983d\") " Dec 11 08:29:27 crc kubenswrapper[4860]: I1211 08:29:27.477972 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b4a3ee06-7c42-4c4a-bce5-d75f9da8983d-logs" (OuterVolumeSpecName: "logs") pod "b4a3ee06-7c42-4c4a-bce5-d75f9da8983d" (UID: "b4a3ee06-7c42-4c4a-bce5-d75f9da8983d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:29:27 crc kubenswrapper[4860]: I1211 08:29:27.478025 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1ee42a0c-5029-44cf-ba8d-720165ac6138-scripts\") pod \"1ee42a0c-5029-44cf-ba8d-720165ac6138\" (UID: \"1ee42a0c-5029-44cf-ba8d-720165ac6138\") " Dec 11 08:29:27 crc kubenswrapper[4860]: I1211 08:29:27.478085 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/b4a3ee06-7c42-4c4a-bce5-d75f9da8983d-horizon-secret-key\") pod \"b4a3ee06-7c42-4c4a-bce5-d75f9da8983d\" (UID: \"b4a3ee06-7c42-4c4a-bce5-d75f9da8983d\") " Dec 11 08:29:27 crc kubenswrapper[4860]: I1211 08:29:27.478115 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b4a3ee06-7c42-4c4a-bce5-d75f9da8983d-scripts" (OuterVolumeSpecName: "scripts") pod "b4a3ee06-7c42-4c4a-bce5-d75f9da8983d" (UID: "b4a3ee06-7c42-4c4a-bce5-d75f9da8983d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:29:27 crc kubenswrapper[4860]: I1211 08:29:27.478299 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1ee42a0c-5029-44cf-ba8d-720165ac6138-logs" (OuterVolumeSpecName: "logs") pod "1ee42a0c-5029-44cf-ba8d-720165ac6138" (UID: "1ee42a0c-5029-44cf-ba8d-720165ac6138"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:29:27 crc kubenswrapper[4860]: I1211 08:29:27.478739 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1ee42a0c-5029-44cf-ba8d-720165ac6138-config-data" (OuterVolumeSpecName: "config-data") pod "1ee42a0c-5029-44cf-ba8d-720165ac6138" (UID: "1ee42a0c-5029-44cf-ba8d-720165ac6138"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:29:27 crc kubenswrapper[4860]: I1211 08:29:27.479101 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1ee42a0c-5029-44cf-ba8d-720165ac6138-scripts" (OuterVolumeSpecName: "scripts") pod "1ee42a0c-5029-44cf-ba8d-720165ac6138" (UID: "1ee42a0c-5029-44cf-ba8d-720165ac6138"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:29:27 crc kubenswrapper[4860]: I1211 08:29:27.479220 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b4a3ee06-7c42-4c4a-bce5-d75f9da8983d-config-data" (OuterVolumeSpecName: "config-data") pod "b4a3ee06-7c42-4c4a-bce5-d75f9da8983d" (UID: "b4a3ee06-7c42-4c4a-bce5-d75f9da8983d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:29:27 crc kubenswrapper[4860]: I1211 08:29:27.480037 4860 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b4a3ee06-7c42-4c4a-bce5-d75f9da8983d-logs\") on node \"crc\" DevicePath \"\"" Dec 11 08:29:27 crc kubenswrapper[4860]: I1211 08:29:27.480061 4860 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b4a3ee06-7c42-4c4a-bce5-d75f9da8983d-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 08:29:27 crc kubenswrapper[4860]: I1211 08:29:27.480076 4860 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1ee42a0c-5029-44cf-ba8d-720165ac6138-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 08:29:27 crc kubenswrapper[4860]: I1211 08:29:27.480087 4860 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1ee42a0c-5029-44cf-ba8d-720165ac6138-logs\") on node \"crc\" DevicePath \"\"" Dec 11 08:29:27 crc kubenswrapper[4860]: I1211 08:29:27.480098 4860 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b4a3ee06-7c42-4c4a-bce5-d75f9da8983d-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 08:29:27 crc kubenswrapper[4860]: I1211 08:29:27.480109 4860 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1ee42a0c-5029-44cf-ba8d-720165ac6138-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 08:29:27 crc kubenswrapper[4860]: I1211 08:29:27.485098 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ee42a0c-5029-44cf-ba8d-720165ac6138-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "1ee42a0c-5029-44cf-ba8d-720165ac6138" (UID: "1ee42a0c-5029-44cf-ba8d-720165ac6138"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:29:27 crc kubenswrapper[4860]: I1211 08:29:27.485103 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b4a3ee06-7c42-4c4a-bce5-d75f9da8983d-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "b4a3ee06-7c42-4c4a-bce5-d75f9da8983d" (UID: "b4a3ee06-7c42-4c4a-bce5-d75f9da8983d"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:29:27 crc kubenswrapper[4860]: I1211 08:29:27.485749 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ee42a0c-5029-44cf-ba8d-720165ac6138-kube-api-access-sgw5k" (OuterVolumeSpecName: "kube-api-access-sgw5k") pod "1ee42a0c-5029-44cf-ba8d-720165ac6138" (UID: "1ee42a0c-5029-44cf-ba8d-720165ac6138"). InnerVolumeSpecName "kube-api-access-sgw5k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:29:27 crc kubenswrapper[4860]: I1211 08:29:27.486601 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b4a3ee06-7c42-4c4a-bce5-d75f9da8983d-kube-api-access-7wlsq" (OuterVolumeSpecName: "kube-api-access-7wlsq") pod "b4a3ee06-7c42-4c4a-bce5-d75f9da8983d" (UID: "b4a3ee06-7c42-4c4a-bce5-d75f9da8983d"). InnerVolumeSpecName "kube-api-access-7wlsq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:29:27 crc kubenswrapper[4860]: I1211 08:29:27.581509 4860 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/b4a3ee06-7c42-4c4a-bce5-d75f9da8983d-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 11 08:29:27 crc kubenswrapper[4860]: I1211 08:29:27.581551 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sgw5k\" (UniqueName: \"kubernetes.io/projected/1ee42a0c-5029-44cf-ba8d-720165ac6138-kube-api-access-sgw5k\") on node \"crc\" DevicePath \"\"" Dec 11 08:29:27 crc kubenswrapper[4860]: I1211 08:29:27.581561 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7wlsq\" (UniqueName: \"kubernetes.io/projected/b4a3ee06-7c42-4c4a-bce5-d75f9da8983d-kube-api-access-7wlsq\") on node \"crc\" DevicePath \"\"" Dec 11 08:29:27 crc kubenswrapper[4860]: I1211 08:29:27.581570 4860 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/1ee42a0c-5029-44cf-ba8d-720165ac6138-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 11 08:29:27 crc kubenswrapper[4860]: E1211 08:29:27.622538 4860 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified" Dec 11 08:29:27 crc kubenswrapper[4860]: E1211 08:29:27.622732 4860 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nd8h689h7fhc8h579h79h55dh669h8fhddh5c9h6bh64fh67fh5cdhffh68dh5ddh644h574h597h677h549h545h5d7hcbh679h579h697h557h58dh5f6q,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vdbx8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(040a0e8c-efd3-4e3b-be51-59e311ae4406): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 11 08:29:27 crc kubenswrapper[4860]: I1211 08:29:27.682403 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d5b6d6b67-pxxrf" Dec 11 08:29:27 crc kubenswrapper[4860]: I1211 08:29:27.702162 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-4fsqr" Dec 11 08:29:27 crc kubenswrapper[4860]: I1211 08:29:27.784613 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d59cd3a6-bd2d-4c74-97d0-265b3fc56dfd-ovsdbserver-sb\") pod \"d59cd3a6-bd2d-4c74-97d0-265b3fc56dfd\" (UID: \"d59cd3a6-bd2d-4c74-97d0-265b3fc56dfd\") " Dec 11 08:29:27 crc kubenswrapper[4860]: I1211 08:29:27.784710 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d59cd3a6-bd2d-4c74-97d0-265b3fc56dfd-dns-swift-storage-0\") pod \"d59cd3a6-bd2d-4c74-97d0-265b3fc56dfd\" (UID: \"d59cd3a6-bd2d-4c74-97d0-265b3fc56dfd\") " Dec 11 08:29:27 crc kubenswrapper[4860]: I1211 08:29:27.784943 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d59cd3a6-bd2d-4c74-97d0-265b3fc56dfd-dns-svc\") pod \"d59cd3a6-bd2d-4c74-97d0-265b3fc56dfd\" (UID: \"d59cd3a6-bd2d-4c74-97d0-265b3fc56dfd\") " Dec 11 08:29:27 crc kubenswrapper[4860]: I1211 08:29:27.785004 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d59cd3a6-bd2d-4c74-97d0-265b3fc56dfd-ovsdbserver-nb\") pod \"d59cd3a6-bd2d-4c74-97d0-265b3fc56dfd\" (UID: \"d59cd3a6-bd2d-4c74-97d0-265b3fc56dfd\") " Dec 11 08:29:27 crc kubenswrapper[4860]: I1211 08:29:27.785074 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d59cd3a6-bd2d-4c74-97d0-265b3fc56dfd-config\") pod \"d59cd3a6-bd2d-4c74-97d0-265b3fc56dfd\" (UID: \"d59cd3a6-bd2d-4c74-97d0-265b3fc56dfd\") " Dec 11 08:29:27 crc kubenswrapper[4860]: I1211 08:29:27.785157 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zz27s\" (UniqueName: \"kubernetes.io/projected/d59cd3a6-bd2d-4c74-97d0-265b3fc56dfd-kube-api-access-zz27s\") pod \"d59cd3a6-bd2d-4c74-97d0-265b3fc56dfd\" (UID: \"d59cd3a6-bd2d-4c74-97d0-265b3fc56dfd\") " Dec 11 08:29:27 crc kubenswrapper[4860]: I1211 08:29:27.790567 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d59cd3a6-bd2d-4c74-97d0-265b3fc56dfd-kube-api-access-zz27s" (OuterVolumeSpecName: "kube-api-access-zz27s") pod "d59cd3a6-bd2d-4c74-97d0-265b3fc56dfd" (UID: "d59cd3a6-bd2d-4c74-97d0-265b3fc56dfd"). InnerVolumeSpecName "kube-api-access-zz27s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:29:27 crc kubenswrapper[4860]: I1211 08:29:27.829856 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d59cd3a6-bd2d-4c74-97d0-265b3fc56dfd-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "d59cd3a6-bd2d-4c74-97d0-265b3fc56dfd" (UID: "d59cd3a6-bd2d-4c74-97d0-265b3fc56dfd"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:29:27 crc kubenswrapper[4860]: I1211 08:29:27.835799 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d59cd3a6-bd2d-4c74-97d0-265b3fc56dfd-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "d59cd3a6-bd2d-4c74-97d0-265b3fc56dfd" (UID: "d59cd3a6-bd2d-4c74-97d0-265b3fc56dfd"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:29:27 crc kubenswrapper[4860]: I1211 08:29:27.837104 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d59cd3a6-bd2d-4c74-97d0-265b3fc56dfd-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "d59cd3a6-bd2d-4c74-97d0-265b3fc56dfd" (UID: "d59cd3a6-bd2d-4c74-97d0-265b3fc56dfd"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:29:27 crc kubenswrapper[4860]: I1211 08:29:27.842447 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d59cd3a6-bd2d-4c74-97d0-265b3fc56dfd-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d59cd3a6-bd2d-4c74-97d0-265b3fc56dfd" (UID: "d59cd3a6-bd2d-4c74-97d0-265b3fc56dfd"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:29:27 crc kubenswrapper[4860]: I1211 08:29:27.850492 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d59cd3a6-bd2d-4c74-97d0-265b3fc56dfd-config" (OuterVolumeSpecName: "config") pod "d59cd3a6-bd2d-4c74-97d0-265b3fc56dfd" (UID: "d59cd3a6-bd2d-4c74-97d0-265b3fc56dfd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:29:27 crc kubenswrapper[4860]: I1211 08:29:27.886673 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/cf433d5d-8c95-4d72-864c-3b74f315dd2b-db-sync-config-data\") pod \"cf433d5d-8c95-4d72-864c-3b74f315dd2b\" (UID: \"cf433d5d-8c95-4d72-864c-3b74f315dd2b\") " Dec 11 08:29:27 crc kubenswrapper[4860]: I1211 08:29:27.886753 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf433d5d-8c95-4d72-864c-3b74f315dd2b-config-data\") pod \"cf433d5d-8c95-4d72-864c-3b74f315dd2b\" (UID: \"cf433d5d-8c95-4d72-864c-3b74f315dd2b\") " Dec 11 08:29:27 crc kubenswrapper[4860]: I1211 08:29:27.886803 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf433d5d-8c95-4d72-864c-3b74f315dd2b-combined-ca-bundle\") pod \"cf433d5d-8c95-4d72-864c-3b74f315dd2b\" (UID: \"cf433d5d-8c95-4d72-864c-3b74f315dd2b\") " Dec 11 08:29:27 crc kubenswrapper[4860]: I1211 08:29:27.886973 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cmvl7\" (UniqueName: \"kubernetes.io/projected/cf433d5d-8c95-4d72-864c-3b74f315dd2b-kube-api-access-cmvl7\") pod \"cf433d5d-8c95-4d72-864c-3b74f315dd2b\" (UID: \"cf433d5d-8c95-4d72-864c-3b74f315dd2b\") " Dec 11 08:29:27 crc kubenswrapper[4860]: I1211 08:29:27.887335 4860 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d59cd3a6-bd2d-4c74-97d0-265b3fc56dfd-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 11 08:29:27 crc kubenswrapper[4860]: I1211 08:29:27.887355 4860 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d59cd3a6-bd2d-4c74-97d0-265b3fc56dfd-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 11 08:29:27 crc kubenswrapper[4860]: I1211 08:29:27.887364 4860 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d59cd3a6-bd2d-4c74-97d0-265b3fc56dfd-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 11 08:29:27 crc kubenswrapper[4860]: I1211 08:29:27.887374 4860 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d59cd3a6-bd2d-4c74-97d0-265b3fc56dfd-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 11 08:29:27 crc kubenswrapper[4860]: I1211 08:29:27.887385 4860 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d59cd3a6-bd2d-4c74-97d0-265b3fc56dfd-config\") on node \"crc\" DevicePath \"\"" Dec 11 08:29:27 crc kubenswrapper[4860]: I1211 08:29:27.887394 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zz27s\" (UniqueName: \"kubernetes.io/projected/d59cd3a6-bd2d-4c74-97d0-265b3fc56dfd-kube-api-access-zz27s\") on node \"crc\" DevicePath \"\"" Dec 11 08:29:27 crc kubenswrapper[4860]: I1211 08:29:27.891002 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cf433d5d-8c95-4d72-864c-3b74f315dd2b-kube-api-access-cmvl7" (OuterVolumeSpecName: "kube-api-access-cmvl7") pod "cf433d5d-8c95-4d72-864c-3b74f315dd2b" (UID: "cf433d5d-8c95-4d72-864c-3b74f315dd2b"). InnerVolumeSpecName "kube-api-access-cmvl7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:29:27 crc kubenswrapper[4860]: I1211 08:29:27.892998 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf433d5d-8c95-4d72-864c-3b74f315dd2b-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "cf433d5d-8c95-4d72-864c-3b74f315dd2b" (UID: "cf433d5d-8c95-4d72-864c-3b74f315dd2b"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:29:27 crc kubenswrapper[4860]: I1211 08:29:27.912465 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf433d5d-8c95-4d72-864c-3b74f315dd2b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cf433d5d-8c95-4d72-864c-3b74f315dd2b" (UID: "cf433d5d-8c95-4d72-864c-3b74f315dd2b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:29:27 crc kubenswrapper[4860]: I1211 08:29:27.932267 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf433d5d-8c95-4d72-864c-3b74f315dd2b-config-data" (OuterVolumeSpecName: "config-data") pod "cf433d5d-8c95-4d72-864c-3b74f315dd2b" (UID: "cf433d5d-8c95-4d72-864c-3b74f315dd2b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:29:27 crc kubenswrapper[4860]: I1211 08:29:27.989259 4860 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/cf433d5d-8c95-4d72-864c-3b74f315dd2b-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 08:29:27 crc kubenswrapper[4860]: I1211 08:29:27.989291 4860 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf433d5d-8c95-4d72-864c-3b74f315dd2b-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 08:29:27 crc kubenswrapper[4860]: I1211 08:29:27.989302 4860 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf433d5d-8c95-4d72-864c-3b74f315dd2b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 08:29:27 crc kubenswrapper[4860]: I1211 08:29:27.989312 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cmvl7\" (UniqueName: \"kubernetes.io/projected/cf433d5d-8c95-4d72-864c-3b74f315dd2b-kube-api-access-cmvl7\") on node \"crc\" DevicePath \"\"" Dec 11 08:29:28 crc kubenswrapper[4860]: I1211 08:29:28.176613 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d5b6d6b67-pxxrf" event={"ID":"d59cd3a6-bd2d-4c74-97d0-265b3fc56dfd","Type":"ContainerDied","Data":"a3afb983675c4e177f06e0451a9e72be824cb1f7e1200cae5923a1122ff43b00"} Dec 11 08:29:28 crc kubenswrapper[4860]: I1211 08:29:28.176711 4860 scope.go:117] "RemoveContainer" containerID="8458e72beef31344067e7052e8cb6285a95889bd51a203b4e3d4a97d4c4e0d2c" Dec 11 08:29:28 crc kubenswrapper[4860]: I1211 08:29:28.176765 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d5b6d6b67-pxxrf" Dec 11 08:29:28 crc kubenswrapper[4860]: I1211 08:29:28.180947 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-768465cc7-f7gwz" Dec 11 08:29:28 crc kubenswrapper[4860]: I1211 08:29:28.180965 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-768465cc7-f7gwz" event={"ID":"1ee42a0c-5029-44cf-ba8d-720165ac6138","Type":"ContainerDied","Data":"dabe4e3751a6cae5d7034bf68fdb500a6600bbf56561b35df1730fca3dbb7ba6"} Dec 11 08:29:28 crc kubenswrapper[4860]: I1211 08:29:28.183007 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7cbbd9698f-ft2xt" Dec 11 08:29:28 crc kubenswrapper[4860]: I1211 08:29:28.183005 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7cbbd9698f-ft2xt" event={"ID":"b4a3ee06-7c42-4c4a-bce5-d75f9da8983d","Type":"ContainerDied","Data":"8b1e36928dc90198ac3111a6937367dae82532daf1167906a97b1c1979bd0a00"} Dec 11 08:29:28 crc kubenswrapper[4860]: I1211 08:29:28.226828 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-4fsqr" Dec 11 08:29:28 crc kubenswrapper[4860]: I1211 08:29:28.229287 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-4fsqr" event={"ID":"cf433d5d-8c95-4d72-864c-3b74f315dd2b","Type":"ContainerDied","Data":"3fd59619fd28c03de0bc2164837abea1e39a9f1b0fcb738f470203d59015276b"} Dec 11 08:29:28 crc kubenswrapper[4860]: I1211 08:29:28.229331 4860 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3fd59619fd28c03de0bc2164837abea1e39a9f1b0fcb738f470203d59015276b" Dec 11 08:29:28 crc kubenswrapper[4860]: I1211 08:29:28.286611 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7cbbd9698f-ft2xt"] Dec 11 08:29:28 crc kubenswrapper[4860]: I1211 08:29:28.299342 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-7cbbd9698f-ft2xt"] Dec 11 08:29:28 crc kubenswrapper[4860]: I1211 08:29:28.315292 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d5b6d6b67-pxxrf"] Dec 11 08:29:28 crc kubenswrapper[4860]: I1211 08:29:28.344627 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6d5b6d6b67-pxxrf"] Dec 11 08:29:28 crc kubenswrapper[4860]: I1211 08:29:28.386775 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-768465cc7-f7gwz"] Dec 11 08:29:28 crc kubenswrapper[4860]: I1211 08:29:28.395849 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-768465cc7-f7gwz"] Dec 11 08:29:29 crc kubenswrapper[4860]: I1211 08:29:29.198411 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57c957c4ff-wxx2m"] Dec 11 08:29:29 crc kubenswrapper[4860]: E1211 08:29:29.199138 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d59cd3a6-bd2d-4c74-97d0-265b3fc56dfd" containerName="init" Dec 11 08:29:29 crc kubenswrapper[4860]: I1211 08:29:29.199152 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="d59cd3a6-bd2d-4c74-97d0-265b3fc56dfd" containerName="init" Dec 11 08:29:29 crc kubenswrapper[4860]: E1211 08:29:29.199196 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d59cd3a6-bd2d-4c74-97d0-265b3fc56dfd" containerName="dnsmasq-dns" Dec 11 08:29:29 crc kubenswrapper[4860]: I1211 08:29:29.199203 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="d59cd3a6-bd2d-4c74-97d0-265b3fc56dfd" containerName="dnsmasq-dns" Dec 11 08:29:29 crc kubenswrapper[4860]: E1211 08:29:29.199225 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf433d5d-8c95-4d72-864c-3b74f315dd2b" containerName="glance-db-sync" Dec 11 08:29:29 crc kubenswrapper[4860]: I1211 08:29:29.199231 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf433d5d-8c95-4d72-864c-3b74f315dd2b" containerName="glance-db-sync" Dec 11 08:29:29 crc kubenswrapper[4860]: I1211 08:29:29.199459 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="d59cd3a6-bd2d-4c74-97d0-265b3fc56dfd" containerName="dnsmasq-dns" Dec 11 08:29:29 crc kubenswrapper[4860]: I1211 08:29:29.199476 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf433d5d-8c95-4d72-864c-3b74f315dd2b" containerName="glance-db-sync" Dec 11 08:29:29 crc kubenswrapper[4860]: I1211 08:29:29.200735 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57c957c4ff-wxx2m" Dec 11 08:29:29 crc kubenswrapper[4860]: I1211 08:29:29.227614 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57c957c4ff-wxx2m"] Dec 11 08:29:29 crc kubenswrapper[4860]: I1211 08:29:29.250452 4860 generic.go:334] "Generic (PLEG): container finished" podID="e3822ab5-d1a7-46da-9533-21b67bd240eb" containerID="650883b438680b9c1ab516bfe7c40142cef24b094b49f32d2439af5c7e7508d6" exitCode=0 Dec 11 08:29:29 crc kubenswrapper[4860]: I1211 08:29:29.250724 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-wgk5t" event={"ID":"e3822ab5-d1a7-46da-9533-21b67bd240eb","Type":"ContainerDied","Data":"650883b438680b9c1ab516bfe7c40142cef24b094b49f32d2439af5c7e7508d6"} Dec 11 08:29:29 crc kubenswrapper[4860]: I1211 08:29:29.314212 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/79626ad3-4978-43f0-b163-fe482aa1dac7-ovsdbserver-sb\") pod \"dnsmasq-dns-57c957c4ff-wxx2m\" (UID: \"79626ad3-4978-43f0-b163-fe482aa1dac7\") " pod="openstack/dnsmasq-dns-57c957c4ff-wxx2m" Dec 11 08:29:29 crc kubenswrapper[4860]: I1211 08:29:29.314611 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/79626ad3-4978-43f0-b163-fe482aa1dac7-dns-svc\") pod \"dnsmasq-dns-57c957c4ff-wxx2m\" (UID: \"79626ad3-4978-43f0-b163-fe482aa1dac7\") " pod="openstack/dnsmasq-dns-57c957c4ff-wxx2m" Dec 11 08:29:29 crc kubenswrapper[4860]: I1211 08:29:29.314666 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/79626ad3-4978-43f0-b163-fe482aa1dac7-ovsdbserver-nb\") pod \"dnsmasq-dns-57c957c4ff-wxx2m\" (UID: \"79626ad3-4978-43f0-b163-fe482aa1dac7\") " pod="openstack/dnsmasq-dns-57c957c4ff-wxx2m" Dec 11 08:29:29 crc kubenswrapper[4860]: I1211 08:29:29.314701 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/79626ad3-4978-43f0-b163-fe482aa1dac7-dns-swift-storage-0\") pod \"dnsmasq-dns-57c957c4ff-wxx2m\" (UID: \"79626ad3-4978-43f0-b163-fe482aa1dac7\") " pod="openstack/dnsmasq-dns-57c957c4ff-wxx2m" Dec 11 08:29:29 crc kubenswrapper[4860]: I1211 08:29:29.314746 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q5qmt\" (UniqueName: \"kubernetes.io/projected/79626ad3-4978-43f0-b163-fe482aa1dac7-kube-api-access-q5qmt\") pod \"dnsmasq-dns-57c957c4ff-wxx2m\" (UID: \"79626ad3-4978-43f0-b163-fe482aa1dac7\") " pod="openstack/dnsmasq-dns-57c957c4ff-wxx2m" Dec 11 08:29:29 crc kubenswrapper[4860]: I1211 08:29:29.314878 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/79626ad3-4978-43f0-b163-fe482aa1dac7-config\") pod \"dnsmasq-dns-57c957c4ff-wxx2m\" (UID: \"79626ad3-4978-43f0-b163-fe482aa1dac7\") " pod="openstack/dnsmasq-dns-57c957c4ff-wxx2m" Dec 11 08:29:29 crc kubenswrapper[4860]: I1211 08:29:29.415944 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/79626ad3-4978-43f0-b163-fe482aa1dac7-dns-svc\") pod \"dnsmasq-dns-57c957c4ff-wxx2m\" (UID: \"79626ad3-4978-43f0-b163-fe482aa1dac7\") " pod="openstack/dnsmasq-dns-57c957c4ff-wxx2m" Dec 11 08:29:29 crc kubenswrapper[4860]: I1211 08:29:29.415998 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/79626ad3-4978-43f0-b163-fe482aa1dac7-ovsdbserver-nb\") pod \"dnsmasq-dns-57c957c4ff-wxx2m\" (UID: \"79626ad3-4978-43f0-b163-fe482aa1dac7\") " pod="openstack/dnsmasq-dns-57c957c4ff-wxx2m" Dec 11 08:29:29 crc kubenswrapper[4860]: I1211 08:29:29.416051 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/79626ad3-4978-43f0-b163-fe482aa1dac7-dns-swift-storage-0\") pod \"dnsmasq-dns-57c957c4ff-wxx2m\" (UID: \"79626ad3-4978-43f0-b163-fe482aa1dac7\") " pod="openstack/dnsmasq-dns-57c957c4ff-wxx2m" Dec 11 08:29:29 crc kubenswrapper[4860]: I1211 08:29:29.416095 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q5qmt\" (UniqueName: \"kubernetes.io/projected/79626ad3-4978-43f0-b163-fe482aa1dac7-kube-api-access-q5qmt\") pod \"dnsmasq-dns-57c957c4ff-wxx2m\" (UID: \"79626ad3-4978-43f0-b163-fe482aa1dac7\") " pod="openstack/dnsmasq-dns-57c957c4ff-wxx2m" Dec 11 08:29:29 crc kubenswrapper[4860]: I1211 08:29:29.416189 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/79626ad3-4978-43f0-b163-fe482aa1dac7-config\") pod \"dnsmasq-dns-57c957c4ff-wxx2m\" (UID: \"79626ad3-4978-43f0-b163-fe482aa1dac7\") " pod="openstack/dnsmasq-dns-57c957c4ff-wxx2m" Dec 11 08:29:29 crc kubenswrapper[4860]: I1211 08:29:29.416220 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/79626ad3-4978-43f0-b163-fe482aa1dac7-ovsdbserver-sb\") pod \"dnsmasq-dns-57c957c4ff-wxx2m\" (UID: \"79626ad3-4978-43f0-b163-fe482aa1dac7\") " pod="openstack/dnsmasq-dns-57c957c4ff-wxx2m" Dec 11 08:29:29 crc kubenswrapper[4860]: I1211 08:29:29.417019 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/79626ad3-4978-43f0-b163-fe482aa1dac7-dns-svc\") pod \"dnsmasq-dns-57c957c4ff-wxx2m\" (UID: \"79626ad3-4978-43f0-b163-fe482aa1dac7\") " pod="openstack/dnsmasq-dns-57c957c4ff-wxx2m" Dec 11 08:29:29 crc kubenswrapper[4860]: I1211 08:29:29.417200 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/79626ad3-4978-43f0-b163-fe482aa1dac7-ovsdbserver-sb\") pod \"dnsmasq-dns-57c957c4ff-wxx2m\" (UID: \"79626ad3-4978-43f0-b163-fe482aa1dac7\") " pod="openstack/dnsmasq-dns-57c957c4ff-wxx2m" Dec 11 08:29:29 crc kubenswrapper[4860]: I1211 08:29:29.417242 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/79626ad3-4978-43f0-b163-fe482aa1dac7-dns-swift-storage-0\") pod \"dnsmasq-dns-57c957c4ff-wxx2m\" (UID: \"79626ad3-4978-43f0-b163-fe482aa1dac7\") " pod="openstack/dnsmasq-dns-57c957c4ff-wxx2m" Dec 11 08:29:29 crc kubenswrapper[4860]: I1211 08:29:29.417773 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/79626ad3-4978-43f0-b163-fe482aa1dac7-config\") pod \"dnsmasq-dns-57c957c4ff-wxx2m\" (UID: \"79626ad3-4978-43f0-b163-fe482aa1dac7\") " pod="openstack/dnsmasq-dns-57c957c4ff-wxx2m" Dec 11 08:29:29 crc kubenswrapper[4860]: I1211 08:29:29.417949 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/79626ad3-4978-43f0-b163-fe482aa1dac7-ovsdbserver-nb\") pod \"dnsmasq-dns-57c957c4ff-wxx2m\" (UID: \"79626ad3-4978-43f0-b163-fe482aa1dac7\") " pod="openstack/dnsmasq-dns-57c957c4ff-wxx2m" Dec 11 08:29:29 crc kubenswrapper[4860]: I1211 08:29:29.463665 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q5qmt\" (UniqueName: \"kubernetes.io/projected/79626ad3-4978-43f0-b163-fe482aa1dac7-kube-api-access-q5qmt\") pod \"dnsmasq-dns-57c957c4ff-wxx2m\" (UID: \"79626ad3-4978-43f0-b163-fe482aa1dac7\") " pod="openstack/dnsmasq-dns-57c957c4ff-wxx2m" Dec 11 08:29:29 crc kubenswrapper[4860]: I1211 08:29:29.528278 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57c957c4ff-wxx2m" Dec 11 08:29:29 crc kubenswrapper[4860]: I1211 08:29:29.603487 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1ee42a0c-5029-44cf-ba8d-720165ac6138" path="/var/lib/kubelet/pods/1ee42a0c-5029-44cf-ba8d-720165ac6138/volumes" Dec 11 08:29:29 crc kubenswrapper[4860]: I1211 08:29:29.603996 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b4a3ee06-7c42-4c4a-bce5-d75f9da8983d" path="/var/lib/kubelet/pods/b4a3ee06-7c42-4c4a-bce5-d75f9da8983d/volumes" Dec 11 08:29:29 crc kubenswrapper[4860]: I1211 08:29:29.604345 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d59cd3a6-bd2d-4c74-97d0-265b3fc56dfd" path="/var/lib/kubelet/pods/d59cd3a6-bd2d-4c74-97d0-265b3fc56dfd/volumes" Dec 11 08:29:29 crc kubenswrapper[4860]: E1211 08:29:29.802879 4860 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Dec 11 08:29:29 crc kubenswrapper[4860]: E1211 08:29:29.803258 4860 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-84jp8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-hccts_openstack(d472d27e-9bc9-4d99-8fd9-697f7f44f57d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 11 08:29:29 crc kubenswrapper[4860]: E1211 08:29:29.804509 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-hccts" podUID="d472d27e-9bc9-4d99-8fd9-697f7f44f57d" Dec 11 08:29:29 crc kubenswrapper[4860]: I1211 08:29:29.836449 4860 scope.go:117] "RemoveContainer" containerID="4baa691e265690619f74abc8c059724e292805c92021c7c8964d08fdb9c12671" Dec 11 08:29:30 crc kubenswrapper[4860]: I1211 08:29:30.021471 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6b45499b8f-8xhjk" Dec 11 08:29:30 crc kubenswrapper[4860]: I1211 08:29:30.119853 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 11 08:29:30 crc kubenswrapper[4860]: I1211 08:29:30.121902 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 11 08:29:30 crc kubenswrapper[4860]: I1211 08:29:30.127483 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 11 08:29:30 crc kubenswrapper[4860]: I1211 08:29:30.128104 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 11 08:29:30 crc kubenswrapper[4860]: I1211 08:29:30.129121 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-9qb2r" Dec 11 08:29:30 crc kubenswrapper[4860]: I1211 08:29:30.131559 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5c9ab8b4-ed9e-46a5-b553-3d085ff20722-scripts\") pod \"5c9ab8b4-ed9e-46a5-b553-3d085ff20722\" (UID: \"5c9ab8b4-ed9e-46a5-b553-3d085ff20722\") " Dec 11 08:29:30 crc kubenswrapper[4860]: I1211 08:29:30.131694 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/5c9ab8b4-ed9e-46a5-b553-3d085ff20722-horizon-secret-key\") pod \"5c9ab8b4-ed9e-46a5-b553-3d085ff20722\" (UID: \"5c9ab8b4-ed9e-46a5-b553-3d085ff20722\") " Dec 11 08:29:30 crc kubenswrapper[4860]: I1211 08:29:30.131793 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5c9ab8b4-ed9e-46a5-b553-3d085ff20722-config-data\") pod \"5c9ab8b4-ed9e-46a5-b553-3d085ff20722\" (UID: \"5c9ab8b4-ed9e-46a5-b553-3d085ff20722\") " Dec 11 08:29:30 crc kubenswrapper[4860]: I1211 08:29:30.131904 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5c9ab8b4-ed9e-46a5-b553-3d085ff20722-logs\") pod \"5c9ab8b4-ed9e-46a5-b553-3d085ff20722\" (UID: \"5c9ab8b4-ed9e-46a5-b553-3d085ff20722\") " Dec 11 08:29:30 crc kubenswrapper[4860]: I1211 08:29:30.132159 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fjxs5\" (UniqueName: \"kubernetes.io/projected/5c9ab8b4-ed9e-46a5-b553-3d085ff20722-kube-api-access-fjxs5\") pod \"5c9ab8b4-ed9e-46a5-b553-3d085ff20722\" (UID: \"5c9ab8b4-ed9e-46a5-b553-3d085ff20722\") " Dec 11 08:29:30 crc kubenswrapper[4860]: I1211 08:29:30.132921 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5c9ab8b4-ed9e-46a5-b553-3d085ff20722-scripts" (OuterVolumeSpecName: "scripts") pod "5c9ab8b4-ed9e-46a5-b553-3d085ff20722" (UID: "5c9ab8b4-ed9e-46a5-b553-3d085ff20722"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:29:30 crc kubenswrapper[4860]: I1211 08:29:30.132934 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5c9ab8b4-ed9e-46a5-b553-3d085ff20722-logs" (OuterVolumeSpecName: "logs") pod "5c9ab8b4-ed9e-46a5-b553-3d085ff20722" (UID: "5c9ab8b4-ed9e-46a5-b553-3d085ff20722"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:29:30 crc kubenswrapper[4860]: I1211 08:29:30.134088 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Dec 11 08:29:30 crc kubenswrapper[4860]: I1211 08:29:30.135139 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5c9ab8b4-ed9e-46a5-b553-3d085ff20722-config-data" (OuterVolumeSpecName: "config-data") pod "5c9ab8b4-ed9e-46a5-b553-3d085ff20722" (UID: "5c9ab8b4-ed9e-46a5-b553-3d085ff20722"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:29:30 crc kubenswrapper[4860]: I1211 08:29:30.135814 4860 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5c9ab8b4-ed9e-46a5-b553-3d085ff20722-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 08:29:30 crc kubenswrapper[4860]: I1211 08:29:30.135834 4860 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5c9ab8b4-ed9e-46a5-b553-3d085ff20722-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 08:29:30 crc kubenswrapper[4860]: I1211 08:29:30.135875 4860 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5c9ab8b4-ed9e-46a5-b553-3d085ff20722-logs\") on node \"crc\" DevicePath \"\"" Dec 11 08:29:30 crc kubenswrapper[4860]: I1211 08:29:30.149421 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c9ab8b4-ed9e-46a5-b553-3d085ff20722-kube-api-access-fjxs5" (OuterVolumeSpecName: "kube-api-access-fjxs5") pod "5c9ab8b4-ed9e-46a5-b553-3d085ff20722" (UID: "5c9ab8b4-ed9e-46a5-b553-3d085ff20722"). InnerVolumeSpecName "kube-api-access-fjxs5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:29:30 crc kubenswrapper[4860]: I1211 08:29:30.149891 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c9ab8b4-ed9e-46a5-b553-3d085ff20722-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "5c9ab8b4-ed9e-46a5-b553-3d085ff20722" (UID: "5c9ab8b4-ed9e-46a5-b553-3d085ff20722"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:29:30 crc kubenswrapper[4860]: I1211 08:29:30.237380 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f153e5dc-0977-40b7-9ed6-ad88919f1081-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"f153e5dc-0977-40b7-9ed6-ad88919f1081\") " pod="openstack/glance-default-external-api-0" Dec 11 08:29:30 crc kubenswrapper[4860]: I1211 08:29:30.237479 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f153e5dc-0977-40b7-9ed6-ad88919f1081-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"f153e5dc-0977-40b7-9ed6-ad88919f1081\") " pod="openstack/glance-default-external-api-0" Dec 11 08:29:30 crc kubenswrapper[4860]: I1211 08:29:30.237553 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f153e5dc-0977-40b7-9ed6-ad88919f1081-logs\") pod \"glance-default-external-api-0\" (UID: \"f153e5dc-0977-40b7-9ed6-ad88919f1081\") " pod="openstack/glance-default-external-api-0" Dec 11 08:29:30 crc kubenswrapper[4860]: I1211 08:29:30.237571 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f153e5dc-0977-40b7-9ed6-ad88919f1081-config-data\") pod \"glance-default-external-api-0\" (UID: \"f153e5dc-0977-40b7-9ed6-ad88919f1081\") " pod="openstack/glance-default-external-api-0" Dec 11 08:29:30 crc kubenswrapper[4860]: I1211 08:29:30.237608 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"f153e5dc-0977-40b7-9ed6-ad88919f1081\") " pod="openstack/glance-default-external-api-0" Dec 11 08:29:30 crc kubenswrapper[4860]: I1211 08:29:30.237661 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f153e5dc-0977-40b7-9ed6-ad88919f1081-scripts\") pod \"glance-default-external-api-0\" (UID: \"f153e5dc-0977-40b7-9ed6-ad88919f1081\") " pod="openstack/glance-default-external-api-0" Dec 11 08:29:30 crc kubenswrapper[4860]: I1211 08:29:30.237688 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vwvc6\" (UniqueName: \"kubernetes.io/projected/f153e5dc-0977-40b7-9ed6-ad88919f1081-kube-api-access-vwvc6\") pod \"glance-default-external-api-0\" (UID: \"f153e5dc-0977-40b7-9ed6-ad88919f1081\") " pod="openstack/glance-default-external-api-0" Dec 11 08:29:30 crc kubenswrapper[4860]: I1211 08:29:30.237783 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fjxs5\" (UniqueName: \"kubernetes.io/projected/5c9ab8b4-ed9e-46a5-b553-3d085ff20722-kube-api-access-fjxs5\") on node \"crc\" DevicePath \"\"" Dec 11 08:29:30 crc kubenswrapper[4860]: I1211 08:29:30.237799 4860 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/5c9ab8b4-ed9e-46a5-b553-3d085ff20722-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 11 08:29:30 crc kubenswrapper[4860]: I1211 08:29:30.262878 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-jqgbc" event={"ID":"eafaa4e1-5187-4406-b55a-9d9349e3a744","Type":"ContainerStarted","Data":"5225f72cd5a3da2be39a5c10f3cec79dbb7979ee0b367d21f16de3045373d813"} Dec 11 08:29:30 crc kubenswrapper[4860]: I1211 08:29:30.266392 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-pfz4t" event={"ID":"4ff61ef1-6a67-4a8e-80e2-11f0154e54c6","Type":"ContainerStarted","Data":"2c222aa49386e8cb938f0cf793517fceae739a76f3c585759558399c58a07d51"} Dec 11 08:29:30 crc kubenswrapper[4860]: I1211 08:29:30.273489 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6b45499b8f-8xhjk" Dec 11 08:29:30 crc kubenswrapper[4860]: I1211 08:29:30.273478 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6b45499b8f-8xhjk" event={"ID":"5c9ab8b4-ed9e-46a5-b553-3d085ff20722","Type":"ContainerDied","Data":"6dff7cdae86985e8baa6bad4ec0cd2b4da501d0c3494a18976771c194d5c5217"} Dec 11 08:29:30 crc kubenswrapper[4860]: E1211 08:29:30.275494 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-hccts" podUID="d472d27e-9bc9-4d99-8fd9-697f7f44f57d" Dec 11 08:29:30 crc kubenswrapper[4860]: I1211 08:29:30.285415 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-jqgbc" podStartSLOduration=4.906027589 podStartE2EDuration="32.285397365s" podCreationTimestamp="2025-12-11 08:28:58 +0000 UTC" firstStartedPulling="2025-12-11 08:29:00.233087796 +0000 UTC m=+1072.961606851" lastFinishedPulling="2025-12-11 08:29:27.612457572 +0000 UTC m=+1100.340976627" observedRunningTime="2025-12-11 08:29:30.283403973 +0000 UTC m=+1103.011923028" watchObservedRunningTime="2025-12-11 08:29:30.285397365 +0000 UTC m=+1103.013916420" Dec 11 08:29:30 crc kubenswrapper[4860]: I1211 08:29:30.322554 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-pfz4t" podStartSLOduration=2.650146894 podStartE2EDuration="32.322503131s" podCreationTimestamp="2025-12-11 08:28:58 +0000 UTC" firstStartedPulling="2025-12-11 08:29:00.077751843 +0000 UTC m=+1072.806270898" lastFinishedPulling="2025-12-11 08:29:29.75010808 +0000 UTC m=+1102.478627135" observedRunningTime="2025-12-11 08:29:30.312078484 +0000 UTC m=+1103.040597539" watchObservedRunningTime="2025-12-11 08:29:30.322503131 +0000 UTC m=+1103.051022186" Dec 11 08:29:30 crc kubenswrapper[4860]: I1211 08:29:30.339337 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f153e5dc-0977-40b7-9ed6-ad88919f1081-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"f153e5dc-0977-40b7-9ed6-ad88919f1081\") " pod="openstack/glance-default-external-api-0" Dec 11 08:29:30 crc kubenswrapper[4860]: I1211 08:29:30.339422 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f153e5dc-0977-40b7-9ed6-ad88919f1081-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"f153e5dc-0977-40b7-9ed6-ad88919f1081\") " pod="openstack/glance-default-external-api-0" Dec 11 08:29:30 crc kubenswrapper[4860]: I1211 08:29:30.339487 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f153e5dc-0977-40b7-9ed6-ad88919f1081-logs\") pod \"glance-default-external-api-0\" (UID: \"f153e5dc-0977-40b7-9ed6-ad88919f1081\") " pod="openstack/glance-default-external-api-0" Dec 11 08:29:30 crc kubenswrapper[4860]: I1211 08:29:30.339509 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f153e5dc-0977-40b7-9ed6-ad88919f1081-config-data\") pod \"glance-default-external-api-0\" (UID: \"f153e5dc-0977-40b7-9ed6-ad88919f1081\") " pod="openstack/glance-default-external-api-0" Dec 11 08:29:30 crc kubenswrapper[4860]: I1211 08:29:30.339549 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"f153e5dc-0977-40b7-9ed6-ad88919f1081\") " pod="openstack/glance-default-external-api-0" Dec 11 08:29:30 crc kubenswrapper[4860]: I1211 08:29:30.339585 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f153e5dc-0977-40b7-9ed6-ad88919f1081-scripts\") pod \"glance-default-external-api-0\" (UID: \"f153e5dc-0977-40b7-9ed6-ad88919f1081\") " pod="openstack/glance-default-external-api-0" Dec 11 08:29:30 crc kubenswrapper[4860]: I1211 08:29:30.339610 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vwvc6\" (UniqueName: \"kubernetes.io/projected/f153e5dc-0977-40b7-9ed6-ad88919f1081-kube-api-access-vwvc6\") pod \"glance-default-external-api-0\" (UID: \"f153e5dc-0977-40b7-9ed6-ad88919f1081\") " pod="openstack/glance-default-external-api-0" Dec 11 08:29:30 crc kubenswrapper[4860]: I1211 08:29:30.340431 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f153e5dc-0977-40b7-9ed6-ad88919f1081-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"f153e5dc-0977-40b7-9ed6-ad88919f1081\") " pod="openstack/glance-default-external-api-0" Dec 11 08:29:30 crc kubenswrapper[4860]: I1211 08:29:30.340592 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f153e5dc-0977-40b7-9ed6-ad88919f1081-logs\") pod \"glance-default-external-api-0\" (UID: \"f153e5dc-0977-40b7-9ed6-ad88919f1081\") " pod="openstack/glance-default-external-api-0" Dec 11 08:29:30 crc kubenswrapper[4860]: I1211 08:29:30.340776 4860 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"f153e5dc-0977-40b7-9ed6-ad88919f1081\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-external-api-0" Dec 11 08:29:30 crc kubenswrapper[4860]: I1211 08:29:30.344750 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f153e5dc-0977-40b7-9ed6-ad88919f1081-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"f153e5dc-0977-40b7-9ed6-ad88919f1081\") " pod="openstack/glance-default-external-api-0" Dec 11 08:29:30 crc kubenswrapper[4860]: I1211 08:29:30.345394 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f153e5dc-0977-40b7-9ed6-ad88919f1081-scripts\") pod \"glance-default-external-api-0\" (UID: \"f153e5dc-0977-40b7-9ed6-ad88919f1081\") " pod="openstack/glance-default-external-api-0" Dec 11 08:29:30 crc kubenswrapper[4860]: I1211 08:29:30.346010 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f153e5dc-0977-40b7-9ed6-ad88919f1081-config-data\") pod \"glance-default-external-api-0\" (UID: \"f153e5dc-0977-40b7-9ed6-ad88919f1081\") " pod="openstack/glance-default-external-api-0" Dec 11 08:29:30 crc kubenswrapper[4860]: I1211 08:29:30.369491 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vwvc6\" (UniqueName: \"kubernetes.io/projected/f153e5dc-0977-40b7-9ed6-ad88919f1081-kube-api-access-vwvc6\") pod \"glance-default-external-api-0\" (UID: \"f153e5dc-0977-40b7-9ed6-ad88919f1081\") " pod="openstack/glance-default-external-api-0" Dec 11 08:29:30 crc kubenswrapper[4860]: I1211 08:29:30.443428 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 11 08:29:30 crc kubenswrapper[4860]: I1211 08:29:30.447926 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 11 08:29:30 crc kubenswrapper[4860]: I1211 08:29:30.452219 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 11 08:29:30 crc kubenswrapper[4860]: I1211 08:29:30.469725 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"f153e5dc-0977-40b7-9ed6-ad88919f1081\") " pod="openstack/glance-default-external-api-0" Dec 11 08:29:30 crc kubenswrapper[4860]: I1211 08:29:30.514501 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 11 08:29:30 crc kubenswrapper[4860]: I1211 08:29:30.547086 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-6b45499b8f-8xhjk"] Dec 11 08:29:30 crc kubenswrapper[4860]: I1211 08:29:30.550027 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67ac03d0-1fbe-46ff-b61d-7d15bcf48f61-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"67ac03d0-1fbe-46ff-b61d-7d15bcf48f61\") " pod="openstack/glance-default-internal-api-0" Dec 11 08:29:30 crc kubenswrapper[4860]: I1211 08:29:30.550086 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67ac03d0-1fbe-46ff-b61d-7d15bcf48f61-scripts\") pod \"glance-default-internal-api-0\" (UID: \"67ac03d0-1fbe-46ff-b61d-7d15bcf48f61\") " pod="openstack/glance-default-internal-api-0" Dec 11 08:29:30 crc kubenswrapper[4860]: I1211 08:29:30.550119 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/67ac03d0-1fbe-46ff-b61d-7d15bcf48f61-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"67ac03d0-1fbe-46ff-b61d-7d15bcf48f61\") " pod="openstack/glance-default-internal-api-0" Dec 11 08:29:30 crc kubenswrapper[4860]: I1211 08:29:30.550165 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67ac03d0-1fbe-46ff-b61d-7d15bcf48f61-config-data\") pod \"glance-default-internal-api-0\" (UID: \"67ac03d0-1fbe-46ff-b61d-7d15bcf48f61\") " pod="openstack/glance-default-internal-api-0" Dec 11 08:29:30 crc kubenswrapper[4860]: I1211 08:29:30.550204 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n2nxk\" (UniqueName: \"kubernetes.io/projected/67ac03d0-1fbe-46ff-b61d-7d15bcf48f61-kube-api-access-n2nxk\") pod \"glance-default-internal-api-0\" (UID: \"67ac03d0-1fbe-46ff-b61d-7d15bcf48f61\") " pod="openstack/glance-default-internal-api-0" Dec 11 08:29:30 crc kubenswrapper[4860]: I1211 08:29:30.550241 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/67ac03d0-1fbe-46ff-b61d-7d15bcf48f61-logs\") pod \"glance-default-internal-api-0\" (UID: \"67ac03d0-1fbe-46ff-b61d-7d15bcf48f61\") " pod="openstack/glance-default-internal-api-0" Dec 11 08:29:30 crc kubenswrapper[4860]: I1211 08:29:30.550279 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"67ac03d0-1fbe-46ff-b61d-7d15bcf48f61\") " pod="openstack/glance-default-internal-api-0" Dec 11 08:29:30 crc kubenswrapper[4860]: I1211 08:29:30.569073 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-6b45499b8f-8xhjk"] Dec 11 08:29:30 crc kubenswrapper[4860]: I1211 08:29:30.592718 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-776755b658-9fwr9"] Dec 11 08:29:30 crc kubenswrapper[4860]: I1211 08:29:30.618738 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6b4697fc68-vnd26"] Dec 11 08:29:30 crc kubenswrapper[4860]: I1211 08:29:30.652969 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67ac03d0-1fbe-46ff-b61d-7d15bcf48f61-config-data\") pod \"glance-default-internal-api-0\" (UID: \"67ac03d0-1fbe-46ff-b61d-7d15bcf48f61\") " pod="openstack/glance-default-internal-api-0" Dec 11 08:29:30 crc kubenswrapper[4860]: I1211 08:29:30.653055 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n2nxk\" (UniqueName: \"kubernetes.io/projected/67ac03d0-1fbe-46ff-b61d-7d15bcf48f61-kube-api-access-n2nxk\") pod \"glance-default-internal-api-0\" (UID: \"67ac03d0-1fbe-46ff-b61d-7d15bcf48f61\") " pod="openstack/glance-default-internal-api-0" Dec 11 08:29:30 crc kubenswrapper[4860]: I1211 08:29:30.653118 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/67ac03d0-1fbe-46ff-b61d-7d15bcf48f61-logs\") pod \"glance-default-internal-api-0\" (UID: \"67ac03d0-1fbe-46ff-b61d-7d15bcf48f61\") " pod="openstack/glance-default-internal-api-0" Dec 11 08:29:30 crc kubenswrapper[4860]: I1211 08:29:30.653189 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"67ac03d0-1fbe-46ff-b61d-7d15bcf48f61\") " pod="openstack/glance-default-internal-api-0" Dec 11 08:29:30 crc kubenswrapper[4860]: I1211 08:29:30.653417 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67ac03d0-1fbe-46ff-b61d-7d15bcf48f61-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"67ac03d0-1fbe-46ff-b61d-7d15bcf48f61\") " pod="openstack/glance-default-internal-api-0" Dec 11 08:29:30 crc kubenswrapper[4860]: I1211 08:29:30.653453 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67ac03d0-1fbe-46ff-b61d-7d15bcf48f61-scripts\") pod \"glance-default-internal-api-0\" (UID: \"67ac03d0-1fbe-46ff-b61d-7d15bcf48f61\") " pod="openstack/glance-default-internal-api-0" Dec 11 08:29:30 crc kubenswrapper[4860]: I1211 08:29:30.653482 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/67ac03d0-1fbe-46ff-b61d-7d15bcf48f61-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"67ac03d0-1fbe-46ff-b61d-7d15bcf48f61\") " pod="openstack/glance-default-internal-api-0" Dec 11 08:29:30 crc kubenswrapper[4860]: I1211 08:29:30.654037 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/67ac03d0-1fbe-46ff-b61d-7d15bcf48f61-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"67ac03d0-1fbe-46ff-b61d-7d15bcf48f61\") " pod="openstack/glance-default-internal-api-0" Dec 11 08:29:30 crc kubenswrapper[4860]: I1211 08:29:30.656059 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/67ac03d0-1fbe-46ff-b61d-7d15bcf48f61-logs\") pod \"glance-default-internal-api-0\" (UID: \"67ac03d0-1fbe-46ff-b61d-7d15bcf48f61\") " pod="openstack/glance-default-internal-api-0" Dec 11 08:29:30 crc kubenswrapper[4860]: I1211 08:29:30.656414 4860 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"67ac03d0-1fbe-46ff-b61d-7d15bcf48f61\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/glance-default-internal-api-0" Dec 11 08:29:30 crc kubenswrapper[4860]: I1211 08:29:30.659968 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67ac03d0-1fbe-46ff-b61d-7d15bcf48f61-scripts\") pod \"glance-default-internal-api-0\" (UID: \"67ac03d0-1fbe-46ff-b61d-7d15bcf48f61\") " pod="openstack/glance-default-internal-api-0" Dec 11 08:29:30 crc kubenswrapper[4860]: I1211 08:29:30.662964 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67ac03d0-1fbe-46ff-b61d-7d15bcf48f61-config-data\") pod \"glance-default-internal-api-0\" (UID: \"67ac03d0-1fbe-46ff-b61d-7d15bcf48f61\") " pod="openstack/glance-default-internal-api-0" Dec 11 08:29:30 crc kubenswrapper[4860]: I1211 08:29:30.663995 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67ac03d0-1fbe-46ff-b61d-7d15bcf48f61-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"67ac03d0-1fbe-46ff-b61d-7d15bcf48f61\") " pod="openstack/glance-default-internal-api-0" Dec 11 08:29:30 crc kubenswrapper[4860]: I1211 08:29:30.680173 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57c957c4ff-wxx2m"] Dec 11 08:29:30 crc kubenswrapper[4860]: I1211 08:29:30.693062 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n2nxk\" (UniqueName: \"kubernetes.io/projected/67ac03d0-1fbe-46ff-b61d-7d15bcf48f61-kube-api-access-n2nxk\") pod \"glance-default-internal-api-0\" (UID: \"67ac03d0-1fbe-46ff-b61d-7d15bcf48f61\") " pod="openstack/glance-default-internal-api-0" Dec 11 08:29:30 crc kubenswrapper[4860]: I1211 08:29:30.695584 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"67ac03d0-1fbe-46ff-b61d-7d15bcf48f61\") " pod="openstack/glance-default-internal-api-0" Dec 11 08:29:30 crc kubenswrapper[4860]: I1211 08:29:30.695726 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-plxxc"] Dec 11 08:29:30 crc kubenswrapper[4860]: I1211 08:29:30.759222 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 11 08:29:30 crc kubenswrapper[4860]: I1211 08:29:30.822746 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-wgk5t" Dec 11 08:29:30 crc kubenswrapper[4860]: I1211 08:29:30.851791 4860 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6d5b6d6b67-pxxrf" podUID="d59cd3a6-bd2d-4c74-97d0-265b3fc56dfd" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.123:5353: i/o timeout" Dec 11 08:29:30 crc kubenswrapper[4860]: I1211 08:29:30.869042 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e3822ab5-d1a7-46da-9533-21b67bd240eb-config\") pod \"e3822ab5-d1a7-46da-9533-21b67bd240eb\" (UID: \"e3822ab5-d1a7-46da-9533-21b67bd240eb\") " Dec 11 08:29:30 crc kubenswrapper[4860]: I1211 08:29:30.869091 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rdsd6\" (UniqueName: \"kubernetes.io/projected/e3822ab5-d1a7-46da-9533-21b67bd240eb-kube-api-access-rdsd6\") pod \"e3822ab5-d1a7-46da-9533-21b67bd240eb\" (UID: \"e3822ab5-d1a7-46da-9533-21b67bd240eb\") " Dec 11 08:29:30 crc kubenswrapper[4860]: I1211 08:29:30.869241 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3822ab5-d1a7-46da-9533-21b67bd240eb-combined-ca-bundle\") pod \"e3822ab5-d1a7-46da-9533-21b67bd240eb\" (UID: \"e3822ab5-d1a7-46da-9533-21b67bd240eb\") " Dec 11 08:29:30 crc kubenswrapper[4860]: I1211 08:29:30.872566 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3822ab5-d1a7-46da-9533-21b67bd240eb-kube-api-access-rdsd6" (OuterVolumeSpecName: "kube-api-access-rdsd6") pod "e3822ab5-d1a7-46da-9533-21b67bd240eb" (UID: "e3822ab5-d1a7-46da-9533-21b67bd240eb"). InnerVolumeSpecName "kube-api-access-rdsd6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:29:30 crc kubenswrapper[4860]: I1211 08:29:30.882182 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 11 08:29:30 crc kubenswrapper[4860]: I1211 08:29:30.893201 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3822ab5-d1a7-46da-9533-21b67bd240eb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e3822ab5-d1a7-46da-9533-21b67bd240eb" (UID: "e3822ab5-d1a7-46da-9533-21b67bd240eb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:29:30 crc kubenswrapper[4860]: I1211 08:29:30.918807 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3822ab5-d1a7-46da-9533-21b67bd240eb-config" (OuterVolumeSpecName: "config") pod "e3822ab5-d1a7-46da-9533-21b67bd240eb" (UID: "e3822ab5-d1a7-46da-9533-21b67bd240eb"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:29:30 crc kubenswrapper[4860]: I1211 08:29:30.971193 4860 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/e3822ab5-d1a7-46da-9533-21b67bd240eb-config\") on node \"crc\" DevicePath \"\"" Dec 11 08:29:30 crc kubenswrapper[4860]: I1211 08:29:30.971577 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rdsd6\" (UniqueName: \"kubernetes.io/projected/e3822ab5-d1a7-46da-9533-21b67bd240eb-kube-api-access-rdsd6\") on node \"crc\" DevicePath \"\"" Dec 11 08:29:30 crc kubenswrapper[4860]: I1211 08:29:30.971596 4860 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3822ab5-d1a7-46da-9533-21b67bd240eb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 08:29:31 crc kubenswrapper[4860]: I1211 08:29:31.299272 4860 generic.go:334] "Generic (PLEG): container finished" podID="79626ad3-4978-43f0-b163-fe482aa1dac7" containerID="dc6c2ac31f2af352f724e5c89081aec87b45b1b7a692bb7fce71b55091cfad48" exitCode=0 Dec 11 08:29:31 crc kubenswrapper[4860]: I1211 08:29:31.299586 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57c957c4ff-wxx2m" event={"ID":"79626ad3-4978-43f0-b163-fe482aa1dac7","Type":"ContainerDied","Data":"dc6c2ac31f2af352f724e5c89081aec87b45b1b7a692bb7fce71b55091cfad48"} Dec 11 08:29:31 crc kubenswrapper[4860]: I1211 08:29:31.299616 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57c957c4ff-wxx2m" event={"ID":"79626ad3-4978-43f0-b163-fe482aa1dac7","Type":"ContainerStarted","Data":"6fcd3aafd910f5701c1a5b87eae989d7c034b47b703be6ececcc925a8cb5f06c"} Dec 11 08:29:31 crc kubenswrapper[4860]: I1211 08:29:31.316542 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-plxxc" event={"ID":"fecaee75-2dc0-4d48-9d40-11fc1ec8e972","Type":"ContainerStarted","Data":"b5e16e2ff7b9a93d183828e46286db5391626046b43b832b915a44d625d1f621"} Dec 11 08:29:31 crc kubenswrapper[4860]: I1211 08:29:31.316607 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-plxxc" event={"ID":"fecaee75-2dc0-4d48-9d40-11fc1ec8e972","Type":"ContainerStarted","Data":"135ecdf1d9cab450a5e7bbb463cf46c95697996cf9020c6c3f85dd181f8d70c8"} Dec 11 08:29:31 crc kubenswrapper[4860]: I1211 08:29:31.332664 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-wgk5t" event={"ID":"e3822ab5-d1a7-46da-9533-21b67bd240eb","Type":"ContainerDied","Data":"3c8a25c0e0a663af7cce9b56a57ec9372e8c0a04ad5fc9e5c618b575b544d25a"} Dec 11 08:29:31 crc kubenswrapper[4860]: I1211 08:29:31.332706 4860 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3c8a25c0e0a663af7cce9b56a57ec9372e8c0a04ad5fc9e5c618b575b544d25a" Dec 11 08:29:31 crc kubenswrapper[4860]: I1211 08:29:31.332778 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-wgk5t" Dec 11 08:29:31 crc kubenswrapper[4860]: I1211 08:29:31.352659 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-776755b658-9fwr9" event={"ID":"8f38e033-5243-4e65-a7e0-29f1282b063f","Type":"ContainerStarted","Data":"e978a4eb3ce30930d0707ffadd7da47a1d76fbcdb63711451ee10b5c7580b359"} Dec 11 08:29:31 crc kubenswrapper[4860]: I1211 08:29:31.362523 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6b4697fc68-vnd26" event={"ID":"78b0f75e-d447-4541-be11-9afe4549d2e1","Type":"ContainerStarted","Data":"373a93d2a393c5cf67b5baa2b8e37d55a5e8200441c2ea72feeb6ff6fe6944e2"} Dec 11 08:29:31 crc kubenswrapper[4860]: I1211 08:29:31.373959 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-plxxc" podStartSLOduration=13.373941069 podStartE2EDuration="13.373941069s" podCreationTimestamp="2025-12-11 08:29:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:29:31.362493999 +0000 UTC m=+1104.091013054" watchObservedRunningTime="2025-12-11 08:29:31.373941069 +0000 UTC m=+1104.102460124" Dec 11 08:29:31 crc kubenswrapper[4860]: I1211 08:29:31.574093 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57c957c4ff-wxx2m"] Dec 11 08:29:31 crc kubenswrapper[4860]: I1211 08:29:31.611512 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5c9ab8b4-ed9e-46a5-b553-3d085ff20722" path="/var/lib/kubelet/pods/5c9ab8b4-ed9e-46a5-b553-3d085ff20722/volumes" Dec 11 08:29:31 crc kubenswrapper[4860]: I1211 08:29:31.646562 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 11 08:29:31 crc kubenswrapper[4860]: I1211 08:29:31.677517 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5ccc5c4795-65k7g"] Dec 11 08:29:31 crc kubenswrapper[4860]: E1211 08:29:31.678074 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3822ab5-d1a7-46da-9533-21b67bd240eb" containerName="neutron-db-sync" Dec 11 08:29:31 crc kubenswrapper[4860]: I1211 08:29:31.678097 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3822ab5-d1a7-46da-9533-21b67bd240eb" containerName="neutron-db-sync" Dec 11 08:29:31 crc kubenswrapper[4860]: I1211 08:29:31.678325 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3822ab5-d1a7-46da-9533-21b67bd240eb" containerName="neutron-db-sync" Dec 11 08:29:31 crc kubenswrapper[4860]: I1211 08:29:31.679511 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc5c4795-65k7g" Dec 11 08:29:31 crc kubenswrapper[4860]: I1211 08:29:31.711382 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5ccc5c4795-65k7g"] Dec 11 08:29:31 crc kubenswrapper[4860]: I1211 08:29:31.749234 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-ccf7c6598-jcbgb"] Dec 11 08:29:31 crc kubenswrapper[4860]: I1211 08:29:31.750756 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-ccf7c6598-jcbgb" Dec 11 08:29:31 crc kubenswrapper[4860]: I1211 08:29:31.759632 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-dxbrv" Dec 11 08:29:31 crc kubenswrapper[4860]: I1211 08:29:31.759916 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 11 08:29:31 crc kubenswrapper[4860]: I1211 08:29:31.767394 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Dec 11 08:29:31 crc kubenswrapper[4860]: I1211 08:29:31.767591 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 11 08:29:31 crc kubenswrapper[4860]: I1211 08:29:31.803774 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fe0e81d5-8772-4047-9b03-fd8127be2b66-ovsdbserver-sb\") pod \"dnsmasq-dns-5ccc5c4795-65k7g\" (UID: \"fe0e81d5-8772-4047-9b03-fd8127be2b66\") " pod="openstack/dnsmasq-dns-5ccc5c4795-65k7g" Dec 11 08:29:31 crc kubenswrapper[4860]: I1211 08:29:31.803824 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hg6tn\" (UniqueName: \"kubernetes.io/projected/fe0e81d5-8772-4047-9b03-fd8127be2b66-kube-api-access-hg6tn\") pod \"dnsmasq-dns-5ccc5c4795-65k7g\" (UID: \"fe0e81d5-8772-4047-9b03-fd8127be2b66\") " pod="openstack/dnsmasq-dns-5ccc5c4795-65k7g" Dec 11 08:29:31 crc kubenswrapper[4860]: I1211 08:29:31.803843 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fe0e81d5-8772-4047-9b03-fd8127be2b66-dns-swift-storage-0\") pod \"dnsmasq-dns-5ccc5c4795-65k7g\" (UID: \"fe0e81d5-8772-4047-9b03-fd8127be2b66\") " pod="openstack/dnsmasq-dns-5ccc5c4795-65k7g" Dec 11 08:29:31 crc kubenswrapper[4860]: I1211 08:29:31.803863 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e076681d-ddec-44a2-9ce4-39b87e5542c6-httpd-config\") pod \"neutron-ccf7c6598-jcbgb\" (UID: \"e076681d-ddec-44a2-9ce4-39b87e5542c6\") " pod="openstack/neutron-ccf7c6598-jcbgb" Dec 11 08:29:31 crc kubenswrapper[4860]: I1211 08:29:31.803904 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e076681d-ddec-44a2-9ce4-39b87e5542c6-config\") pod \"neutron-ccf7c6598-jcbgb\" (UID: \"e076681d-ddec-44a2-9ce4-39b87e5542c6\") " pod="openstack/neutron-ccf7c6598-jcbgb" Dec 11 08:29:31 crc kubenswrapper[4860]: I1211 08:29:31.803927 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fe0e81d5-8772-4047-9b03-fd8127be2b66-dns-svc\") pod \"dnsmasq-dns-5ccc5c4795-65k7g\" (UID: \"fe0e81d5-8772-4047-9b03-fd8127be2b66\") " pod="openstack/dnsmasq-dns-5ccc5c4795-65k7g" Dec 11 08:29:31 crc kubenswrapper[4860]: I1211 08:29:31.803945 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e076681d-ddec-44a2-9ce4-39b87e5542c6-ovndb-tls-certs\") pod \"neutron-ccf7c6598-jcbgb\" (UID: \"e076681d-ddec-44a2-9ce4-39b87e5542c6\") " pod="openstack/neutron-ccf7c6598-jcbgb" Dec 11 08:29:31 crc kubenswrapper[4860]: I1211 08:29:31.803962 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fe0e81d5-8772-4047-9b03-fd8127be2b66-config\") pod \"dnsmasq-dns-5ccc5c4795-65k7g\" (UID: \"fe0e81d5-8772-4047-9b03-fd8127be2b66\") " pod="openstack/dnsmasq-dns-5ccc5c4795-65k7g" Dec 11 08:29:31 crc kubenswrapper[4860]: I1211 08:29:31.803990 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c7c9x\" (UniqueName: \"kubernetes.io/projected/e076681d-ddec-44a2-9ce4-39b87e5542c6-kube-api-access-c7c9x\") pod \"neutron-ccf7c6598-jcbgb\" (UID: \"e076681d-ddec-44a2-9ce4-39b87e5542c6\") " pod="openstack/neutron-ccf7c6598-jcbgb" Dec 11 08:29:31 crc kubenswrapper[4860]: I1211 08:29:31.804042 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e076681d-ddec-44a2-9ce4-39b87e5542c6-combined-ca-bundle\") pod \"neutron-ccf7c6598-jcbgb\" (UID: \"e076681d-ddec-44a2-9ce4-39b87e5542c6\") " pod="openstack/neutron-ccf7c6598-jcbgb" Dec 11 08:29:31 crc kubenswrapper[4860]: I1211 08:29:31.804062 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fe0e81d5-8772-4047-9b03-fd8127be2b66-ovsdbserver-nb\") pod \"dnsmasq-dns-5ccc5c4795-65k7g\" (UID: \"fe0e81d5-8772-4047-9b03-fd8127be2b66\") " pod="openstack/dnsmasq-dns-5ccc5c4795-65k7g" Dec 11 08:29:31 crc kubenswrapper[4860]: I1211 08:29:31.824750 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-ccf7c6598-jcbgb"] Dec 11 08:29:31 crc kubenswrapper[4860]: I1211 08:29:31.886272 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 11 08:29:31 crc kubenswrapper[4860]: I1211 08:29:31.905463 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fe0e81d5-8772-4047-9b03-fd8127be2b66-dns-svc\") pod \"dnsmasq-dns-5ccc5c4795-65k7g\" (UID: \"fe0e81d5-8772-4047-9b03-fd8127be2b66\") " pod="openstack/dnsmasq-dns-5ccc5c4795-65k7g" Dec 11 08:29:31 crc kubenswrapper[4860]: I1211 08:29:31.905511 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e076681d-ddec-44a2-9ce4-39b87e5542c6-ovndb-tls-certs\") pod \"neutron-ccf7c6598-jcbgb\" (UID: \"e076681d-ddec-44a2-9ce4-39b87e5542c6\") " pod="openstack/neutron-ccf7c6598-jcbgb" Dec 11 08:29:31 crc kubenswrapper[4860]: I1211 08:29:31.905530 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fe0e81d5-8772-4047-9b03-fd8127be2b66-config\") pod \"dnsmasq-dns-5ccc5c4795-65k7g\" (UID: \"fe0e81d5-8772-4047-9b03-fd8127be2b66\") " pod="openstack/dnsmasq-dns-5ccc5c4795-65k7g" Dec 11 08:29:31 crc kubenswrapper[4860]: I1211 08:29:31.905572 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c7c9x\" (UniqueName: \"kubernetes.io/projected/e076681d-ddec-44a2-9ce4-39b87e5542c6-kube-api-access-c7c9x\") pod \"neutron-ccf7c6598-jcbgb\" (UID: \"e076681d-ddec-44a2-9ce4-39b87e5542c6\") " pod="openstack/neutron-ccf7c6598-jcbgb" Dec 11 08:29:31 crc kubenswrapper[4860]: I1211 08:29:31.905635 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e076681d-ddec-44a2-9ce4-39b87e5542c6-combined-ca-bundle\") pod \"neutron-ccf7c6598-jcbgb\" (UID: \"e076681d-ddec-44a2-9ce4-39b87e5542c6\") " pod="openstack/neutron-ccf7c6598-jcbgb" Dec 11 08:29:31 crc kubenswrapper[4860]: I1211 08:29:31.905668 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fe0e81d5-8772-4047-9b03-fd8127be2b66-ovsdbserver-nb\") pod \"dnsmasq-dns-5ccc5c4795-65k7g\" (UID: \"fe0e81d5-8772-4047-9b03-fd8127be2b66\") " pod="openstack/dnsmasq-dns-5ccc5c4795-65k7g" Dec 11 08:29:31 crc kubenswrapper[4860]: I1211 08:29:31.905738 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fe0e81d5-8772-4047-9b03-fd8127be2b66-ovsdbserver-sb\") pod \"dnsmasq-dns-5ccc5c4795-65k7g\" (UID: \"fe0e81d5-8772-4047-9b03-fd8127be2b66\") " pod="openstack/dnsmasq-dns-5ccc5c4795-65k7g" Dec 11 08:29:31 crc kubenswrapper[4860]: I1211 08:29:31.905756 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hg6tn\" (UniqueName: \"kubernetes.io/projected/fe0e81d5-8772-4047-9b03-fd8127be2b66-kube-api-access-hg6tn\") pod \"dnsmasq-dns-5ccc5c4795-65k7g\" (UID: \"fe0e81d5-8772-4047-9b03-fd8127be2b66\") " pod="openstack/dnsmasq-dns-5ccc5c4795-65k7g" Dec 11 08:29:31 crc kubenswrapper[4860]: I1211 08:29:31.905774 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fe0e81d5-8772-4047-9b03-fd8127be2b66-dns-swift-storage-0\") pod \"dnsmasq-dns-5ccc5c4795-65k7g\" (UID: \"fe0e81d5-8772-4047-9b03-fd8127be2b66\") " pod="openstack/dnsmasq-dns-5ccc5c4795-65k7g" Dec 11 08:29:31 crc kubenswrapper[4860]: I1211 08:29:31.905793 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e076681d-ddec-44a2-9ce4-39b87e5542c6-httpd-config\") pod \"neutron-ccf7c6598-jcbgb\" (UID: \"e076681d-ddec-44a2-9ce4-39b87e5542c6\") " pod="openstack/neutron-ccf7c6598-jcbgb" Dec 11 08:29:31 crc kubenswrapper[4860]: I1211 08:29:31.905827 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e076681d-ddec-44a2-9ce4-39b87e5542c6-config\") pod \"neutron-ccf7c6598-jcbgb\" (UID: \"e076681d-ddec-44a2-9ce4-39b87e5542c6\") " pod="openstack/neutron-ccf7c6598-jcbgb" Dec 11 08:29:31 crc kubenswrapper[4860]: I1211 08:29:31.912830 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/e076681d-ddec-44a2-9ce4-39b87e5542c6-config\") pod \"neutron-ccf7c6598-jcbgb\" (UID: \"e076681d-ddec-44a2-9ce4-39b87e5542c6\") " pod="openstack/neutron-ccf7c6598-jcbgb" Dec 11 08:29:31 crc kubenswrapper[4860]: I1211 08:29:31.913748 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fe0e81d5-8772-4047-9b03-fd8127be2b66-ovsdbserver-nb\") pod \"dnsmasq-dns-5ccc5c4795-65k7g\" (UID: \"fe0e81d5-8772-4047-9b03-fd8127be2b66\") " pod="openstack/dnsmasq-dns-5ccc5c4795-65k7g" Dec 11 08:29:31 crc kubenswrapper[4860]: I1211 08:29:31.914465 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fe0e81d5-8772-4047-9b03-fd8127be2b66-ovsdbserver-sb\") pod \"dnsmasq-dns-5ccc5c4795-65k7g\" (UID: \"fe0e81d5-8772-4047-9b03-fd8127be2b66\") " pod="openstack/dnsmasq-dns-5ccc5c4795-65k7g" Dec 11 08:29:31 crc kubenswrapper[4860]: I1211 08:29:31.914996 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fe0e81d5-8772-4047-9b03-fd8127be2b66-config\") pod \"dnsmasq-dns-5ccc5c4795-65k7g\" (UID: \"fe0e81d5-8772-4047-9b03-fd8127be2b66\") " pod="openstack/dnsmasq-dns-5ccc5c4795-65k7g" Dec 11 08:29:31 crc kubenswrapper[4860]: I1211 08:29:31.915570 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fe0e81d5-8772-4047-9b03-fd8127be2b66-dns-swift-storage-0\") pod \"dnsmasq-dns-5ccc5c4795-65k7g\" (UID: \"fe0e81d5-8772-4047-9b03-fd8127be2b66\") " pod="openstack/dnsmasq-dns-5ccc5c4795-65k7g" Dec 11 08:29:31 crc kubenswrapper[4860]: I1211 08:29:31.916540 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fe0e81d5-8772-4047-9b03-fd8127be2b66-dns-svc\") pod \"dnsmasq-dns-5ccc5c4795-65k7g\" (UID: \"fe0e81d5-8772-4047-9b03-fd8127be2b66\") " pod="openstack/dnsmasq-dns-5ccc5c4795-65k7g" Dec 11 08:29:31 crc kubenswrapper[4860]: I1211 08:29:31.923438 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e076681d-ddec-44a2-9ce4-39b87e5542c6-ovndb-tls-certs\") pod \"neutron-ccf7c6598-jcbgb\" (UID: \"e076681d-ddec-44a2-9ce4-39b87e5542c6\") " pod="openstack/neutron-ccf7c6598-jcbgb" Dec 11 08:29:31 crc kubenswrapper[4860]: I1211 08:29:31.927999 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e076681d-ddec-44a2-9ce4-39b87e5542c6-httpd-config\") pod \"neutron-ccf7c6598-jcbgb\" (UID: \"e076681d-ddec-44a2-9ce4-39b87e5542c6\") " pod="openstack/neutron-ccf7c6598-jcbgb" Dec 11 08:29:31 crc kubenswrapper[4860]: I1211 08:29:31.937678 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e076681d-ddec-44a2-9ce4-39b87e5542c6-combined-ca-bundle\") pod \"neutron-ccf7c6598-jcbgb\" (UID: \"e076681d-ddec-44a2-9ce4-39b87e5542c6\") " pod="openstack/neutron-ccf7c6598-jcbgb" Dec 11 08:29:31 crc kubenswrapper[4860]: I1211 08:29:31.944988 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c7c9x\" (UniqueName: \"kubernetes.io/projected/e076681d-ddec-44a2-9ce4-39b87e5542c6-kube-api-access-c7c9x\") pod \"neutron-ccf7c6598-jcbgb\" (UID: \"e076681d-ddec-44a2-9ce4-39b87e5542c6\") " pod="openstack/neutron-ccf7c6598-jcbgb" Dec 11 08:29:31 crc kubenswrapper[4860]: I1211 08:29:31.945265 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hg6tn\" (UniqueName: \"kubernetes.io/projected/fe0e81d5-8772-4047-9b03-fd8127be2b66-kube-api-access-hg6tn\") pod \"dnsmasq-dns-5ccc5c4795-65k7g\" (UID: \"fe0e81d5-8772-4047-9b03-fd8127be2b66\") " pod="openstack/dnsmasq-dns-5ccc5c4795-65k7g" Dec 11 08:29:32 crc kubenswrapper[4860]: I1211 08:29:32.022185 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc5c4795-65k7g" Dec 11 08:29:32 crc kubenswrapper[4860]: I1211 08:29:32.098765 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-ccf7c6598-jcbgb" Dec 11 08:29:32 crc kubenswrapper[4860]: I1211 08:29:32.419905 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"67ac03d0-1fbe-46ff-b61d-7d15bcf48f61","Type":"ContainerStarted","Data":"cf7f8eb5286c424bc2e307f364204e4fd8b87eebfdefe6d3d38b792b1a669680"} Dec 11 08:29:32 crc kubenswrapper[4860]: I1211 08:29:32.452451 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f153e5dc-0977-40b7-9ed6-ad88919f1081","Type":"ContainerStarted","Data":"26481a7b9014dd8624ee0495fc60e935a037101148bb0ab94008cb2f17236c22"} Dec 11 08:29:32 crc kubenswrapper[4860]: I1211 08:29:32.726252 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 11 08:29:32 crc kubenswrapper[4860]: I1211 08:29:32.854789 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 11 08:29:32 crc kubenswrapper[4860]: I1211 08:29:32.933439 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5ccc5c4795-65k7g"] Dec 11 08:29:32 crc kubenswrapper[4860]: W1211 08:29:32.977984 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfe0e81d5_8772_4047_9b03_fd8127be2b66.slice/crio-b29d22420160c71e4e32523ec98bf724a176d783a853ef0b4d9dd854f65c61da WatchSource:0}: Error finding container b29d22420160c71e4e32523ec98bf724a176d783a853ef0b4d9dd854f65c61da: Status 404 returned error can't find the container with id b29d22420160c71e4e32523ec98bf724a176d783a853ef0b4d9dd854f65c61da Dec 11 08:29:33 crc kubenswrapper[4860]: I1211 08:29:33.071484 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-ccf7c6598-jcbgb"] Dec 11 08:29:33 crc kubenswrapper[4860]: I1211 08:29:33.511197 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-776755b658-9fwr9" event={"ID":"8f38e033-5243-4e65-a7e0-29f1282b063f","Type":"ContainerStarted","Data":"0aab949e04f4b3a4703fdc0bff0fb82f6a451c3c24e3537820a25282e91ff2c8"} Dec 11 08:29:33 crc kubenswrapper[4860]: I1211 08:29:33.518024 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6b4697fc68-vnd26" event={"ID":"78b0f75e-d447-4541-be11-9afe4549d2e1","Type":"ContainerStarted","Data":"bbedd88f1dfff1226aaf3f48bbf35d80abc292960604a43d67ebad9bb8615c40"} Dec 11 08:29:33 crc kubenswrapper[4860]: I1211 08:29:33.518081 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6b4697fc68-vnd26" event={"ID":"78b0f75e-d447-4541-be11-9afe4549d2e1","Type":"ContainerStarted","Data":"233afb77e54535211da784324c8064e9c2468182f2cbb9f864e447e11335e749"} Dec 11 08:29:33 crc kubenswrapper[4860]: I1211 08:29:33.528774 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57c957c4ff-wxx2m" event={"ID":"79626ad3-4978-43f0-b163-fe482aa1dac7","Type":"ContainerStarted","Data":"88aedbd26922eb622cb2c1a6ca543159d73a1595095e7fc2731e96c2fe0838cd"} Dec 11 08:29:33 crc kubenswrapper[4860]: I1211 08:29:33.528944 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57c957c4ff-wxx2m" podUID="79626ad3-4978-43f0-b163-fe482aa1dac7" containerName="dnsmasq-dns" containerID="cri-o://88aedbd26922eb622cb2c1a6ca543159d73a1595095e7fc2731e96c2fe0838cd" gracePeriod=10 Dec 11 08:29:33 crc kubenswrapper[4860]: I1211 08:29:33.529049 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57c957c4ff-wxx2m" Dec 11 08:29:33 crc kubenswrapper[4860]: I1211 08:29:33.540948 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f153e5dc-0977-40b7-9ed6-ad88919f1081","Type":"ContainerStarted","Data":"fbe50653cd9fe78c494f42fa61f7c07b554d82034fe17558bd1e3ed88df56056"} Dec 11 08:29:33 crc kubenswrapper[4860]: I1211 08:29:33.549188 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-6b4697fc68-vnd26" podStartSLOduration=25.777630187 podStartE2EDuration="27.549166908s" podCreationTimestamp="2025-12-11 08:29:06 +0000 UTC" firstStartedPulling="2025-12-11 08:29:30.484824903 +0000 UTC m=+1103.213343958" lastFinishedPulling="2025-12-11 08:29:32.256361624 +0000 UTC m=+1104.984880679" observedRunningTime="2025-12-11 08:29:33.546827445 +0000 UTC m=+1106.275346500" watchObservedRunningTime="2025-12-11 08:29:33.549166908 +0000 UTC m=+1106.277685963" Dec 11 08:29:33 crc kubenswrapper[4860]: I1211 08:29:33.572061 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"67ac03d0-1fbe-46ff-b61d-7d15bcf48f61","Type":"ContainerStarted","Data":"5d29695ad5aece20f85dcf129d4f9a829958abfeec4b99546f362ca3993bb5d0"} Dec 11 08:29:33 crc kubenswrapper[4860]: I1211 08:29:33.600883 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-ccf7c6598-jcbgb" event={"ID":"e076681d-ddec-44a2-9ce4-39b87e5542c6","Type":"ContainerStarted","Data":"f409338159e92c6ed18765cf7e90fd5ab1cabe9c8f2ab78e9b25db7c4312cd65"} Dec 11 08:29:33 crc kubenswrapper[4860]: I1211 08:29:33.601963 4860 generic.go:334] "Generic (PLEG): container finished" podID="fe0e81d5-8772-4047-9b03-fd8127be2b66" containerID="e600d5682ae00beaea5e17f2097e7aaf07fbca21b4d3422f88dd24fd87141681" exitCode=0 Dec 11 08:29:33 crc kubenswrapper[4860]: I1211 08:29:33.602049 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc5c4795-65k7g" event={"ID":"fe0e81d5-8772-4047-9b03-fd8127be2b66","Type":"ContainerDied","Data":"e600d5682ae00beaea5e17f2097e7aaf07fbca21b4d3422f88dd24fd87141681"} Dec 11 08:29:33 crc kubenswrapper[4860]: I1211 08:29:33.602094 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc5c4795-65k7g" event={"ID":"fe0e81d5-8772-4047-9b03-fd8127be2b66","Type":"ContainerStarted","Data":"b29d22420160c71e4e32523ec98bf724a176d783a853ef0b4d9dd854f65c61da"} Dec 11 08:29:33 crc kubenswrapper[4860]: I1211 08:29:33.606360 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57c957c4ff-wxx2m" podStartSLOduration=4.606335945 podStartE2EDuration="4.606335945s" podCreationTimestamp="2025-12-11 08:29:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:29:33.594275506 +0000 UTC m=+1106.322794561" watchObservedRunningTime="2025-12-11 08:29:33.606335945 +0000 UTC m=+1106.334855000" Dec 11 08:29:33 crc kubenswrapper[4860]: I1211 08:29:33.611813 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"040a0e8c-efd3-4e3b-be51-59e311ae4406","Type":"ContainerStarted","Data":"2c85cf7db795596f54f14c6ed32043967f06c05f5cab51172fbd00666aa45279"} Dec 11 08:29:34 crc kubenswrapper[4860]: I1211 08:29:34.217918 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57c957c4ff-wxx2m" Dec 11 08:29:34 crc kubenswrapper[4860]: I1211 08:29:34.271272 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/79626ad3-4978-43f0-b163-fe482aa1dac7-dns-svc\") pod \"79626ad3-4978-43f0-b163-fe482aa1dac7\" (UID: \"79626ad3-4978-43f0-b163-fe482aa1dac7\") " Dec 11 08:29:34 crc kubenswrapper[4860]: I1211 08:29:34.271402 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/79626ad3-4978-43f0-b163-fe482aa1dac7-dns-swift-storage-0\") pod \"79626ad3-4978-43f0-b163-fe482aa1dac7\" (UID: \"79626ad3-4978-43f0-b163-fe482aa1dac7\") " Dec 11 08:29:34 crc kubenswrapper[4860]: I1211 08:29:34.271457 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/79626ad3-4978-43f0-b163-fe482aa1dac7-ovsdbserver-sb\") pod \"79626ad3-4978-43f0-b163-fe482aa1dac7\" (UID: \"79626ad3-4978-43f0-b163-fe482aa1dac7\") " Dec 11 08:29:34 crc kubenswrapper[4860]: I1211 08:29:34.273920 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q5qmt\" (UniqueName: \"kubernetes.io/projected/79626ad3-4978-43f0-b163-fe482aa1dac7-kube-api-access-q5qmt\") pod \"79626ad3-4978-43f0-b163-fe482aa1dac7\" (UID: \"79626ad3-4978-43f0-b163-fe482aa1dac7\") " Dec 11 08:29:34 crc kubenswrapper[4860]: I1211 08:29:34.274024 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/79626ad3-4978-43f0-b163-fe482aa1dac7-config\") pod \"79626ad3-4978-43f0-b163-fe482aa1dac7\" (UID: \"79626ad3-4978-43f0-b163-fe482aa1dac7\") " Dec 11 08:29:34 crc kubenswrapper[4860]: I1211 08:29:34.274061 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/79626ad3-4978-43f0-b163-fe482aa1dac7-ovsdbserver-nb\") pod \"79626ad3-4978-43f0-b163-fe482aa1dac7\" (UID: \"79626ad3-4978-43f0-b163-fe482aa1dac7\") " Dec 11 08:29:34 crc kubenswrapper[4860]: I1211 08:29:34.316979 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/79626ad3-4978-43f0-b163-fe482aa1dac7-kube-api-access-q5qmt" (OuterVolumeSpecName: "kube-api-access-q5qmt") pod "79626ad3-4978-43f0-b163-fe482aa1dac7" (UID: "79626ad3-4978-43f0-b163-fe482aa1dac7"). InnerVolumeSpecName "kube-api-access-q5qmt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:29:34 crc kubenswrapper[4860]: I1211 08:29:34.371319 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/79626ad3-4978-43f0-b163-fe482aa1dac7-config" (OuterVolumeSpecName: "config") pod "79626ad3-4978-43f0-b163-fe482aa1dac7" (UID: "79626ad3-4978-43f0-b163-fe482aa1dac7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:29:34 crc kubenswrapper[4860]: I1211 08:29:34.378044 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q5qmt\" (UniqueName: \"kubernetes.io/projected/79626ad3-4978-43f0-b163-fe482aa1dac7-kube-api-access-q5qmt\") on node \"crc\" DevicePath \"\"" Dec 11 08:29:34 crc kubenswrapper[4860]: I1211 08:29:34.378069 4860 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/79626ad3-4978-43f0-b163-fe482aa1dac7-config\") on node \"crc\" DevicePath \"\"" Dec 11 08:29:34 crc kubenswrapper[4860]: I1211 08:29:34.385083 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/79626ad3-4978-43f0-b163-fe482aa1dac7-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "79626ad3-4978-43f0-b163-fe482aa1dac7" (UID: "79626ad3-4978-43f0-b163-fe482aa1dac7"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:29:34 crc kubenswrapper[4860]: I1211 08:29:34.402360 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/79626ad3-4978-43f0-b163-fe482aa1dac7-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "79626ad3-4978-43f0-b163-fe482aa1dac7" (UID: "79626ad3-4978-43f0-b163-fe482aa1dac7"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:29:34 crc kubenswrapper[4860]: I1211 08:29:34.434051 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/79626ad3-4978-43f0-b163-fe482aa1dac7-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "79626ad3-4978-43f0-b163-fe482aa1dac7" (UID: "79626ad3-4978-43f0-b163-fe482aa1dac7"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:29:34 crc kubenswrapper[4860]: I1211 08:29:34.493833 4860 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/79626ad3-4978-43f0-b163-fe482aa1dac7-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 11 08:29:34 crc kubenswrapper[4860]: I1211 08:29:34.493869 4860 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/79626ad3-4978-43f0-b163-fe482aa1dac7-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 11 08:29:34 crc kubenswrapper[4860]: I1211 08:29:34.493879 4860 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/79626ad3-4978-43f0-b163-fe482aa1dac7-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 11 08:29:34 crc kubenswrapper[4860]: I1211 08:29:34.512791 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/79626ad3-4978-43f0-b163-fe482aa1dac7-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "79626ad3-4978-43f0-b163-fe482aa1dac7" (UID: "79626ad3-4978-43f0-b163-fe482aa1dac7"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:29:34 crc kubenswrapper[4860]: I1211 08:29:34.599046 4860 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/79626ad3-4978-43f0-b163-fe482aa1dac7-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 11 08:29:34 crc kubenswrapper[4860]: I1211 08:29:34.658146 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc5c4795-65k7g" event={"ID":"fe0e81d5-8772-4047-9b03-fd8127be2b66","Type":"ContainerStarted","Data":"5c1825a4e2c7afa36121cf72f71a0143cd39e74617e9246b561d6909c66cc58d"} Dec 11 08:29:34 crc kubenswrapper[4860]: I1211 08:29:34.659225 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5ccc5c4795-65k7g" Dec 11 08:29:34 crc kubenswrapper[4860]: I1211 08:29:34.681857 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-776755b658-9fwr9" event={"ID":"8f38e033-5243-4e65-a7e0-29f1282b063f","Type":"ContainerStarted","Data":"948c71446dedd81878e5444ab6a4dbd7fbfe0ce699ae4e007ae7d21b188368ec"} Dec 11 08:29:34 crc kubenswrapper[4860]: I1211 08:29:34.704520 4860 generic.go:334] "Generic (PLEG): container finished" podID="79626ad3-4978-43f0-b163-fe482aa1dac7" containerID="88aedbd26922eb622cb2c1a6ca543159d73a1595095e7fc2731e96c2fe0838cd" exitCode=0 Dec 11 08:29:34 crc kubenswrapper[4860]: I1211 08:29:34.704609 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57c957c4ff-wxx2m" event={"ID":"79626ad3-4978-43f0-b163-fe482aa1dac7","Type":"ContainerDied","Data":"88aedbd26922eb622cb2c1a6ca543159d73a1595095e7fc2731e96c2fe0838cd"} Dec 11 08:29:34 crc kubenswrapper[4860]: I1211 08:29:34.704636 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57c957c4ff-wxx2m" event={"ID":"79626ad3-4978-43f0-b163-fe482aa1dac7","Type":"ContainerDied","Data":"6fcd3aafd910f5701c1a5b87eae989d7c034b47b703be6ececcc925a8cb5f06c"} Dec 11 08:29:34 crc kubenswrapper[4860]: I1211 08:29:34.704668 4860 scope.go:117] "RemoveContainer" containerID="88aedbd26922eb622cb2c1a6ca543159d73a1595095e7fc2731e96c2fe0838cd" Dec 11 08:29:34 crc kubenswrapper[4860]: I1211 08:29:34.704811 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57c957c4ff-wxx2m" Dec 11 08:29:34 crc kubenswrapper[4860]: I1211 08:29:34.727506 4860 generic.go:334] "Generic (PLEG): container finished" podID="eafaa4e1-5187-4406-b55a-9d9349e3a744" containerID="5225f72cd5a3da2be39a5c10f3cec79dbb7979ee0b367d21f16de3045373d813" exitCode=0 Dec 11 08:29:34 crc kubenswrapper[4860]: I1211 08:29:34.727568 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-jqgbc" event={"ID":"eafaa4e1-5187-4406-b55a-9d9349e3a744","Type":"ContainerDied","Data":"5225f72cd5a3da2be39a5c10f3cec79dbb7979ee0b367d21f16de3045373d813"} Dec 11 08:29:34 crc kubenswrapper[4860]: I1211 08:29:34.744138 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-ccf7c6598-jcbgb" event={"ID":"e076681d-ddec-44a2-9ce4-39b87e5542c6","Type":"ContainerStarted","Data":"fd59343d1602e30d80627c28c579dc28a98e01ed5bdc68b6006a2b053ba08373"} Dec 11 08:29:34 crc kubenswrapper[4860]: I1211 08:29:34.744409 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-ccf7c6598-jcbgb" event={"ID":"e076681d-ddec-44a2-9ce4-39b87e5542c6","Type":"ContainerStarted","Data":"fcaf6875741f0622368b408b61196853ecbfe68ddd9ec0f7cf84241df6b96926"} Dec 11 08:29:34 crc kubenswrapper[4860]: I1211 08:29:34.744633 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-ccf7c6598-jcbgb" Dec 11 08:29:34 crc kubenswrapper[4860]: I1211 08:29:34.746551 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-776755b658-9fwr9" podStartSLOduration=26.782582795 podStartE2EDuration="28.746526823s" podCreationTimestamp="2025-12-11 08:29:06 +0000 UTC" firstStartedPulling="2025-12-11 08:29:30.451407473 +0000 UTC m=+1103.179926528" lastFinishedPulling="2025-12-11 08:29:32.415351501 +0000 UTC m=+1105.143870556" observedRunningTime="2025-12-11 08:29:34.725177611 +0000 UTC m=+1107.453696666" watchObservedRunningTime="2025-12-11 08:29:34.746526823 +0000 UTC m=+1107.475045878" Dec 11 08:29:34 crc kubenswrapper[4860]: I1211 08:29:34.747663 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5ccc5c4795-65k7g" podStartSLOduration=3.747658118 podStartE2EDuration="3.747658118s" podCreationTimestamp="2025-12-11 08:29:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:29:34.68825008 +0000 UTC m=+1107.416769135" watchObservedRunningTime="2025-12-11 08:29:34.747658118 +0000 UTC m=+1107.476177173" Dec 11 08:29:34 crc kubenswrapper[4860]: I1211 08:29:34.772181 4860 scope.go:117] "RemoveContainer" containerID="dc6c2ac31f2af352f724e5c89081aec87b45b1b7a692bb7fce71b55091cfad48" Dec 11 08:29:34 crc kubenswrapper[4860]: I1211 08:29:34.792344 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57c957c4ff-wxx2m"] Dec 11 08:29:34 crc kubenswrapper[4860]: I1211 08:29:34.806427 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57c957c4ff-wxx2m"] Dec 11 08:29:34 crc kubenswrapper[4860]: I1211 08:29:34.818962 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-ccf7c6598-jcbgb" podStartSLOduration=3.818937888 podStartE2EDuration="3.818937888s" podCreationTimestamp="2025-12-11 08:29:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:29:34.795385198 +0000 UTC m=+1107.523904253" watchObservedRunningTime="2025-12-11 08:29:34.818937888 +0000 UTC m=+1107.547456953" Dec 11 08:29:34 crc kubenswrapper[4860]: I1211 08:29:34.839108 4860 scope.go:117] "RemoveContainer" containerID="88aedbd26922eb622cb2c1a6ca543159d73a1595095e7fc2731e96c2fe0838cd" Dec 11 08:29:34 crc kubenswrapper[4860]: E1211 08:29:34.842221 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"88aedbd26922eb622cb2c1a6ca543159d73a1595095e7fc2731e96c2fe0838cd\": container with ID starting with 88aedbd26922eb622cb2c1a6ca543159d73a1595095e7fc2731e96c2fe0838cd not found: ID does not exist" containerID="88aedbd26922eb622cb2c1a6ca543159d73a1595095e7fc2731e96c2fe0838cd" Dec 11 08:29:34 crc kubenswrapper[4860]: I1211 08:29:34.842281 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"88aedbd26922eb622cb2c1a6ca543159d73a1595095e7fc2731e96c2fe0838cd"} err="failed to get container status \"88aedbd26922eb622cb2c1a6ca543159d73a1595095e7fc2731e96c2fe0838cd\": rpc error: code = NotFound desc = could not find container \"88aedbd26922eb622cb2c1a6ca543159d73a1595095e7fc2731e96c2fe0838cd\": container with ID starting with 88aedbd26922eb622cb2c1a6ca543159d73a1595095e7fc2731e96c2fe0838cd not found: ID does not exist" Dec 11 08:29:34 crc kubenswrapper[4860]: I1211 08:29:34.842321 4860 scope.go:117] "RemoveContainer" containerID="dc6c2ac31f2af352f724e5c89081aec87b45b1b7a692bb7fce71b55091cfad48" Dec 11 08:29:34 crc kubenswrapper[4860]: E1211 08:29:34.842771 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dc6c2ac31f2af352f724e5c89081aec87b45b1b7a692bb7fce71b55091cfad48\": container with ID starting with dc6c2ac31f2af352f724e5c89081aec87b45b1b7a692bb7fce71b55091cfad48 not found: ID does not exist" containerID="dc6c2ac31f2af352f724e5c89081aec87b45b1b7a692bb7fce71b55091cfad48" Dec 11 08:29:34 crc kubenswrapper[4860]: I1211 08:29:34.842822 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc6c2ac31f2af352f724e5c89081aec87b45b1b7a692bb7fce71b55091cfad48"} err="failed to get container status \"dc6c2ac31f2af352f724e5c89081aec87b45b1b7a692bb7fce71b55091cfad48\": rpc error: code = NotFound desc = could not find container \"dc6c2ac31f2af352f724e5c89081aec87b45b1b7a692bb7fce71b55091cfad48\": container with ID starting with dc6c2ac31f2af352f724e5c89081aec87b45b1b7a692bb7fce71b55091cfad48 not found: ID does not exist" Dec 11 08:29:35 crc kubenswrapper[4860]: I1211 08:29:35.589225 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="79626ad3-4978-43f0-b163-fe482aa1dac7" path="/var/lib/kubelet/pods/79626ad3-4978-43f0-b163-fe482aa1dac7/volumes" Dec 11 08:29:35 crc kubenswrapper[4860]: I1211 08:29:35.757344 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f153e5dc-0977-40b7-9ed6-ad88919f1081","Type":"ContainerStarted","Data":"d04c5444a6476bde178dcd80a442b4df937e8d9d28bf7bb2fee81b80764223de"} Dec 11 08:29:35 crc kubenswrapper[4860]: I1211 08:29:35.757722 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="f153e5dc-0977-40b7-9ed6-ad88919f1081" containerName="glance-log" containerID="cri-o://fbe50653cd9fe78c494f42fa61f7c07b554d82034fe17558bd1e3ed88df56056" gracePeriod=30 Dec 11 08:29:35 crc kubenswrapper[4860]: I1211 08:29:35.757767 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="f153e5dc-0977-40b7-9ed6-ad88919f1081" containerName="glance-httpd" containerID="cri-o://d04c5444a6476bde178dcd80a442b4df937e8d9d28bf7bb2fee81b80764223de" gracePeriod=30 Dec 11 08:29:35 crc kubenswrapper[4860]: I1211 08:29:35.766656 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"67ac03d0-1fbe-46ff-b61d-7d15bcf48f61","Type":"ContainerStarted","Data":"66ff380bf4d057ea3618be3f90a82f97b08d4418f62be5241db90a97a3ccf924"} Dec 11 08:29:35 crc kubenswrapper[4860]: I1211 08:29:35.766809 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="67ac03d0-1fbe-46ff-b61d-7d15bcf48f61" containerName="glance-log" containerID="cri-o://5d29695ad5aece20f85dcf129d4f9a829958abfeec4b99546f362ca3993bb5d0" gracePeriod=30 Dec 11 08:29:35 crc kubenswrapper[4860]: I1211 08:29:35.767036 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="67ac03d0-1fbe-46ff-b61d-7d15bcf48f61" containerName="glance-httpd" containerID="cri-o://66ff380bf4d057ea3618be3f90a82f97b08d4418f62be5241db90a97a3ccf924" gracePeriod=30 Dec 11 08:29:35 crc kubenswrapper[4860]: I1211 08:29:35.768452 4860 generic.go:334] "Generic (PLEG): container finished" podID="4ff61ef1-6a67-4a8e-80e2-11f0154e54c6" containerID="2c222aa49386e8cb938f0cf793517fceae739a76f3c585759558399c58a07d51" exitCode=0 Dec 11 08:29:35 crc kubenswrapper[4860]: I1211 08:29:35.769224 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-pfz4t" event={"ID":"4ff61ef1-6a67-4a8e-80e2-11f0154e54c6","Type":"ContainerDied","Data":"2c222aa49386e8cb938f0cf793517fceae739a76f3c585759558399c58a07d51"} Dec 11 08:29:35 crc kubenswrapper[4860]: I1211 08:29:35.785931 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=6.785910861 podStartE2EDuration="6.785910861s" podCreationTimestamp="2025-12-11 08:29:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:29:35.780436649 +0000 UTC m=+1108.508955714" watchObservedRunningTime="2025-12-11 08:29:35.785910861 +0000 UTC m=+1108.514429916" Dec 11 08:29:35 crc kubenswrapper[4860]: I1211 08:29:35.828394 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=6.828363815 podStartE2EDuration="6.828363815s" podCreationTimestamp="2025-12-11 08:29:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:29:35.826127125 +0000 UTC m=+1108.554646180" watchObservedRunningTime="2025-12-11 08:29:35.828363815 +0000 UTC m=+1108.556883030" Dec 11 08:29:36 crc kubenswrapper[4860]: I1211 08:29:36.306776 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-5fccb6555c-xwc9b"] Dec 11 08:29:36 crc kubenswrapper[4860]: E1211 08:29:36.307271 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79626ad3-4978-43f0-b163-fe482aa1dac7" containerName="dnsmasq-dns" Dec 11 08:29:36 crc kubenswrapper[4860]: I1211 08:29:36.307287 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="79626ad3-4978-43f0-b163-fe482aa1dac7" containerName="dnsmasq-dns" Dec 11 08:29:36 crc kubenswrapper[4860]: E1211 08:29:36.307300 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79626ad3-4978-43f0-b163-fe482aa1dac7" containerName="init" Dec 11 08:29:36 crc kubenswrapper[4860]: I1211 08:29:36.307306 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="79626ad3-4978-43f0-b163-fe482aa1dac7" containerName="init" Dec 11 08:29:36 crc kubenswrapper[4860]: I1211 08:29:36.307501 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="79626ad3-4978-43f0-b163-fe482aa1dac7" containerName="dnsmasq-dns" Dec 11 08:29:36 crc kubenswrapper[4860]: I1211 08:29:36.308569 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5fccb6555c-xwc9b" Dec 11 08:29:36 crc kubenswrapper[4860]: I1211 08:29:36.313544 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Dec 11 08:29:36 crc kubenswrapper[4860]: I1211 08:29:36.313905 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Dec 11 08:29:36 crc kubenswrapper[4860]: I1211 08:29:36.320599 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5fccb6555c-xwc9b"] Dec 11 08:29:36 crc kubenswrapper[4860]: I1211 08:29:36.397980 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-jqgbc" Dec 11 08:29:36 crc kubenswrapper[4860]: I1211 08:29:36.448184 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7cxwk\" (UniqueName: \"kubernetes.io/projected/eafaa4e1-5187-4406-b55a-9d9349e3a744-kube-api-access-7cxwk\") pod \"eafaa4e1-5187-4406-b55a-9d9349e3a744\" (UID: \"eafaa4e1-5187-4406-b55a-9d9349e3a744\") " Dec 11 08:29:36 crc kubenswrapper[4860]: I1211 08:29:36.448365 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eafaa4e1-5187-4406-b55a-9d9349e3a744-config-data\") pod \"eafaa4e1-5187-4406-b55a-9d9349e3a744\" (UID: \"eafaa4e1-5187-4406-b55a-9d9349e3a744\") " Dec 11 08:29:36 crc kubenswrapper[4860]: I1211 08:29:36.448405 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eafaa4e1-5187-4406-b55a-9d9349e3a744-combined-ca-bundle\") pod \"eafaa4e1-5187-4406-b55a-9d9349e3a744\" (UID: \"eafaa4e1-5187-4406-b55a-9d9349e3a744\") " Dec 11 08:29:36 crc kubenswrapper[4860]: I1211 08:29:36.448472 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eafaa4e1-5187-4406-b55a-9d9349e3a744-scripts\") pod \"eafaa4e1-5187-4406-b55a-9d9349e3a744\" (UID: \"eafaa4e1-5187-4406-b55a-9d9349e3a744\") " Dec 11 08:29:36 crc kubenswrapper[4860]: I1211 08:29:36.448541 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eafaa4e1-5187-4406-b55a-9d9349e3a744-logs\") pod \"eafaa4e1-5187-4406-b55a-9d9349e3a744\" (UID: \"eafaa4e1-5187-4406-b55a-9d9349e3a744\") " Dec 11 08:29:36 crc kubenswrapper[4860]: I1211 08:29:36.448904 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7739fee0-ccf5-49e9-bc21-4acd841daf55-ovndb-tls-certs\") pod \"neutron-5fccb6555c-xwc9b\" (UID: \"7739fee0-ccf5-49e9-bc21-4acd841daf55\") " pod="openstack/neutron-5fccb6555c-xwc9b" Dec 11 08:29:36 crc kubenswrapper[4860]: I1211 08:29:36.448931 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/7739fee0-ccf5-49e9-bc21-4acd841daf55-httpd-config\") pod \"neutron-5fccb6555c-xwc9b\" (UID: \"7739fee0-ccf5-49e9-bc21-4acd841daf55\") " pod="openstack/neutron-5fccb6555c-xwc9b" Dec 11 08:29:36 crc kubenswrapper[4860]: I1211 08:29:36.449003 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sgcld\" (UniqueName: \"kubernetes.io/projected/7739fee0-ccf5-49e9-bc21-4acd841daf55-kube-api-access-sgcld\") pod \"neutron-5fccb6555c-xwc9b\" (UID: \"7739fee0-ccf5-49e9-bc21-4acd841daf55\") " pod="openstack/neutron-5fccb6555c-xwc9b" Dec 11 08:29:36 crc kubenswrapper[4860]: I1211 08:29:36.449039 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7739fee0-ccf5-49e9-bc21-4acd841daf55-internal-tls-certs\") pod \"neutron-5fccb6555c-xwc9b\" (UID: \"7739fee0-ccf5-49e9-bc21-4acd841daf55\") " pod="openstack/neutron-5fccb6555c-xwc9b" Dec 11 08:29:36 crc kubenswrapper[4860]: I1211 08:29:36.449057 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/7739fee0-ccf5-49e9-bc21-4acd841daf55-config\") pod \"neutron-5fccb6555c-xwc9b\" (UID: \"7739fee0-ccf5-49e9-bc21-4acd841daf55\") " pod="openstack/neutron-5fccb6555c-xwc9b" Dec 11 08:29:36 crc kubenswrapper[4860]: I1211 08:29:36.449071 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7739fee0-ccf5-49e9-bc21-4acd841daf55-public-tls-certs\") pod \"neutron-5fccb6555c-xwc9b\" (UID: \"7739fee0-ccf5-49e9-bc21-4acd841daf55\") " pod="openstack/neutron-5fccb6555c-xwc9b" Dec 11 08:29:36 crc kubenswrapper[4860]: I1211 08:29:36.449168 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7739fee0-ccf5-49e9-bc21-4acd841daf55-combined-ca-bundle\") pod \"neutron-5fccb6555c-xwc9b\" (UID: \"7739fee0-ccf5-49e9-bc21-4acd841daf55\") " pod="openstack/neutron-5fccb6555c-xwc9b" Dec 11 08:29:36 crc kubenswrapper[4860]: I1211 08:29:36.452716 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eafaa4e1-5187-4406-b55a-9d9349e3a744-logs" (OuterVolumeSpecName: "logs") pod "eafaa4e1-5187-4406-b55a-9d9349e3a744" (UID: "eafaa4e1-5187-4406-b55a-9d9349e3a744"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:29:36 crc kubenswrapper[4860]: I1211 08:29:36.457372 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eafaa4e1-5187-4406-b55a-9d9349e3a744-kube-api-access-7cxwk" (OuterVolumeSpecName: "kube-api-access-7cxwk") pod "eafaa4e1-5187-4406-b55a-9d9349e3a744" (UID: "eafaa4e1-5187-4406-b55a-9d9349e3a744"). InnerVolumeSpecName "kube-api-access-7cxwk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:29:36 crc kubenswrapper[4860]: I1211 08:29:36.462874 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eafaa4e1-5187-4406-b55a-9d9349e3a744-scripts" (OuterVolumeSpecName: "scripts") pod "eafaa4e1-5187-4406-b55a-9d9349e3a744" (UID: "eafaa4e1-5187-4406-b55a-9d9349e3a744"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:29:36 crc kubenswrapper[4860]: I1211 08:29:36.488023 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eafaa4e1-5187-4406-b55a-9d9349e3a744-config-data" (OuterVolumeSpecName: "config-data") pod "eafaa4e1-5187-4406-b55a-9d9349e3a744" (UID: "eafaa4e1-5187-4406-b55a-9d9349e3a744"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:29:36 crc kubenswrapper[4860]: I1211 08:29:36.520386 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eafaa4e1-5187-4406-b55a-9d9349e3a744-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "eafaa4e1-5187-4406-b55a-9d9349e3a744" (UID: "eafaa4e1-5187-4406-b55a-9d9349e3a744"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:29:36 crc kubenswrapper[4860]: I1211 08:29:36.550958 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7739fee0-ccf5-49e9-bc21-4acd841daf55-combined-ca-bundle\") pod \"neutron-5fccb6555c-xwc9b\" (UID: \"7739fee0-ccf5-49e9-bc21-4acd841daf55\") " pod="openstack/neutron-5fccb6555c-xwc9b" Dec 11 08:29:36 crc kubenswrapper[4860]: I1211 08:29:36.551021 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7739fee0-ccf5-49e9-bc21-4acd841daf55-ovndb-tls-certs\") pod \"neutron-5fccb6555c-xwc9b\" (UID: \"7739fee0-ccf5-49e9-bc21-4acd841daf55\") " pod="openstack/neutron-5fccb6555c-xwc9b" Dec 11 08:29:36 crc kubenswrapper[4860]: I1211 08:29:36.551044 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/7739fee0-ccf5-49e9-bc21-4acd841daf55-httpd-config\") pod \"neutron-5fccb6555c-xwc9b\" (UID: \"7739fee0-ccf5-49e9-bc21-4acd841daf55\") " pod="openstack/neutron-5fccb6555c-xwc9b" Dec 11 08:29:36 crc kubenswrapper[4860]: I1211 08:29:36.551099 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sgcld\" (UniqueName: \"kubernetes.io/projected/7739fee0-ccf5-49e9-bc21-4acd841daf55-kube-api-access-sgcld\") pod \"neutron-5fccb6555c-xwc9b\" (UID: \"7739fee0-ccf5-49e9-bc21-4acd841daf55\") " pod="openstack/neutron-5fccb6555c-xwc9b" Dec 11 08:29:36 crc kubenswrapper[4860]: I1211 08:29:36.551130 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7739fee0-ccf5-49e9-bc21-4acd841daf55-internal-tls-certs\") pod \"neutron-5fccb6555c-xwc9b\" (UID: \"7739fee0-ccf5-49e9-bc21-4acd841daf55\") " pod="openstack/neutron-5fccb6555c-xwc9b" Dec 11 08:29:36 crc kubenswrapper[4860]: I1211 08:29:36.551149 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/7739fee0-ccf5-49e9-bc21-4acd841daf55-config\") pod \"neutron-5fccb6555c-xwc9b\" (UID: \"7739fee0-ccf5-49e9-bc21-4acd841daf55\") " pod="openstack/neutron-5fccb6555c-xwc9b" Dec 11 08:29:36 crc kubenswrapper[4860]: I1211 08:29:36.551166 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7739fee0-ccf5-49e9-bc21-4acd841daf55-public-tls-certs\") pod \"neutron-5fccb6555c-xwc9b\" (UID: \"7739fee0-ccf5-49e9-bc21-4acd841daf55\") " pod="openstack/neutron-5fccb6555c-xwc9b" Dec 11 08:29:36 crc kubenswrapper[4860]: I1211 08:29:36.551257 4860 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eafaa4e1-5187-4406-b55a-9d9349e3a744-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 08:29:36 crc kubenswrapper[4860]: I1211 08:29:36.551268 4860 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eafaa4e1-5187-4406-b55a-9d9349e3a744-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 08:29:36 crc kubenswrapper[4860]: I1211 08:29:36.551277 4860 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eafaa4e1-5187-4406-b55a-9d9349e3a744-logs\") on node \"crc\" DevicePath \"\"" Dec 11 08:29:36 crc kubenswrapper[4860]: I1211 08:29:36.551286 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7cxwk\" (UniqueName: \"kubernetes.io/projected/eafaa4e1-5187-4406-b55a-9d9349e3a744-kube-api-access-7cxwk\") on node \"crc\" DevicePath \"\"" Dec 11 08:29:36 crc kubenswrapper[4860]: I1211 08:29:36.551296 4860 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eafaa4e1-5187-4406-b55a-9d9349e3a744-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 08:29:36 crc kubenswrapper[4860]: I1211 08:29:36.556498 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7739fee0-ccf5-49e9-bc21-4acd841daf55-public-tls-certs\") pod \"neutron-5fccb6555c-xwc9b\" (UID: \"7739fee0-ccf5-49e9-bc21-4acd841daf55\") " pod="openstack/neutron-5fccb6555c-xwc9b" Dec 11 08:29:36 crc kubenswrapper[4860]: I1211 08:29:36.557519 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/7739fee0-ccf5-49e9-bc21-4acd841daf55-config\") pod \"neutron-5fccb6555c-xwc9b\" (UID: \"7739fee0-ccf5-49e9-bc21-4acd841daf55\") " pod="openstack/neutron-5fccb6555c-xwc9b" Dec 11 08:29:36 crc kubenswrapper[4860]: I1211 08:29:36.559294 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7739fee0-ccf5-49e9-bc21-4acd841daf55-ovndb-tls-certs\") pod \"neutron-5fccb6555c-xwc9b\" (UID: \"7739fee0-ccf5-49e9-bc21-4acd841daf55\") " pod="openstack/neutron-5fccb6555c-xwc9b" Dec 11 08:29:36 crc kubenswrapper[4860]: I1211 08:29:36.559573 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7739fee0-ccf5-49e9-bc21-4acd841daf55-combined-ca-bundle\") pod \"neutron-5fccb6555c-xwc9b\" (UID: \"7739fee0-ccf5-49e9-bc21-4acd841daf55\") " pod="openstack/neutron-5fccb6555c-xwc9b" Dec 11 08:29:36 crc kubenswrapper[4860]: I1211 08:29:36.560102 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/7739fee0-ccf5-49e9-bc21-4acd841daf55-httpd-config\") pod \"neutron-5fccb6555c-xwc9b\" (UID: \"7739fee0-ccf5-49e9-bc21-4acd841daf55\") " pod="openstack/neutron-5fccb6555c-xwc9b" Dec 11 08:29:36 crc kubenswrapper[4860]: I1211 08:29:36.560877 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7739fee0-ccf5-49e9-bc21-4acd841daf55-internal-tls-certs\") pod \"neutron-5fccb6555c-xwc9b\" (UID: \"7739fee0-ccf5-49e9-bc21-4acd841daf55\") " pod="openstack/neutron-5fccb6555c-xwc9b" Dec 11 08:29:36 crc kubenswrapper[4860]: I1211 08:29:36.590881 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sgcld\" (UniqueName: \"kubernetes.io/projected/7739fee0-ccf5-49e9-bc21-4acd841daf55-kube-api-access-sgcld\") pod \"neutron-5fccb6555c-xwc9b\" (UID: \"7739fee0-ccf5-49e9-bc21-4acd841daf55\") " pod="openstack/neutron-5fccb6555c-xwc9b" Dec 11 08:29:36 crc kubenswrapper[4860]: I1211 08:29:36.698766 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-776755b658-9fwr9" Dec 11 08:29:36 crc kubenswrapper[4860]: I1211 08:29:36.702752 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-776755b658-9fwr9" Dec 11 08:29:36 crc kubenswrapper[4860]: I1211 08:29:36.725812 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5fccb6555c-xwc9b" Dec 11 08:29:36 crc kubenswrapper[4860]: I1211 08:29:36.799172 4860 generic.go:334] "Generic (PLEG): container finished" podID="f153e5dc-0977-40b7-9ed6-ad88919f1081" containerID="d04c5444a6476bde178dcd80a442b4df937e8d9d28bf7bb2fee81b80764223de" exitCode=0 Dec 11 08:29:36 crc kubenswrapper[4860]: I1211 08:29:36.799381 4860 generic.go:334] "Generic (PLEG): container finished" podID="f153e5dc-0977-40b7-9ed6-ad88919f1081" containerID="fbe50653cd9fe78c494f42fa61f7c07b554d82034fe17558bd1e3ed88df56056" exitCode=143 Dec 11 08:29:36 crc kubenswrapper[4860]: I1211 08:29:36.799422 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f153e5dc-0977-40b7-9ed6-ad88919f1081","Type":"ContainerDied","Data":"d04c5444a6476bde178dcd80a442b4df937e8d9d28bf7bb2fee81b80764223de"} Dec 11 08:29:36 crc kubenswrapper[4860]: I1211 08:29:36.799450 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f153e5dc-0977-40b7-9ed6-ad88919f1081","Type":"ContainerDied","Data":"fbe50653cd9fe78c494f42fa61f7c07b554d82034fe17558bd1e3ed88df56056"} Dec 11 08:29:36 crc kubenswrapper[4860]: I1211 08:29:36.833311 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 11 08:29:36 crc kubenswrapper[4860]: I1211 08:29:36.833925 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-6b4697fc68-vnd26" Dec 11 08:29:36 crc kubenswrapper[4860]: I1211 08:29:36.834737 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-6b4697fc68-vnd26" Dec 11 08:29:36 crc kubenswrapper[4860]: I1211 08:29:36.844402 4860 generic.go:334] "Generic (PLEG): container finished" podID="67ac03d0-1fbe-46ff-b61d-7d15bcf48f61" containerID="66ff380bf4d057ea3618be3f90a82f97b08d4418f62be5241db90a97a3ccf924" exitCode=0 Dec 11 08:29:36 crc kubenswrapper[4860]: I1211 08:29:36.844434 4860 generic.go:334] "Generic (PLEG): container finished" podID="67ac03d0-1fbe-46ff-b61d-7d15bcf48f61" containerID="5d29695ad5aece20f85dcf129d4f9a829958abfeec4b99546f362ca3993bb5d0" exitCode=143 Dec 11 08:29:36 crc kubenswrapper[4860]: I1211 08:29:36.844477 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"67ac03d0-1fbe-46ff-b61d-7d15bcf48f61","Type":"ContainerDied","Data":"66ff380bf4d057ea3618be3f90a82f97b08d4418f62be5241db90a97a3ccf924"} Dec 11 08:29:36 crc kubenswrapper[4860]: I1211 08:29:36.844503 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"67ac03d0-1fbe-46ff-b61d-7d15bcf48f61","Type":"ContainerDied","Data":"5d29695ad5aece20f85dcf129d4f9a829958abfeec4b99546f362ca3993bb5d0"} Dec 11 08:29:36 crc kubenswrapper[4860]: I1211 08:29:36.844521 4860 scope.go:117] "RemoveContainer" containerID="66ff380bf4d057ea3618be3f90a82f97b08d4418f62be5241db90a97a3ccf924" Dec 11 08:29:36 crc kubenswrapper[4860]: I1211 08:29:36.855258 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-jqgbc" Dec 11 08:29:36 crc kubenswrapper[4860]: I1211 08:29:36.855320 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-jqgbc" event={"ID":"eafaa4e1-5187-4406-b55a-9d9349e3a744","Type":"ContainerDied","Data":"180aa31559929b074de4c1faefdc518264286cf7496070de82b469144a62a978"} Dec 11 08:29:36 crc kubenswrapper[4860]: I1211 08:29:36.855367 4860 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="180aa31559929b074de4c1faefdc518264286cf7496070de82b469144a62a978" Dec 11 08:29:36 crc kubenswrapper[4860]: I1211 08:29:36.859692 4860 generic.go:334] "Generic (PLEG): container finished" podID="fecaee75-2dc0-4d48-9d40-11fc1ec8e972" containerID="b5e16e2ff7b9a93d183828e46286db5391626046b43b832b915a44d625d1f621" exitCode=0 Dec 11 08:29:36 crc kubenswrapper[4860]: I1211 08:29:36.859829 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-plxxc" event={"ID":"fecaee75-2dc0-4d48-9d40-11fc1ec8e972","Type":"ContainerDied","Data":"b5e16e2ff7b9a93d183828e46286db5391626046b43b832b915a44d625d1f621"} Dec 11 08:29:36 crc kubenswrapper[4860]: I1211 08:29:36.931890 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-55ccdb6bbd-grdt6"] Dec 11 08:29:36 crc kubenswrapper[4860]: E1211 08:29:36.932456 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67ac03d0-1fbe-46ff-b61d-7d15bcf48f61" containerName="glance-log" Dec 11 08:29:36 crc kubenswrapper[4860]: I1211 08:29:36.932476 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="67ac03d0-1fbe-46ff-b61d-7d15bcf48f61" containerName="glance-log" Dec 11 08:29:36 crc kubenswrapper[4860]: E1211 08:29:36.932494 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67ac03d0-1fbe-46ff-b61d-7d15bcf48f61" containerName="glance-httpd" Dec 11 08:29:36 crc kubenswrapper[4860]: I1211 08:29:36.932500 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="67ac03d0-1fbe-46ff-b61d-7d15bcf48f61" containerName="glance-httpd" Dec 11 08:29:36 crc kubenswrapper[4860]: E1211 08:29:36.932515 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eafaa4e1-5187-4406-b55a-9d9349e3a744" containerName="placement-db-sync" Dec 11 08:29:36 crc kubenswrapper[4860]: I1211 08:29:36.932541 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="eafaa4e1-5187-4406-b55a-9d9349e3a744" containerName="placement-db-sync" Dec 11 08:29:36 crc kubenswrapper[4860]: I1211 08:29:36.932744 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="67ac03d0-1fbe-46ff-b61d-7d15bcf48f61" containerName="glance-httpd" Dec 11 08:29:36 crc kubenswrapper[4860]: I1211 08:29:36.932761 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="eafaa4e1-5187-4406-b55a-9d9349e3a744" containerName="placement-db-sync" Dec 11 08:29:36 crc kubenswrapper[4860]: I1211 08:29:36.932776 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="67ac03d0-1fbe-46ff-b61d-7d15bcf48f61" containerName="glance-log" Dec 11 08:29:36 crc kubenswrapper[4860]: I1211 08:29:36.933751 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-55ccdb6bbd-grdt6" Dec 11 08:29:36 crc kubenswrapper[4860]: I1211 08:29:36.941310 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Dec 11 08:29:36 crc kubenswrapper[4860]: I1211 08:29:36.941542 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 11 08:29:36 crc kubenswrapper[4860]: I1211 08:29:36.947516 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 11 08:29:36 crc kubenswrapper[4860]: I1211 08:29:36.948437 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Dec 11 08:29:36 crc kubenswrapper[4860]: I1211 08:29:36.948769 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-9dlgs" Dec 11 08:29:36 crc kubenswrapper[4860]: I1211 08:29:36.962628 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67ac03d0-1fbe-46ff-b61d-7d15bcf48f61-combined-ca-bundle\") pod \"67ac03d0-1fbe-46ff-b61d-7d15bcf48f61\" (UID: \"67ac03d0-1fbe-46ff-b61d-7d15bcf48f61\") " Dec 11 08:29:36 crc kubenswrapper[4860]: I1211 08:29:36.962672 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67ac03d0-1fbe-46ff-b61d-7d15bcf48f61-scripts\") pod \"67ac03d0-1fbe-46ff-b61d-7d15bcf48f61\" (UID: \"67ac03d0-1fbe-46ff-b61d-7d15bcf48f61\") " Dec 11 08:29:36 crc kubenswrapper[4860]: I1211 08:29:36.966294 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-55ccdb6bbd-grdt6"] Dec 11 08:29:36 crc kubenswrapper[4860]: I1211 08:29:36.967547 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"67ac03d0-1fbe-46ff-b61d-7d15bcf48f61\" (UID: \"67ac03d0-1fbe-46ff-b61d-7d15bcf48f61\") " Dec 11 08:29:36 crc kubenswrapper[4860]: I1211 08:29:36.967605 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/67ac03d0-1fbe-46ff-b61d-7d15bcf48f61-httpd-run\") pod \"67ac03d0-1fbe-46ff-b61d-7d15bcf48f61\" (UID: \"67ac03d0-1fbe-46ff-b61d-7d15bcf48f61\") " Dec 11 08:29:36 crc kubenswrapper[4860]: I1211 08:29:36.967830 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67ac03d0-1fbe-46ff-b61d-7d15bcf48f61-config-data\") pod \"67ac03d0-1fbe-46ff-b61d-7d15bcf48f61\" (UID: \"67ac03d0-1fbe-46ff-b61d-7d15bcf48f61\") " Dec 11 08:29:36 crc kubenswrapper[4860]: I1211 08:29:36.967870 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/67ac03d0-1fbe-46ff-b61d-7d15bcf48f61-logs\") pod \"67ac03d0-1fbe-46ff-b61d-7d15bcf48f61\" (UID: \"67ac03d0-1fbe-46ff-b61d-7d15bcf48f61\") " Dec 11 08:29:36 crc kubenswrapper[4860]: I1211 08:29:36.967914 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n2nxk\" (UniqueName: \"kubernetes.io/projected/67ac03d0-1fbe-46ff-b61d-7d15bcf48f61-kube-api-access-n2nxk\") pod \"67ac03d0-1fbe-46ff-b61d-7d15bcf48f61\" (UID: \"67ac03d0-1fbe-46ff-b61d-7d15bcf48f61\") " Dec 11 08:29:36 crc kubenswrapper[4860]: I1211 08:29:36.969112 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/67ac03d0-1fbe-46ff-b61d-7d15bcf48f61-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "67ac03d0-1fbe-46ff-b61d-7d15bcf48f61" (UID: "67ac03d0-1fbe-46ff-b61d-7d15bcf48f61"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:29:36 crc kubenswrapper[4860]: I1211 08:29:36.973838 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67ac03d0-1fbe-46ff-b61d-7d15bcf48f61-scripts" (OuterVolumeSpecName: "scripts") pod "67ac03d0-1fbe-46ff-b61d-7d15bcf48f61" (UID: "67ac03d0-1fbe-46ff-b61d-7d15bcf48f61"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:29:36 crc kubenswrapper[4860]: I1211 08:29:36.978653 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "glance") pod "67ac03d0-1fbe-46ff-b61d-7d15bcf48f61" (UID: "67ac03d0-1fbe-46ff-b61d-7d15bcf48f61"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 11 08:29:36 crc kubenswrapper[4860]: I1211 08:29:36.978857 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/67ac03d0-1fbe-46ff-b61d-7d15bcf48f61-logs" (OuterVolumeSpecName: "logs") pod "67ac03d0-1fbe-46ff-b61d-7d15bcf48f61" (UID: "67ac03d0-1fbe-46ff-b61d-7d15bcf48f61"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:29:36 crc kubenswrapper[4860]: I1211 08:29:36.988443 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67ac03d0-1fbe-46ff-b61d-7d15bcf48f61-kube-api-access-n2nxk" (OuterVolumeSpecName: "kube-api-access-n2nxk") pod "67ac03d0-1fbe-46ff-b61d-7d15bcf48f61" (UID: "67ac03d0-1fbe-46ff-b61d-7d15bcf48f61"). InnerVolumeSpecName "kube-api-access-n2nxk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:29:37 crc kubenswrapper[4860]: I1211 08:29:37.048378 4860 scope.go:117] "RemoveContainer" containerID="5d29695ad5aece20f85dcf129d4f9a829958abfeec4b99546f362ca3993bb5d0" Dec 11 08:29:37 crc kubenswrapper[4860]: I1211 08:29:37.072653 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/78683923-6a60-4a9a-bb81-51ae61f55376-internal-tls-certs\") pod \"placement-55ccdb6bbd-grdt6\" (UID: \"78683923-6a60-4a9a-bb81-51ae61f55376\") " pod="openstack/placement-55ccdb6bbd-grdt6" Dec 11 08:29:37 crc kubenswrapper[4860]: I1211 08:29:37.072719 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/78683923-6a60-4a9a-bb81-51ae61f55376-public-tls-certs\") pod \"placement-55ccdb6bbd-grdt6\" (UID: \"78683923-6a60-4a9a-bb81-51ae61f55376\") " pod="openstack/placement-55ccdb6bbd-grdt6" Dec 11 08:29:37 crc kubenswrapper[4860]: I1211 08:29:37.072756 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78683923-6a60-4a9a-bb81-51ae61f55376-combined-ca-bundle\") pod \"placement-55ccdb6bbd-grdt6\" (UID: \"78683923-6a60-4a9a-bb81-51ae61f55376\") " pod="openstack/placement-55ccdb6bbd-grdt6" Dec 11 08:29:37 crc kubenswrapper[4860]: I1211 08:29:37.072780 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/78683923-6a60-4a9a-bb81-51ae61f55376-scripts\") pod \"placement-55ccdb6bbd-grdt6\" (UID: \"78683923-6a60-4a9a-bb81-51ae61f55376\") " pod="openstack/placement-55ccdb6bbd-grdt6" Dec 11 08:29:37 crc kubenswrapper[4860]: I1211 08:29:37.072856 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78683923-6a60-4a9a-bb81-51ae61f55376-config-data\") pod \"placement-55ccdb6bbd-grdt6\" (UID: \"78683923-6a60-4a9a-bb81-51ae61f55376\") " pod="openstack/placement-55ccdb6bbd-grdt6" Dec 11 08:29:37 crc kubenswrapper[4860]: I1211 08:29:37.072879 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p4czj\" (UniqueName: \"kubernetes.io/projected/78683923-6a60-4a9a-bb81-51ae61f55376-kube-api-access-p4czj\") pod \"placement-55ccdb6bbd-grdt6\" (UID: \"78683923-6a60-4a9a-bb81-51ae61f55376\") " pod="openstack/placement-55ccdb6bbd-grdt6" Dec 11 08:29:37 crc kubenswrapper[4860]: I1211 08:29:37.072943 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/78683923-6a60-4a9a-bb81-51ae61f55376-logs\") pod \"placement-55ccdb6bbd-grdt6\" (UID: \"78683923-6a60-4a9a-bb81-51ae61f55376\") " pod="openstack/placement-55ccdb6bbd-grdt6" Dec 11 08:29:37 crc kubenswrapper[4860]: I1211 08:29:37.073006 4860 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Dec 11 08:29:37 crc kubenswrapper[4860]: I1211 08:29:37.073122 4860 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/67ac03d0-1fbe-46ff-b61d-7d15bcf48f61-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 11 08:29:37 crc kubenswrapper[4860]: I1211 08:29:37.073167 4860 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/67ac03d0-1fbe-46ff-b61d-7d15bcf48f61-logs\") on node \"crc\" DevicePath \"\"" Dec 11 08:29:37 crc kubenswrapper[4860]: I1211 08:29:37.073182 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n2nxk\" (UniqueName: \"kubernetes.io/projected/67ac03d0-1fbe-46ff-b61d-7d15bcf48f61-kube-api-access-n2nxk\") on node \"crc\" DevicePath \"\"" Dec 11 08:29:37 crc kubenswrapper[4860]: I1211 08:29:37.073194 4860 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67ac03d0-1fbe-46ff-b61d-7d15bcf48f61-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 08:29:37 crc kubenswrapper[4860]: I1211 08:29:37.097924 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67ac03d0-1fbe-46ff-b61d-7d15bcf48f61-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "67ac03d0-1fbe-46ff-b61d-7d15bcf48f61" (UID: "67ac03d0-1fbe-46ff-b61d-7d15bcf48f61"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:29:37 crc kubenswrapper[4860]: I1211 08:29:37.106226 4860 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Dec 11 08:29:37 crc kubenswrapper[4860]: I1211 08:29:37.119502 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67ac03d0-1fbe-46ff-b61d-7d15bcf48f61-config-data" (OuterVolumeSpecName: "config-data") pod "67ac03d0-1fbe-46ff-b61d-7d15bcf48f61" (UID: "67ac03d0-1fbe-46ff-b61d-7d15bcf48f61"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:29:37 crc kubenswrapper[4860]: I1211 08:29:37.181968 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/78683923-6a60-4a9a-bb81-51ae61f55376-internal-tls-certs\") pod \"placement-55ccdb6bbd-grdt6\" (UID: \"78683923-6a60-4a9a-bb81-51ae61f55376\") " pod="openstack/placement-55ccdb6bbd-grdt6" Dec 11 08:29:37 crc kubenswrapper[4860]: I1211 08:29:37.188008 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/78683923-6a60-4a9a-bb81-51ae61f55376-public-tls-certs\") pod \"placement-55ccdb6bbd-grdt6\" (UID: \"78683923-6a60-4a9a-bb81-51ae61f55376\") " pod="openstack/placement-55ccdb6bbd-grdt6" Dec 11 08:29:37 crc kubenswrapper[4860]: I1211 08:29:37.188135 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78683923-6a60-4a9a-bb81-51ae61f55376-combined-ca-bundle\") pod \"placement-55ccdb6bbd-grdt6\" (UID: \"78683923-6a60-4a9a-bb81-51ae61f55376\") " pod="openstack/placement-55ccdb6bbd-grdt6" Dec 11 08:29:37 crc kubenswrapper[4860]: I1211 08:29:37.188196 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/78683923-6a60-4a9a-bb81-51ae61f55376-scripts\") pod \"placement-55ccdb6bbd-grdt6\" (UID: \"78683923-6a60-4a9a-bb81-51ae61f55376\") " pod="openstack/placement-55ccdb6bbd-grdt6" Dec 11 08:29:37 crc kubenswrapper[4860]: I1211 08:29:37.188358 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78683923-6a60-4a9a-bb81-51ae61f55376-config-data\") pod \"placement-55ccdb6bbd-grdt6\" (UID: \"78683923-6a60-4a9a-bb81-51ae61f55376\") " pod="openstack/placement-55ccdb6bbd-grdt6" Dec 11 08:29:37 crc kubenswrapper[4860]: I1211 08:29:37.188390 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p4czj\" (UniqueName: \"kubernetes.io/projected/78683923-6a60-4a9a-bb81-51ae61f55376-kube-api-access-p4czj\") pod \"placement-55ccdb6bbd-grdt6\" (UID: \"78683923-6a60-4a9a-bb81-51ae61f55376\") " pod="openstack/placement-55ccdb6bbd-grdt6" Dec 11 08:29:37 crc kubenswrapper[4860]: I1211 08:29:37.188566 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/78683923-6a60-4a9a-bb81-51ae61f55376-logs\") pod \"placement-55ccdb6bbd-grdt6\" (UID: \"78683923-6a60-4a9a-bb81-51ae61f55376\") " pod="openstack/placement-55ccdb6bbd-grdt6" Dec 11 08:29:37 crc kubenswrapper[4860]: I1211 08:29:37.188698 4860 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Dec 11 08:29:37 crc kubenswrapper[4860]: I1211 08:29:37.188714 4860 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67ac03d0-1fbe-46ff-b61d-7d15bcf48f61-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 08:29:37 crc kubenswrapper[4860]: I1211 08:29:37.188730 4860 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67ac03d0-1fbe-46ff-b61d-7d15bcf48f61-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 08:29:37 crc kubenswrapper[4860]: I1211 08:29:37.189131 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/78683923-6a60-4a9a-bb81-51ae61f55376-logs\") pod \"placement-55ccdb6bbd-grdt6\" (UID: \"78683923-6a60-4a9a-bb81-51ae61f55376\") " pod="openstack/placement-55ccdb6bbd-grdt6" Dec 11 08:29:37 crc kubenswrapper[4860]: I1211 08:29:37.203394 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/78683923-6a60-4a9a-bb81-51ae61f55376-internal-tls-certs\") pod \"placement-55ccdb6bbd-grdt6\" (UID: \"78683923-6a60-4a9a-bb81-51ae61f55376\") " pod="openstack/placement-55ccdb6bbd-grdt6" Dec 11 08:29:37 crc kubenswrapper[4860]: I1211 08:29:37.205518 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78683923-6a60-4a9a-bb81-51ae61f55376-combined-ca-bundle\") pod \"placement-55ccdb6bbd-grdt6\" (UID: \"78683923-6a60-4a9a-bb81-51ae61f55376\") " pod="openstack/placement-55ccdb6bbd-grdt6" Dec 11 08:29:37 crc kubenswrapper[4860]: I1211 08:29:37.212137 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/78683923-6a60-4a9a-bb81-51ae61f55376-scripts\") pod \"placement-55ccdb6bbd-grdt6\" (UID: \"78683923-6a60-4a9a-bb81-51ae61f55376\") " pod="openstack/placement-55ccdb6bbd-grdt6" Dec 11 08:29:37 crc kubenswrapper[4860]: I1211 08:29:37.212896 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78683923-6a60-4a9a-bb81-51ae61f55376-config-data\") pod \"placement-55ccdb6bbd-grdt6\" (UID: \"78683923-6a60-4a9a-bb81-51ae61f55376\") " pod="openstack/placement-55ccdb6bbd-grdt6" Dec 11 08:29:37 crc kubenswrapper[4860]: I1211 08:29:37.218819 4860 scope.go:117] "RemoveContainer" containerID="66ff380bf4d057ea3618be3f90a82f97b08d4418f62be5241db90a97a3ccf924" Dec 11 08:29:37 crc kubenswrapper[4860]: I1211 08:29:37.223380 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p4czj\" (UniqueName: \"kubernetes.io/projected/78683923-6a60-4a9a-bb81-51ae61f55376-kube-api-access-p4czj\") pod \"placement-55ccdb6bbd-grdt6\" (UID: \"78683923-6a60-4a9a-bb81-51ae61f55376\") " pod="openstack/placement-55ccdb6bbd-grdt6" Dec 11 08:29:37 crc kubenswrapper[4860]: I1211 08:29:37.225290 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/78683923-6a60-4a9a-bb81-51ae61f55376-public-tls-certs\") pod \"placement-55ccdb6bbd-grdt6\" (UID: \"78683923-6a60-4a9a-bb81-51ae61f55376\") " pod="openstack/placement-55ccdb6bbd-grdt6" Dec 11 08:29:37 crc kubenswrapper[4860]: E1211 08:29:37.257629 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"66ff380bf4d057ea3618be3f90a82f97b08d4418f62be5241db90a97a3ccf924\": container with ID starting with 66ff380bf4d057ea3618be3f90a82f97b08d4418f62be5241db90a97a3ccf924 not found: ID does not exist" containerID="66ff380bf4d057ea3618be3f90a82f97b08d4418f62be5241db90a97a3ccf924" Dec 11 08:29:37 crc kubenswrapper[4860]: I1211 08:29:37.257706 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"66ff380bf4d057ea3618be3f90a82f97b08d4418f62be5241db90a97a3ccf924"} err="failed to get container status \"66ff380bf4d057ea3618be3f90a82f97b08d4418f62be5241db90a97a3ccf924\": rpc error: code = NotFound desc = could not find container \"66ff380bf4d057ea3618be3f90a82f97b08d4418f62be5241db90a97a3ccf924\": container with ID starting with 66ff380bf4d057ea3618be3f90a82f97b08d4418f62be5241db90a97a3ccf924 not found: ID does not exist" Dec 11 08:29:37 crc kubenswrapper[4860]: I1211 08:29:37.257737 4860 scope.go:117] "RemoveContainer" containerID="5d29695ad5aece20f85dcf129d4f9a829958abfeec4b99546f362ca3993bb5d0" Dec 11 08:29:37 crc kubenswrapper[4860]: E1211 08:29:37.262417 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5d29695ad5aece20f85dcf129d4f9a829958abfeec4b99546f362ca3993bb5d0\": container with ID starting with 5d29695ad5aece20f85dcf129d4f9a829958abfeec4b99546f362ca3993bb5d0 not found: ID does not exist" containerID="5d29695ad5aece20f85dcf129d4f9a829958abfeec4b99546f362ca3993bb5d0" Dec 11 08:29:37 crc kubenswrapper[4860]: I1211 08:29:37.262469 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d29695ad5aece20f85dcf129d4f9a829958abfeec4b99546f362ca3993bb5d0"} err="failed to get container status \"5d29695ad5aece20f85dcf129d4f9a829958abfeec4b99546f362ca3993bb5d0\": rpc error: code = NotFound desc = could not find container \"5d29695ad5aece20f85dcf129d4f9a829958abfeec4b99546f362ca3993bb5d0\": container with ID starting with 5d29695ad5aece20f85dcf129d4f9a829958abfeec4b99546f362ca3993bb5d0 not found: ID does not exist" Dec 11 08:29:37 crc kubenswrapper[4860]: I1211 08:29:37.265913 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 11 08:29:37 crc kubenswrapper[4860]: I1211 08:29:37.316585 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"f153e5dc-0977-40b7-9ed6-ad88919f1081\" (UID: \"f153e5dc-0977-40b7-9ed6-ad88919f1081\") " Dec 11 08:29:37 crc kubenswrapper[4860]: I1211 08:29:37.316689 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f153e5dc-0977-40b7-9ed6-ad88919f1081-logs\") pod \"f153e5dc-0977-40b7-9ed6-ad88919f1081\" (UID: \"f153e5dc-0977-40b7-9ed6-ad88919f1081\") " Dec 11 08:29:37 crc kubenswrapper[4860]: I1211 08:29:37.317051 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vwvc6\" (UniqueName: \"kubernetes.io/projected/f153e5dc-0977-40b7-9ed6-ad88919f1081-kube-api-access-vwvc6\") pod \"f153e5dc-0977-40b7-9ed6-ad88919f1081\" (UID: \"f153e5dc-0977-40b7-9ed6-ad88919f1081\") " Dec 11 08:29:37 crc kubenswrapper[4860]: I1211 08:29:37.317119 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f153e5dc-0977-40b7-9ed6-ad88919f1081-scripts\") pod \"f153e5dc-0977-40b7-9ed6-ad88919f1081\" (UID: \"f153e5dc-0977-40b7-9ed6-ad88919f1081\") " Dec 11 08:29:37 crc kubenswrapper[4860]: I1211 08:29:37.317208 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f153e5dc-0977-40b7-9ed6-ad88919f1081-httpd-run\") pod \"f153e5dc-0977-40b7-9ed6-ad88919f1081\" (UID: \"f153e5dc-0977-40b7-9ed6-ad88919f1081\") " Dec 11 08:29:37 crc kubenswrapper[4860]: I1211 08:29:37.317239 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f153e5dc-0977-40b7-9ed6-ad88919f1081-config-data\") pod \"f153e5dc-0977-40b7-9ed6-ad88919f1081\" (UID: \"f153e5dc-0977-40b7-9ed6-ad88919f1081\") " Dec 11 08:29:37 crc kubenswrapper[4860]: I1211 08:29:37.317286 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f153e5dc-0977-40b7-9ed6-ad88919f1081-combined-ca-bundle\") pod \"f153e5dc-0977-40b7-9ed6-ad88919f1081\" (UID: \"f153e5dc-0977-40b7-9ed6-ad88919f1081\") " Dec 11 08:29:37 crc kubenswrapper[4860]: I1211 08:29:37.318474 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f153e5dc-0977-40b7-9ed6-ad88919f1081-logs" (OuterVolumeSpecName: "logs") pod "f153e5dc-0977-40b7-9ed6-ad88919f1081" (UID: "f153e5dc-0977-40b7-9ed6-ad88919f1081"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:29:37 crc kubenswrapper[4860]: I1211 08:29:37.322714 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f153e5dc-0977-40b7-9ed6-ad88919f1081-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "f153e5dc-0977-40b7-9ed6-ad88919f1081" (UID: "f153e5dc-0977-40b7-9ed6-ad88919f1081"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:29:37 crc kubenswrapper[4860]: I1211 08:29:37.334343 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f153e5dc-0977-40b7-9ed6-ad88919f1081-scripts" (OuterVolumeSpecName: "scripts") pod "f153e5dc-0977-40b7-9ed6-ad88919f1081" (UID: "f153e5dc-0977-40b7-9ed6-ad88919f1081"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:29:37 crc kubenswrapper[4860]: I1211 08:29:37.334961 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "glance") pod "f153e5dc-0977-40b7-9ed6-ad88919f1081" (UID: "f153e5dc-0977-40b7-9ed6-ad88919f1081"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 11 08:29:37 crc kubenswrapper[4860]: I1211 08:29:37.338095 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f153e5dc-0977-40b7-9ed6-ad88919f1081-kube-api-access-vwvc6" (OuterVolumeSpecName: "kube-api-access-vwvc6") pod "f153e5dc-0977-40b7-9ed6-ad88919f1081" (UID: "f153e5dc-0977-40b7-9ed6-ad88919f1081"). InnerVolumeSpecName "kube-api-access-vwvc6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:29:37 crc kubenswrapper[4860]: I1211 08:29:37.422196 4860 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Dec 11 08:29:37 crc kubenswrapper[4860]: I1211 08:29:37.422526 4860 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f153e5dc-0977-40b7-9ed6-ad88919f1081-logs\") on node \"crc\" DevicePath \"\"" Dec 11 08:29:37 crc kubenswrapper[4860]: I1211 08:29:37.422539 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vwvc6\" (UniqueName: \"kubernetes.io/projected/f153e5dc-0977-40b7-9ed6-ad88919f1081-kube-api-access-vwvc6\") on node \"crc\" DevicePath \"\"" Dec 11 08:29:37 crc kubenswrapper[4860]: I1211 08:29:37.422554 4860 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f153e5dc-0977-40b7-9ed6-ad88919f1081-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 08:29:37 crc kubenswrapper[4860]: I1211 08:29:37.422565 4860 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f153e5dc-0977-40b7-9ed6-ad88919f1081-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 11 08:29:37 crc kubenswrapper[4860]: I1211 08:29:37.435929 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f153e5dc-0977-40b7-9ed6-ad88919f1081-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f153e5dc-0977-40b7-9ed6-ad88919f1081" (UID: "f153e5dc-0977-40b7-9ed6-ad88919f1081"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:29:37 crc kubenswrapper[4860]: I1211 08:29:37.473422 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-55ccdb6bbd-grdt6" Dec 11 08:29:37 crc kubenswrapper[4860]: I1211 08:29:37.496583 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-pfz4t" Dec 11 08:29:37 crc kubenswrapper[4860]: I1211 08:29:37.526254 4860 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f153e5dc-0977-40b7-9ed6-ad88919f1081-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 08:29:37 crc kubenswrapper[4860]: I1211 08:29:37.529837 4860 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Dec 11 08:29:37 crc kubenswrapper[4860]: I1211 08:29:37.601970 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f153e5dc-0977-40b7-9ed6-ad88919f1081-config-data" (OuterVolumeSpecName: "config-data") pod "f153e5dc-0977-40b7-9ed6-ad88919f1081" (UID: "f153e5dc-0977-40b7-9ed6-ad88919f1081"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:29:37 crc kubenswrapper[4860]: I1211 08:29:37.626974 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7pxv\" (UniqueName: \"kubernetes.io/projected/4ff61ef1-6a67-4a8e-80e2-11f0154e54c6-kube-api-access-x7pxv\") pod \"4ff61ef1-6a67-4a8e-80e2-11f0154e54c6\" (UID: \"4ff61ef1-6a67-4a8e-80e2-11f0154e54c6\") " Dec 11 08:29:37 crc kubenswrapper[4860]: I1211 08:29:37.627052 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ff61ef1-6a67-4a8e-80e2-11f0154e54c6-combined-ca-bundle\") pod \"4ff61ef1-6a67-4a8e-80e2-11f0154e54c6\" (UID: \"4ff61ef1-6a67-4a8e-80e2-11f0154e54c6\") " Dec 11 08:29:37 crc kubenswrapper[4860]: I1211 08:29:37.627164 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4ff61ef1-6a67-4a8e-80e2-11f0154e54c6-db-sync-config-data\") pod \"4ff61ef1-6a67-4a8e-80e2-11f0154e54c6\" (UID: \"4ff61ef1-6a67-4a8e-80e2-11f0154e54c6\") " Dec 11 08:29:37 crc kubenswrapper[4860]: I1211 08:29:37.627561 4860 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Dec 11 08:29:37 crc kubenswrapper[4860]: I1211 08:29:37.627572 4860 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f153e5dc-0977-40b7-9ed6-ad88919f1081-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 08:29:37 crc kubenswrapper[4860]: I1211 08:29:37.633494 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ff61ef1-6a67-4a8e-80e2-11f0154e54c6-kube-api-access-x7pxv" (OuterVolumeSpecName: "kube-api-access-x7pxv") pod "4ff61ef1-6a67-4a8e-80e2-11f0154e54c6" (UID: "4ff61ef1-6a67-4a8e-80e2-11f0154e54c6"). InnerVolumeSpecName "kube-api-access-x7pxv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:29:37 crc kubenswrapper[4860]: I1211 08:29:37.637751 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ff61ef1-6a67-4a8e-80e2-11f0154e54c6-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "4ff61ef1-6a67-4a8e-80e2-11f0154e54c6" (UID: "4ff61ef1-6a67-4a8e-80e2-11f0154e54c6"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:29:37 crc kubenswrapper[4860]: I1211 08:29:37.640364 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5fccb6555c-xwc9b"] Dec 11 08:29:37 crc kubenswrapper[4860]: I1211 08:29:37.677787 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ff61ef1-6a67-4a8e-80e2-11f0154e54c6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4ff61ef1-6a67-4a8e-80e2-11f0154e54c6" (UID: "4ff61ef1-6a67-4a8e-80e2-11f0154e54c6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:29:37 crc kubenswrapper[4860]: I1211 08:29:37.732017 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7pxv\" (UniqueName: \"kubernetes.io/projected/4ff61ef1-6a67-4a8e-80e2-11f0154e54c6-kube-api-access-x7pxv\") on node \"crc\" DevicePath \"\"" Dec 11 08:29:37 crc kubenswrapper[4860]: I1211 08:29:37.732341 4860 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ff61ef1-6a67-4a8e-80e2-11f0154e54c6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 08:29:37 crc kubenswrapper[4860]: I1211 08:29:37.732354 4860 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4ff61ef1-6a67-4a8e-80e2-11f0154e54c6-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 08:29:37 crc kubenswrapper[4860]: I1211 08:29:37.925862 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f153e5dc-0977-40b7-9ed6-ad88919f1081","Type":"ContainerDied","Data":"26481a7b9014dd8624ee0495fc60e935a037101148bb0ab94008cb2f17236c22"} Dec 11 08:29:37 crc kubenswrapper[4860]: I1211 08:29:37.925963 4860 scope.go:117] "RemoveContainer" containerID="d04c5444a6476bde178dcd80a442b4df937e8d9d28bf7bb2fee81b80764223de" Dec 11 08:29:37 crc kubenswrapper[4860]: I1211 08:29:37.927057 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 11 08:29:37 crc kubenswrapper[4860]: I1211 08:29:37.947155 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"67ac03d0-1fbe-46ff-b61d-7d15bcf48f61","Type":"ContainerDied","Data":"cf7f8eb5286c424bc2e307f364204e4fd8b87eebfdefe6d3d38b792b1a669680"} Dec 11 08:29:37 crc kubenswrapper[4860]: I1211 08:29:37.947201 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 11 08:29:37 crc kubenswrapper[4860]: I1211 08:29:37.956014 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5fccb6555c-xwc9b" event={"ID":"7739fee0-ccf5-49e9-bc21-4acd841daf55","Type":"ContainerStarted","Data":"048268d6ab7d0d33f8e8ec1254909544392614c4c2797be2c59d28c3e1405371"} Dec 11 08:29:37 crc kubenswrapper[4860]: I1211 08:29:37.985392 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 11 08:29:37 crc kubenswrapper[4860]: I1211 08:29:37.997306 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-pfz4t" Dec 11 08:29:37 crc kubenswrapper[4860]: I1211 08:29:37.999324 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-pfz4t" event={"ID":"4ff61ef1-6a67-4a8e-80e2-11f0154e54c6","Type":"ContainerDied","Data":"6ada87093bbc32624bf5ed6b56c88f8ae8afb66ca49ea4598f7c4115649d070a"} Dec 11 08:29:37 crc kubenswrapper[4860]: I1211 08:29:37.999364 4860 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6ada87093bbc32624bf5ed6b56c88f8ae8afb66ca49ea4598f7c4115649d070a" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.016926 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.033213 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.051003 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.063730 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 11 08:29:38 crc kubenswrapper[4860]: E1211 08:29:38.064243 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ff61ef1-6a67-4a8e-80e2-11f0154e54c6" containerName="barbican-db-sync" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.064261 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ff61ef1-6a67-4a8e-80e2-11f0154e54c6" containerName="barbican-db-sync" Dec 11 08:29:38 crc kubenswrapper[4860]: E1211 08:29:38.064274 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f153e5dc-0977-40b7-9ed6-ad88919f1081" containerName="glance-httpd" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.064283 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="f153e5dc-0977-40b7-9ed6-ad88919f1081" containerName="glance-httpd" Dec 11 08:29:38 crc kubenswrapper[4860]: E1211 08:29:38.064307 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f153e5dc-0977-40b7-9ed6-ad88919f1081" containerName="glance-log" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.064313 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="f153e5dc-0977-40b7-9ed6-ad88919f1081" containerName="glance-log" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.064501 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="f153e5dc-0977-40b7-9ed6-ad88919f1081" containerName="glance-log" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.064536 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="f153e5dc-0977-40b7-9ed6-ad88919f1081" containerName="glance-httpd" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.064550 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ff61ef1-6a67-4a8e-80e2-11f0154e54c6" containerName="barbican-db-sync" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.065581 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.074727 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.074969 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.075219 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.092162 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-9qb2r" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.096343 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.097941 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.107009 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.107215 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.138430 4860 scope.go:117] "RemoveContainer" containerID="fbe50653cd9fe78c494f42fa61f7c07b554d82034fe17558bd1e3ed88df56056" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.145787 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.205618 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.257149 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-568d665c75-4vptt"] Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.258888 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-568d665c75-4vptt" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.262214 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.263395 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3d65995-e403-4e92-b7aa-b39b8084d1bc-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"a3d65995-e403-4e92-b7aa-b39b8084d1bc\") " pod="openstack/glance-default-external-api-0" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.263459 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cwwwd\" (UniqueName: \"kubernetes.io/projected/83082bff-37e1-4ac1-9074-ae9d69ed66b0-kube-api-access-cwwwd\") pod \"glance-default-internal-api-0\" (UID: \"83082bff-37e1-4ac1-9074-ae9d69ed66b0\") " pod="openstack/glance-default-internal-api-0" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.263490 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a3d65995-e403-4e92-b7aa-b39b8084d1bc-scripts\") pod \"glance-default-external-api-0\" (UID: \"a3d65995-e403-4e92-b7aa-b39b8084d1bc\") " pod="openstack/glance-default-external-api-0" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.263505 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/83082bff-37e1-4ac1-9074-ae9d69ed66b0-logs\") pod \"glance-default-internal-api-0\" (UID: \"83082bff-37e1-4ac1-9074-ae9d69ed66b0\") " pod="openstack/glance-default-internal-api-0" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.263543 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a3d65995-e403-4e92-b7aa-b39b8084d1bc-logs\") pod \"glance-default-external-api-0\" (UID: \"a3d65995-e403-4e92-b7aa-b39b8084d1bc\") " pod="openstack/glance-default-external-api-0" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.263560 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"a3d65995-e403-4e92-b7aa-b39b8084d1bc\") " pod="openstack/glance-default-external-api-0" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.263585 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/83082bff-37e1-4ac1-9074-ae9d69ed66b0-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"83082bff-37e1-4ac1-9074-ae9d69ed66b0\") " pod="openstack/glance-default-internal-api-0" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.263604 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a3d65995-e403-4e92-b7aa-b39b8084d1bc-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"a3d65995-e403-4e92-b7aa-b39b8084d1bc\") " pod="openstack/glance-default-external-api-0" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.267174 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-xpbmn" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.267377 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.263636 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/83082bff-37e1-4ac1-9074-ae9d69ed66b0-scripts\") pod \"glance-default-internal-api-0\" (UID: \"83082bff-37e1-4ac1-9074-ae9d69ed66b0\") " pod="openstack/glance-default-internal-api-0" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.267848 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83082bff-37e1-4ac1-9074-ae9d69ed66b0-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"83082bff-37e1-4ac1-9074-ae9d69ed66b0\") " pod="openstack/glance-default-internal-api-0" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.267912 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dngvk\" (UniqueName: \"kubernetes.io/projected/a3d65995-e403-4e92-b7aa-b39b8084d1bc-kube-api-access-dngvk\") pod \"glance-default-external-api-0\" (UID: \"a3d65995-e403-4e92-b7aa-b39b8084d1bc\") " pod="openstack/glance-default-external-api-0" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.267928 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/83082bff-37e1-4ac1-9074-ae9d69ed66b0-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"83082bff-37e1-4ac1-9074-ae9d69ed66b0\") " pod="openstack/glance-default-internal-api-0" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.268008 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a3d65995-e403-4e92-b7aa-b39b8084d1bc-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"a3d65995-e403-4e92-b7aa-b39b8084d1bc\") " pod="openstack/glance-default-external-api-0" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.268040 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83082bff-37e1-4ac1-9074-ae9d69ed66b0-config-data\") pod \"glance-default-internal-api-0\" (UID: \"83082bff-37e1-4ac1-9074-ae9d69ed66b0\") " pod="openstack/glance-default-internal-api-0" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.268070 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3d65995-e403-4e92-b7aa-b39b8084d1bc-config-data\") pod \"glance-default-external-api-0\" (UID: \"a3d65995-e403-4e92-b7aa-b39b8084d1bc\") " pod="openstack/glance-default-external-api-0" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.268091 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"83082bff-37e1-4ac1-9074-ae9d69ed66b0\") " pod="openstack/glance-default-internal-api-0" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.277713 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-684c9ddd98-t29bd"] Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.279803 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-684c9ddd98-t29bd" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.285182 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.302301 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-684c9ddd98-t29bd"] Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.318674 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-55ccdb6bbd-grdt6"] Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.338931 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-568d665c75-4vptt"] Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.351220 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5ccc5c4795-65k7g"] Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.351515 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5ccc5c4795-65k7g" podUID="fe0e81d5-8772-4047-9b03-fd8127be2b66" containerName="dnsmasq-dns" containerID="cri-o://5c1825a4e2c7afa36121cf72f71a0143cd39e74617e9246b561d6909c66cc58d" gracePeriod=10 Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.369971 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a3d65995-e403-4e92-b7aa-b39b8084d1bc-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"a3d65995-e403-4e92-b7aa-b39b8084d1bc\") " pod="openstack/glance-default-external-api-0" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.370012 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83082bff-37e1-4ac1-9074-ae9d69ed66b0-config-data\") pod \"glance-default-internal-api-0\" (UID: \"83082bff-37e1-4ac1-9074-ae9d69ed66b0\") " pod="openstack/glance-default-internal-api-0" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.370036 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3d65995-e403-4e92-b7aa-b39b8084d1bc-config-data\") pod \"glance-default-external-api-0\" (UID: \"a3d65995-e403-4e92-b7aa-b39b8084d1bc\") " pod="openstack/glance-default-external-api-0" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.370054 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"83082bff-37e1-4ac1-9074-ae9d69ed66b0\") " pod="openstack/glance-default-internal-api-0" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.370087 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d8954df2-1f06-4bbe-bb33-a5ac8fdf9675-config-data-custom\") pod \"barbican-keystone-listener-684c9ddd98-t29bd\" (UID: \"d8954df2-1f06-4bbe-bb33-a5ac8fdf9675\") " pod="openstack/barbican-keystone-listener-684c9ddd98-t29bd" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.370112 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3d65995-e403-4e92-b7aa-b39b8084d1bc-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"a3d65995-e403-4e92-b7aa-b39b8084d1bc\") " pod="openstack/glance-default-external-api-0" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.370133 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8954df2-1f06-4bbe-bb33-a5ac8fdf9675-config-data\") pod \"barbican-keystone-listener-684c9ddd98-t29bd\" (UID: \"d8954df2-1f06-4bbe-bb33-a5ac8fdf9675\") " pod="openstack/barbican-keystone-listener-684c9ddd98-t29bd" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.370165 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l5l5d\" (UniqueName: \"kubernetes.io/projected/a7e1def2-f417-40f5-a7dc-c63148858073-kube-api-access-l5l5d\") pod \"barbican-worker-568d665c75-4vptt\" (UID: \"a7e1def2-f417-40f5-a7dc-c63148858073\") " pod="openstack/barbican-worker-568d665c75-4vptt" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.370191 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7e1def2-f417-40f5-a7dc-c63148858073-combined-ca-bundle\") pod \"barbican-worker-568d665c75-4vptt\" (UID: \"a7e1def2-f417-40f5-a7dc-c63148858073\") " pod="openstack/barbican-worker-568d665c75-4vptt" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.370219 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cwwwd\" (UniqueName: \"kubernetes.io/projected/83082bff-37e1-4ac1-9074-ae9d69ed66b0-kube-api-access-cwwwd\") pod \"glance-default-internal-api-0\" (UID: \"83082bff-37e1-4ac1-9074-ae9d69ed66b0\") " pod="openstack/glance-default-internal-api-0" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.370351 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a7e1def2-f417-40f5-a7dc-c63148858073-logs\") pod \"barbican-worker-568d665c75-4vptt\" (UID: \"a7e1def2-f417-40f5-a7dc-c63148858073\") " pod="openstack/barbican-worker-568d665c75-4vptt" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.370454 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a7e1def2-f417-40f5-a7dc-c63148858073-config-data-custom\") pod \"barbican-worker-568d665c75-4vptt\" (UID: \"a7e1def2-f417-40f5-a7dc-c63148858073\") " pod="openstack/barbican-worker-568d665c75-4vptt" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.370480 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a3d65995-e403-4e92-b7aa-b39b8084d1bc-scripts\") pod \"glance-default-external-api-0\" (UID: \"a3d65995-e403-4e92-b7aa-b39b8084d1bc\") " pod="openstack/glance-default-external-api-0" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.370530 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/83082bff-37e1-4ac1-9074-ae9d69ed66b0-logs\") pod \"glance-default-internal-api-0\" (UID: \"83082bff-37e1-4ac1-9074-ae9d69ed66b0\") " pod="openstack/glance-default-internal-api-0" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.370599 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8954df2-1f06-4bbe-bb33-a5ac8fdf9675-combined-ca-bundle\") pod \"barbican-keystone-listener-684c9ddd98-t29bd\" (UID: \"d8954df2-1f06-4bbe-bb33-a5ac8fdf9675\") " pod="openstack/barbican-keystone-listener-684c9ddd98-t29bd" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.370622 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7e1def2-f417-40f5-a7dc-c63148858073-config-data\") pod \"barbican-worker-568d665c75-4vptt\" (UID: \"a7e1def2-f417-40f5-a7dc-c63148858073\") " pod="openstack/barbican-worker-568d665c75-4vptt" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.370672 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d8954df2-1f06-4bbe-bb33-a5ac8fdf9675-logs\") pod \"barbican-keystone-listener-684c9ddd98-t29bd\" (UID: \"d8954df2-1f06-4bbe-bb33-a5ac8fdf9675\") " pod="openstack/barbican-keystone-listener-684c9ddd98-t29bd" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.370699 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a3d65995-e403-4e92-b7aa-b39b8084d1bc-logs\") pod \"glance-default-external-api-0\" (UID: \"a3d65995-e403-4e92-b7aa-b39b8084d1bc\") " pod="openstack/glance-default-external-api-0" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.370715 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"a3d65995-e403-4e92-b7aa-b39b8084d1bc\") " pod="openstack/glance-default-external-api-0" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.370789 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/83082bff-37e1-4ac1-9074-ae9d69ed66b0-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"83082bff-37e1-4ac1-9074-ae9d69ed66b0\") " pod="openstack/glance-default-internal-api-0" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.370846 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a3d65995-e403-4e92-b7aa-b39b8084d1bc-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"a3d65995-e403-4e92-b7aa-b39b8084d1bc\") " pod="openstack/glance-default-external-api-0" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.370875 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x2f7t\" (UniqueName: \"kubernetes.io/projected/d8954df2-1f06-4bbe-bb33-a5ac8fdf9675-kube-api-access-x2f7t\") pod \"barbican-keystone-listener-684c9ddd98-t29bd\" (UID: \"d8954df2-1f06-4bbe-bb33-a5ac8fdf9675\") " pod="openstack/barbican-keystone-listener-684c9ddd98-t29bd" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.371121 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/83082bff-37e1-4ac1-9074-ae9d69ed66b0-scripts\") pod \"glance-default-internal-api-0\" (UID: \"83082bff-37e1-4ac1-9074-ae9d69ed66b0\") " pod="openstack/glance-default-internal-api-0" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.371193 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83082bff-37e1-4ac1-9074-ae9d69ed66b0-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"83082bff-37e1-4ac1-9074-ae9d69ed66b0\") " pod="openstack/glance-default-internal-api-0" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.371252 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/83082bff-37e1-4ac1-9074-ae9d69ed66b0-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"83082bff-37e1-4ac1-9074-ae9d69ed66b0\") " pod="openstack/glance-default-internal-api-0" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.371435 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dngvk\" (UniqueName: \"kubernetes.io/projected/a3d65995-e403-4e92-b7aa-b39b8084d1bc-kube-api-access-dngvk\") pod \"glance-default-external-api-0\" (UID: \"a3d65995-e403-4e92-b7aa-b39b8084d1bc\") " pod="openstack/glance-default-external-api-0" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.372513 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-688c87cc99-sbps7"] Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.373454 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/83082bff-37e1-4ac1-9074-ae9d69ed66b0-logs\") pod \"glance-default-internal-api-0\" (UID: \"83082bff-37e1-4ac1-9074-ae9d69ed66b0\") " pod="openstack/glance-default-internal-api-0" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.373781 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/83082bff-37e1-4ac1-9074-ae9d69ed66b0-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"83082bff-37e1-4ac1-9074-ae9d69ed66b0\") " pod="openstack/glance-default-internal-api-0" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.374178 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a3d65995-e403-4e92-b7aa-b39b8084d1bc-logs\") pod \"glance-default-external-api-0\" (UID: \"a3d65995-e403-4e92-b7aa-b39b8084d1bc\") " pod="openstack/glance-default-external-api-0" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.374465 4860 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"a3d65995-e403-4e92-b7aa-b39b8084d1bc\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-external-api-0" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.375177 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-688c87cc99-sbps7" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.376180 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a3d65995-e403-4e92-b7aa-b39b8084d1bc-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"a3d65995-e403-4e92-b7aa-b39b8084d1bc\") " pod="openstack/glance-default-external-api-0" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.377258 4860 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"83082bff-37e1-4ac1-9074-ae9d69ed66b0\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/glance-default-internal-api-0" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.379861 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/83082bff-37e1-4ac1-9074-ae9d69ed66b0-scripts\") pod \"glance-default-internal-api-0\" (UID: \"83082bff-37e1-4ac1-9074-ae9d69ed66b0\") " pod="openstack/glance-default-internal-api-0" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.389190 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/83082bff-37e1-4ac1-9074-ae9d69ed66b0-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"83082bff-37e1-4ac1-9074-ae9d69ed66b0\") " pod="openstack/glance-default-internal-api-0" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.390857 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3d65995-e403-4e92-b7aa-b39b8084d1bc-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"a3d65995-e403-4e92-b7aa-b39b8084d1bc\") " pod="openstack/glance-default-external-api-0" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.393464 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83082bff-37e1-4ac1-9074-ae9d69ed66b0-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"83082bff-37e1-4ac1-9074-ae9d69ed66b0\") " pod="openstack/glance-default-internal-api-0" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.394781 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83082bff-37e1-4ac1-9074-ae9d69ed66b0-config-data\") pod \"glance-default-internal-api-0\" (UID: \"83082bff-37e1-4ac1-9074-ae9d69ed66b0\") " pod="openstack/glance-default-internal-api-0" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.395345 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a3d65995-e403-4e92-b7aa-b39b8084d1bc-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"a3d65995-e403-4e92-b7aa-b39b8084d1bc\") " pod="openstack/glance-default-external-api-0" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.404505 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dngvk\" (UniqueName: \"kubernetes.io/projected/a3d65995-e403-4e92-b7aa-b39b8084d1bc-kube-api-access-dngvk\") pod \"glance-default-external-api-0\" (UID: \"a3d65995-e403-4e92-b7aa-b39b8084d1bc\") " pod="openstack/glance-default-external-api-0" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.412448 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3d65995-e403-4e92-b7aa-b39b8084d1bc-config-data\") pod \"glance-default-external-api-0\" (UID: \"a3d65995-e403-4e92-b7aa-b39b8084d1bc\") " pod="openstack/glance-default-external-api-0" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.413463 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a3d65995-e403-4e92-b7aa-b39b8084d1bc-scripts\") pod \"glance-default-external-api-0\" (UID: \"a3d65995-e403-4e92-b7aa-b39b8084d1bc\") " pod="openstack/glance-default-external-api-0" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.424883 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cwwwd\" (UniqueName: \"kubernetes.io/projected/83082bff-37e1-4ac1-9074-ae9d69ed66b0-kube-api-access-cwwwd\") pod \"glance-default-internal-api-0\" (UID: \"83082bff-37e1-4ac1-9074-ae9d69ed66b0\") " pod="openstack/glance-default-internal-api-0" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.428922 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"a3d65995-e403-4e92-b7aa-b39b8084d1bc\") " pod="openstack/glance-default-external-api-0" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.437843 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"83082bff-37e1-4ac1-9074-ae9d69ed66b0\") " pod="openstack/glance-default-internal-api-0" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.447129 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.457034 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-688c87cc99-sbps7"] Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.476053 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5bc875db-b1de-4250-b9e7-bdd3a54bb08f-dns-swift-storage-0\") pod \"dnsmasq-dns-688c87cc99-sbps7\" (UID: \"5bc875db-b1de-4250-b9e7-bdd3a54bb08f\") " pod="openstack/dnsmasq-dns-688c87cc99-sbps7" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.476105 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5bc875db-b1de-4250-b9e7-bdd3a54bb08f-ovsdbserver-nb\") pod \"dnsmasq-dns-688c87cc99-sbps7\" (UID: \"5bc875db-b1de-4250-b9e7-bdd3a54bb08f\") " pod="openstack/dnsmasq-dns-688c87cc99-sbps7" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.476149 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5bc875db-b1de-4250-b9e7-bdd3a54bb08f-dns-svc\") pod \"dnsmasq-dns-688c87cc99-sbps7\" (UID: \"5bc875db-b1de-4250-b9e7-bdd3a54bb08f\") " pod="openstack/dnsmasq-dns-688c87cc99-sbps7" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.476269 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d8954df2-1f06-4bbe-bb33-a5ac8fdf9675-config-data-custom\") pod \"barbican-keystone-listener-684c9ddd98-t29bd\" (UID: \"d8954df2-1f06-4bbe-bb33-a5ac8fdf9675\") " pod="openstack/barbican-keystone-listener-684c9ddd98-t29bd" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.476309 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8954df2-1f06-4bbe-bb33-a5ac8fdf9675-config-data\") pod \"barbican-keystone-listener-684c9ddd98-t29bd\" (UID: \"d8954df2-1f06-4bbe-bb33-a5ac8fdf9675\") " pod="openstack/barbican-keystone-listener-684c9ddd98-t29bd" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.476340 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l5l5d\" (UniqueName: \"kubernetes.io/projected/a7e1def2-f417-40f5-a7dc-c63148858073-kube-api-access-l5l5d\") pod \"barbican-worker-568d665c75-4vptt\" (UID: \"a7e1def2-f417-40f5-a7dc-c63148858073\") " pod="openstack/barbican-worker-568d665c75-4vptt" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.476360 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5bc875db-b1de-4250-b9e7-bdd3a54bb08f-config\") pod \"dnsmasq-dns-688c87cc99-sbps7\" (UID: \"5bc875db-b1de-4250-b9e7-bdd3a54bb08f\") " pod="openstack/dnsmasq-dns-688c87cc99-sbps7" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.476389 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7e1def2-f417-40f5-a7dc-c63148858073-combined-ca-bundle\") pod \"barbican-worker-568d665c75-4vptt\" (UID: \"a7e1def2-f417-40f5-a7dc-c63148858073\") " pod="openstack/barbican-worker-568d665c75-4vptt" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.527627 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a7e1def2-f417-40f5-a7dc-c63148858073-logs\") pod \"barbican-worker-568d665c75-4vptt\" (UID: \"a7e1def2-f417-40f5-a7dc-c63148858073\") " pod="openstack/barbican-worker-568d665c75-4vptt" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.527696 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a7e1def2-f417-40f5-a7dc-c63148858073-config-data-custom\") pod \"barbican-worker-568d665c75-4vptt\" (UID: \"a7e1def2-f417-40f5-a7dc-c63148858073\") " pod="openstack/barbican-worker-568d665c75-4vptt" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.527748 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5bc875db-b1de-4250-b9e7-bdd3a54bb08f-ovsdbserver-sb\") pod \"dnsmasq-dns-688c87cc99-sbps7\" (UID: \"5bc875db-b1de-4250-b9e7-bdd3a54bb08f\") " pod="openstack/dnsmasq-dns-688c87cc99-sbps7" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.527802 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8954df2-1f06-4bbe-bb33-a5ac8fdf9675-combined-ca-bundle\") pod \"barbican-keystone-listener-684c9ddd98-t29bd\" (UID: \"d8954df2-1f06-4bbe-bb33-a5ac8fdf9675\") " pod="openstack/barbican-keystone-listener-684c9ddd98-t29bd" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.527846 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7e1def2-f417-40f5-a7dc-c63148858073-config-data\") pod \"barbican-worker-568d665c75-4vptt\" (UID: \"a7e1def2-f417-40f5-a7dc-c63148858073\") " pod="openstack/barbican-worker-568d665c75-4vptt" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.527875 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d8954df2-1f06-4bbe-bb33-a5ac8fdf9675-logs\") pod \"barbican-keystone-listener-684c9ddd98-t29bd\" (UID: \"d8954df2-1f06-4bbe-bb33-a5ac8fdf9675\") " pod="openstack/barbican-keystone-listener-684c9ddd98-t29bd" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.527951 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x2f7t\" (UniqueName: \"kubernetes.io/projected/d8954df2-1f06-4bbe-bb33-a5ac8fdf9675-kube-api-access-x2f7t\") pod \"barbican-keystone-listener-684c9ddd98-t29bd\" (UID: \"d8954df2-1f06-4bbe-bb33-a5ac8fdf9675\") " pod="openstack/barbican-keystone-listener-684c9ddd98-t29bd" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.527989 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2bp67\" (UniqueName: \"kubernetes.io/projected/5bc875db-b1de-4250-b9e7-bdd3a54bb08f-kube-api-access-2bp67\") pod \"dnsmasq-dns-688c87cc99-sbps7\" (UID: \"5bc875db-b1de-4250-b9e7-bdd3a54bb08f\") " pod="openstack/dnsmasq-dns-688c87cc99-sbps7" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.535416 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7e1def2-f417-40f5-a7dc-c63148858073-combined-ca-bundle\") pod \"barbican-worker-568d665c75-4vptt\" (UID: \"a7e1def2-f417-40f5-a7dc-c63148858073\") " pod="openstack/barbican-worker-568d665c75-4vptt" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.535957 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a7e1def2-f417-40f5-a7dc-c63148858073-logs\") pod \"barbican-worker-568d665c75-4vptt\" (UID: \"a7e1def2-f417-40f5-a7dc-c63148858073\") " pod="openstack/barbican-worker-568d665c75-4vptt" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.540603 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d8954df2-1f06-4bbe-bb33-a5ac8fdf9675-logs\") pod \"barbican-keystone-listener-684c9ddd98-t29bd\" (UID: \"d8954df2-1f06-4bbe-bb33-a5ac8fdf9675\") " pod="openstack/barbican-keystone-listener-684c9ddd98-t29bd" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.542423 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7e1def2-f417-40f5-a7dc-c63148858073-config-data\") pod \"barbican-worker-568d665c75-4vptt\" (UID: \"a7e1def2-f417-40f5-a7dc-c63148858073\") " pod="openstack/barbican-worker-568d665c75-4vptt" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.548314 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a7e1def2-f417-40f5-a7dc-c63148858073-config-data-custom\") pod \"barbican-worker-568d665c75-4vptt\" (UID: \"a7e1def2-f417-40f5-a7dc-c63148858073\") " pod="openstack/barbican-worker-568d665c75-4vptt" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.555726 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8954df2-1f06-4bbe-bb33-a5ac8fdf9675-combined-ca-bundle\") pod \"barbican-keystone-listener-684c9ddd98-t29bd\" (UID: \"d8954df2-1f06-4bbe-bb33-a5ac8fdf9675\") " pod="openstack/barbican-keystone-listener-684c9ddd98-t29bd" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.561310 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l5l5d\" (UniqueName: \"kubernetes.io/projected/a7e1def2-f417-40f5-a7dc-c63148858073-kube-api-access-l5l5d\") pod \"barbican-worker-568d665c75-4vptt\" (UID: \"a7e1def2-f417-40f5-a7dc-c63148858073\") " pod="openstack/barbican-worker-568d665c75-4vptt" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.572489 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d8954df2-1f06-4bbe-bb33-a5ac8fdf9675-config-data-custom\") pod \"barbican-keystone-listener-684c9ddd98-t29bd\" (UID: \"d8954df2-1f06-4bbe-bb33-a5ac8fdf9675\") " pod="openstack/barbican-keystone-listener-684c9ddd98-t29bd" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.575207 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8954df2-1f06-4bbe-bb33-a5ac8fdf9675-config-data\") pod \"barbican-keystone-listener-684c9ddd98-t29bd\" (UID: \"d8954df2-1f06-4bbe-bb33-a5ac8fdf9675\") " pod="openstack/barbican-keystone-listener-684c9ddd98-t29bd" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.609770 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x2f7t\" (UniqueName: \"kubernetes.io/projected/d8954df2-1f06-4bbe-bb33-a5ac8fdf9675-kube-api-access-x2f7t\") pod \"barbican-keystone-listener-684c9ddd98-t29bd\" (UID: \"d8954df2-1f06-4bbe-bb33-a5ac8fdf9675\") " pod="openstack/barbican-keystone-listener-684c9ddd98-t29bd" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.613711 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-9bc765b68-htbxs"] Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.616542 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-9bc765b68-htbxs" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.623879 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-9bc765b68-htbxs"] Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.627756 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.643297 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2bp67\" (UniqueName: \"kubernetes.io/projected/5bc875db-b1de-4250-b9e7-bdd3a54bb08f-kube-api-access-2bp67\") pod \"dnsmasq-dns-688c87cc99-sbps7\" (UID: \"5bc875db-b1de-4250-b9e7-bdd3a54bb08f\") " pod="openstack/dnsmasq-dns-688c87cc99-sbps7" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.643415 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5bc875db-b1de-4250-b9e7-bdd3a54bb08f-dns-swift-storage-0\") pod \"dnsmasq-dns-688c87cc99-sbps7\" (UID: \"5bc875db-b1de-4250-b9e7-bdd3a54bb08f\") " pod="openstack/dnsmasq-dns-688c87cc99-sbps7" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.643438 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5bc875db-b1de-4250-b9e7-bdd3a54bb08f-ovsdbserver-nb\") pod \"dnsmasq-dns-688c87cc99-sbps7\" (UID: \"5bc875db-b1de-4250-b9e7-bdd3a54bb08f\") " pod="openstack/dnsmasq-dns-688c87cc99-sbps7" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.643526 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5bc875db-b1de-4250-b9e7-bdd3a54bb08f-dns-svc\") pod \"dnsmasq-dns-688c87cc99-sbps7\" (UID: \"5bc875db-b1de-4250-b9e7-bdd3a54bb08f\") " pod="openstack/dnsmasq-dns-688c87cc99-sbps7" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.643613 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5bc875db-b1de-4250-b9e7-bdd3a54bb08f-config\") pod \"dnsmasq-dns-688c87cc99-sbps7\" (UID: \"5bc875db-b1de-4250-b9e7-bdd3a54bb08f\") " pod="openstack/dnsmasq-dns-688c87cc99-sbps7" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.643795 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5bc875db-b1de-4250-b9e7-bdd3a54bb08f-ovsdbserver-sb\") pod \"dnsmasq-dns-688c87cc99-sbps7\" (UID: \"5bc875db-b1de-4250-b9e7-bdd3a54bb08f\") " pod="openstack/dnsmasq-dns-688c87cc99-sbps7" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.647951 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5bc875db-b1de-4250-b9e7-bdd3a54bb08f-dns-swift-storage-0\") pod \"dnsmasq-dns-688c87cc99-sbps7\" (UID: \"5bc875db-b1de-4250-b9e7-bdd3a54bb08f\") " pod="openstack/dnsmasq-dns-688c87cc99-sbps7" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.648531 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5bc875db-b1de-4250-b9e7-bdd3a54bb08f-config\") pod \"dnsmasq-dns-688c87cc99-sbps7\" (UID: \"5bc875db-b1de-4250-b9e7-bdd3a54bb08f\") " pod="openstack/dnsmasq-dns-688c87cc99-sbps7" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.649491 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5bc875db-b1de-4250-b9e7-bdd3a54bb08f-dns-svc\") pod \"dnsmasq-dns-688c87cc99-sbps7\" (UID: \"5bc875db-b1de-4250-b9e7-bdd3a54bb08f\") " pod="openstack/dnsmasq-dns-688c87cc99-sbps7" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.650391 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5bc875db-b1de-4250-b9e7-bdd3a54bb08f-ovsdbserver-sb\") pod \"dnsmasq-dns-688c87cc99-sbps7\" (UID: \"5bc875db-b1de-4250-b9e7-bdd3a54bb08f\") " pod="openstack/dnsmasq-dns-688c87cc99-sbps7" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.651086 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5bc875db-b1de-4250-b9e7-bdd3a54bb08f-ovsdbserver-nb\") pod \"dnsmasq-dns-688c87cc99-sbps7\" (UID: \"5bc875db-b1de-4250-b9e7-bdd3a54bb08f\") " pod="openstack/dnsmasq-dns-688c87cc99-sbps7" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.687472 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-568d665c75-4vptt" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.696002 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.739343 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-684c9ddd98-t29bd" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.749298 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e28c0dc-a169-423c-ae52-4392a9a27a3e-combined-ca-bundle\") pod \"barbican-api-9bc765b68-htbxs\" (UID: \"0e28c0dc-a169-423c-ae52-4392a9a27a3e\") " pod="openstack/barbican-api-9bc765b68-htbxs" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.749362 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0e28c0dc-a169-423c-ae52-4392a9a27a3e-logs\") pod \"barbican-api-9bc765b68-htbxs\" (UID: \"0e28c0dc-a169-423c-ae52-4392a9a27a3e\") " pod="openstack/barbican-api-9bc765b68-htbxs" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.749543 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mxrqv\" (UniqueName: \"kubernetes.io/projected/0e28c0dc-a169-423c-ae52-4392a9a27a3e-kube-api-access-mxrqv\") pod \"barbican-api-9bc765b68-htbxs\" (UID: \"0e28c0dc-a169-423c-ae52-4392a9a27a3e\") " pod="openstack/barbican-api-9bc765b68-htbxs" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.749599 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0e28c0dc-a169-423c-ae52-4392a9a27a3e-config-data-custom\") pod \"barbican-api-9bc765b68-htbxs\" (UID: \"0e28c0dc-a169-423c-ae52-4392a9a27a3e\") " pod="openstack/barbican-api-9bc765b68-htbxs" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.749749 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e28c0dc-a169-423c-ae52-4392a9a27a3e-config-data\") pod \"barbican-api-9bc765b68-htbxs\" (UID: \"0e28c0dc-a169-423c-ae52-4392a9a27a3e\") " pod="openstack/barbican-api-9bc765b68-htbxs" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.751776 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2bp67\" (UniqueName: \"kubernetes.io/projected/5bc875db-b1de-4250-b9e7-bdd3a54bb08f-kube-api-access-2bp67\") pod \"dnsmasq-dns-688c87cc99-sbps7\" (UID: \"5bc875db-b1de-4250-b9e7-bdd3a54bb08f\") " pod="openstack/dnsmasq-dns-688c87cc99-sbps7" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.846556 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-plxxc" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.851660 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0e28c0dc-a169-423c-ae52-4392a9a27a3e-config-data-custom\") pod \"barbican-api-9bc765b68-htbxs\" (UID: \"0e28c0dc-a169-423c-ae52-4392a9a27a3e\") " pod="openstack/barbican-api-9bc765b68-htbxs" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.851764 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e28c0dc-a169-423c-ae52-4392a9a27a3e-config-data\") pod \"barbican-api-9bc765b68-htbxs\" (UID: \"0e28c0dc-a169-423c-ae52-4392a9a27a3e\") " pod="openstack/barbican-api-9bc765b68-htbxs" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.851814 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e28c0dc-a169-423c-ae52-4392a9a27a3e-combined-ca-bundle\") pod \"barbican-api-9bc765b68-htbxs\" (UID: \"0e28c0dc-a169-423c-ae52-4392a9a27a3e\") " pod="openstack/barbican-api-9bc765b68-htbxs" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.851839 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0e28c0dc-a169-423c-ae52-4392a9a27a3e-logs\") pod \"barbican-api-9bc765b68-htbxs\" (UID: \"0e28c0dc-a169-423c-ae52-4392a9a27a3e\") " pod="openstack/barbican-api-9bc765b68-htbxs" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.851914 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mxrqv\" (UniqueName: \"kubernetes.io/projected/0e28c0dc-a169-423c-ae52-4392a9a27a3e-kube-api-access-mxrqv\") pod \"barbican-api-9bc765b68-htbxs\" (UID: \"0e28c0dc-a169-423c-ae52-4392a9a27a3e\") " pod="openstack/barbican-api-9bc765b68-htbxs" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.855167 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0e28c0dc-a169-423c-ae52-4392a9a27a3e-logs\") pod \"barbican-api-9bc765b68-htbxs\" (UID: \"0e28c0dc-a169-423c-ae52-4392a9a27a3e\") " pod="openstack/barbican-api-9bc765b68-htbxs" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.861271 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0e28c0dc-a169-423c-ae52-4392a9a27a3e-config-data-custom\") pod \"barbican-api-9bc765b68-htbxs\" (UID: \"0e28c0dc-a169-423c-ae52-4392a9a27a3e\") " pod="openstack/barbican-api-9bc765b68-htbxs" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.881594 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e28c0dc-a169-423c-ae52-4392a9a27a3e-combined-ca-bundle\") pod \"barbican-api-9bc765b68-htbxs\" (UID: \"0e28c0dc-a169-423c-ae52-4392a9a27a3e\") " pod="openstack/barbican-api-9bc765b68-htbxs" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.882994 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e28c0dc-a169-423c-ae52-4392a9a27a3e-config-data\") pod \"barbican-api-9bc765b68-htbxs\" (UID: \"0e28c0dc-a169-423c-ae52-4392a9a27a3e\") " pod="openstack/barbican-api-9bc765b68-htbxs" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.902351 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mxrqv\" (UniqueName: \"kubernetes.io/projected/0e28c0dc-a169-423c-ae52-4392a9a27a3e-kube-api-access-mxrqv\") pod \"barbican-api-9bc765b68-htbxs\" (UID: \"0e28c0dc-a169-423c-ae52-4392a9a27a3e\") " pod="openstack/barbican-api-9bc765b68-htbxs" Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.958359 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/fecaee75-2dc0-4d48-9d40-11fc1ec8e972-credential-keys\") pod \"fecaee75-2dc0-4d48-9d40-11fc1ec8e972\" (UID: \"fecaee75-2dc0-4d48-9d40-11fc1ec8e972\") " Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.958701 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fecaee75-2dc0-4d48-9d40-11fc1ec8e972-scripts\") pod \"fecaee75-2dc0-4d48-9d40-11fc1ec8e972\" (UID: \"fecaee75-2dc0-4d48-9d40-11fc1ec8e972\") " Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.958732 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/fecaee75-2dc0-4d48-9d40-11fc1ec8e972-fernet-keys\") pod \"fecaee75-2dc0-4d48-9d40-11fc1ec8e972\" (UID: \"fecaee75-2dc0-4d48-9d40-11fc1ec8e972\") " Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.958847 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fecaee75-2dc0-4d48-9d40-11fc1ec8e972-config-data\") pod \"fecaee75-2dc0-4d48-9d40-11fc1ec8e972\" (UID: \"fecaee75-2dc0-4d48-9d40-11fc1ec8e972\") " Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.958873 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fecaee75-2dc0-4d48-9d40-11fc1ec8e972-combined-ca-bundle\") pod \"fecaee75-2dc0-4d48-9d40-11fc1ec8e972\" (UID: \"fecaee75-2dc0-4d48-9d40-11fc1ec8e972\") " Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.958982 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xzhjl\" (UniqueName: \"kubernetes.io/projected/fecaee75-2dc0-4d48-9d40-11fc1ec8e972-kube-api-access-xzhjl\") pod \"fecaee75-2dc0-4d48-9d40-11fc1ec8e972\" (UID: \"fecaee75-2dc0-4d48-9d40-11fc1ec8e972\") " Dec 11 08:29:38 crc kubenswrapper[4860]: I1211 08:29:38.992195 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fecaee75-2dc0-4d48-9d40-11fc1ec8e972-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "fecaee75-2dc0-4d48-9d40-11fc1ec8e972" (UID: "fecaee75-2dc0-4d48-9d40-11fc1ec8e972"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:29:39 crc kubenswrapper[4860]: I1211 08:29:39.017003 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fecaee75-2dc0-4d48-9d40-11fc1ec8e972-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "fecaee75-2dc0-4d48-9d40-11fc1ec8e972" (UID: "fecaee75-2dc0-4d48-9d40-11fc1ec8e972"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:29:39 crc kubenswrapper[4860]: I1211 08:29:39.017110 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fecaee75-2dc0-4d48-9d40-11fc1ec8e972-scripts" (OuterVolumeSpecName: "scripts") pod "fecaee75-2dc0-4d48-9d40-11fc1ec8e972" (UID: "fecaee75-2dc0-4d48-9d40-11fc1ec8e972"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:29:39 crc kubenswrapper[4860]: I1211 08:29:39.018479 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fecaee75-2dc0-4d48-9d40-11fc1ec8e972-kube-api-access-xzhjl" (OuterVolumeSpecName: "kube-api-access-xzhjl") pod "fecaee75-2dc0-4d48-9d40-11fc1ec8e972" (UID: "fecaee75-2dc0-4d48-9d40-11fc1ec8e972"). InnerVolumeSpecName "kube-api-access-xzhjl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:29:39 crc kubenswrapper[4860]: I1211 08:29:39.055317 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fecaee75-2dc0-4d48-9d40-11fc1ec8e972-config-data" (OuterVolumeSpecName: "config-data") pod "fecaee75-2dc0-4d48-9d40-11fc1ec8e972" (UID: "fecaee75-2dc0-4d48-9d40-11fc1ec8e972"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:29:39 crc kubenswrapper[4860]: I1211 08:29:39.063305 4860 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fecaee75-2dc0-4d48-9d40-11fc1ec8e972-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 08:29:39 crc kubenswrapper[4860]: I1211 08:29:39.063340 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xzhjl\" (UniqueName: \"kubernetes.io/projected/fecaee75-2dc0-4d48-9d40-11fc1ec8e972-kube-api-access-xzhjl\") on node \"crc\" DevicePath \"\"" Dec 11 08:29:39 crc kubenswrapper[4860]: I1211 08:29:39.063355 4860 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/fecaee75-2dc0-4d48-9d40-11fc1ec8e972-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 11 08:29:39 crc kubenswrapper[4860]: I1211 08:29:39.063364 4860 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fecaee75-2dc0-4d48-9d40-11fc1ec8e972-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 08:29:39 crc kubenswrapper[4860]: I1211 08:29:39.063373 4860 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/fecaee75-2dc0-4d48-9d40-11fc1ec8e972-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 11 08:29:39 crc kubenswrapper[4860]: I1211 08:29:39.063531 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-688c87cc99-sbps7" Dec 11 08:29:39 crc kubenswrapper[4860]: I1211 08:29:39.072412 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-9bc765b68-htbxs" Dec 11 08:29:39 crc kubenswrapper[4860]: I1211 08:29:39.085701 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fecaee75-2dc0-4d48-9d40-11fc1ec8e972-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fecaee75-2dc0-4d48-9d40-11fc1ec8e972" (UID: "fecaee75-2dc0-4d48-9d40-11fc1ec8e972"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:29:39 crc kubenswrapper[4860]: I1211 08:29:39.168683 4860 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fecaee75-2dc0-4d48-9d40-11fc1ec8e972-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 08:29:39 crc kubenswrapper[4860]: I1211 08:29:39.184431 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-748d5bd978-4sz6v"] Dec 11 08:29:39 crc kubenswrapper[4860]: E1211 08:29:39.188455 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fecaee75-2dc0-4d48-9d40-11fc1ec8e972" containerName="keystone-bootstrap" Dec 11 08:29:39 crc kubenswrapper[4860]: I1211 08:29:39.188487 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="fecaee75-2dc0-4d48-9d40-11fc1ec8e972" containerName="keystone-bootstrap" Dec 11 08:29:39 crc kubenswrapper[4860]: I1211 08:29:39.188857 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="fecaee75-2dc0-4d48-9d40-11fc1ec8e972" containerName="keystone-bootstrap" Dec 11 08:29:39 crc kubenswrapper[4860]: I1211 08:29:39.190518 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-748d5bd978-4sz6v" Dec 11 08:29:39 crc kubenswrapper[4860]: I1211 08:29:39.198190 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Dec 11 08:29:39 crc kubenswrapper[4860]: I1211 08:29:39.198568 4860 generic.go:334] "Generic (PLEG): container finished" podID="fe0e81d5-8772-4047-9b03-fd8127be2b66" containerID="5c1825a4e2c7afa36121cf72f71a0143cd39e74617e9246b561d6909c66cc58d" exitCode=0 Dec 11 08:29:39 crc kubenswrapper[4860]: I1211 08:29:39.198630 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc5c4795-65k7g" event={"ID":"fe0e81d5-8772-4047-9b03-fd8127be2b66","Type":"ContainerDied","Data":"5c1825a4e2c7afa36121cf72f71a0143cd39e74617e9246b561d6909c66cc58d"} Dec 11 08:29:39 crc kubenswrapper[4860]: I1211 08:29:39.202613 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-748d5bd978-4sz6v"] Dec 11 08:29:39 crc kubenswrapper[4860]: I1211 08:29:39.209530 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Dec 11 08:29:39 crc kubenswrapper[4860]: I1211 08:29:39.214819 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-55ccdb6bbd-grdt6" event={"ID":"78683923-6a60-4a9a-bb81-51ae61f55376","Type":"ContainerStarted","Data":"a81ea6712d7f2f980fea83db838212db28779cdf659468b8cb46e6498f403f33"} Dec 11 08:29:39 crc kubenswrapper[4860]: I1211 08:29:39.218223 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5fccb6555c-xwc9b" event={"ID":"7739fee0-ccf5-49e9-bc21-4acd841daf55","Type":"ContainerStarted","Data":"e26010f8c1a00ef85ba94e0dc3534a9aeedc991f927175fa96e54a8c3a083a8e"} Dec 11 08:29:39 crc kubenswrapper[4860]: I1211 08:29:39.218247 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5fccb6555c-xwc9b" event={"ID":"7739fee0-ccf5-49e9-bc21-4acd841daf55","Type":"ContainerStarted","Data":"33548b189ef3656d65a7211ca07bcf11f0990906378767d0864fe0991fe2bbbd"} Dec 11 08:29:39 crc kubenswrapper[4860]: I1211 08:29:39.218959 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-5fccb6555c-xwc9b" Dec 11 08:29:39 crc kubenswrapper[4860]: I1211 08:29:39.221862 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-plxxc" event={"ID":"fecaee75-2dc0-4d48-9d40-11fc1ec8e972","Type":"ContainerDied","Data":"135ecdf1d9cab450a5e7bbb463cf46c95697996cf9020c6c3f85dd181f8d70c8"} Dec 11 08:29:39 crc kubenswrapper[4860]: I1211 08:29:39.221884 4860 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="135ecdf1d9cab450a5e7bbb463cf46c95697996cf9020c6c3f85dd181f8d70c8" Dec 11 08:29:39 crc kubenswrapper[4860]: I1211 08:29:39.221974 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-plxxc" Dec 11 08:29:39 crc kubenswrapper[4860]: I1211 08:29:39.270941 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-5fccb6555c-xwc9b" podStartSLOduration=3.270919317 podStartE2EDuration="3.270919317s" podCreationTimestamp="2025-12-11 08:29:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:29:39.259271272 +0000 UTC m=+1111.987790327" watchObservedRunningTime="2025-12-11 08:29:39.270919317 +0000 UTC m=+1111.999438372" Dec 11 08:29:39 crc kubenswrapper[4860]: I1211 08:29:39.281052 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e208d74e-f50f-443c-aa2d-eccdb584413d-combined-ca-bundle\") pod \"keystone-748d5bd978-4sz6v\" (UID: \"e208d74e-f50f-443c-aa2d-eccdb584413d\") " pod="openstack/keystone-748d5bd978-4sz6v" Dec 11 08:29:39 crc kubenswrapper[4860]: I1211 08:29:39.297743 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lvmrl\" (UniqueName: \"kubernetes.io/projected/e208d74e-f50f-443c-aa2d-eccdb584413d-kube-api-access-lvmrl\") pod \"keystone-748d5bd978-4sz6v\" (UID: \"e208d74e-f50f-443c-aa2d-eccdb584413d\") " pod="openstack/keystone-748d5bd978-4sz6v" Dec 11 08:29:39 crc kubenswrapper[4860]: I1211 08:29:39.298009 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e208d74e-f50f-443c-aa2d-eccdb584413d-config-data\") pod \"keystone-748d5bd978-4sz6v\" (UID: \"e208d74e-f50f-443c-aa2d-eccdb584413d\") " pod="openstack/keystone-748d5bd978-4sz6v" Dec 11 08:29:39 crc kubenswrapper[4860]: I1211 08:29:39.298038 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e208d74e-f50f-443c-aa2d-eccdb584413d-internal-tls-certs\") pod \"keystone-748d5bd978-4sz6v\" (UID: \"e208d74e-f50f-443c-aa2d-eccdb584413d\") " pod="openstack/keystone-748d5bd978-4sz6v" Dec 11 08:29:39 crc kubenswrapper[4860]: I1211 08:29:39.298095 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e208d74e-f50f-443c-aa2d-eccdb584413d-fernet-keys\") pod \"keystone-748d5bd978-4sz6v\" (UID: \"e208d74e-f50f-443c-aa2d-eccdb584413d\") " pod="openstack/keystone-748d5bd978-4sz6v" Dec 11 08:29:39 crc kubenswrapper[4860]: I1211 08:29:39.298126 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e208d74e-f50f-443c-aa2d-eccdb584413d-scripts\") pod \"keystone-748d5bd978-4sz6v\" (UID: \"e208d74e-f50f-443c-aa2d-eccdb584413d\") " pod="openstack/keystone-748d5bd978-4sz6v" Dec 11 08:29:39 crc kubenswrapper[4860]: I1211 08:29:39.298157 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/e208d74e-f50f-443c-aa2d-eccdb584413d-credential-keys\") pod \"keystone-748d5bd978-4sz6v\" (UID: \"e208d74e-f50f-443c-aa2d-eccdb584413d\") " pod="openstack/keystone-748d5bd978-4sz6v" Dec 11 08:29:39 crc kubenswrapper[4860]: I1211 08:29:39.298195 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e208d74e-f50f-443c-aa2d-eccdb584413d-public-tls-certs\") pod \"keystone-748d5bd978-4sz6v\" (UID: \"e208d74e-f50f-443c-aa2d-eccdb584413d\") " pod="openstack/keystone-748d5bd978-4sz6v" Dec 11 08:29:39 crc kubenswrapper[4860]: I1211 08:29:39.317083 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc5c4795-65k7g" Dec 11 08:29:39 crc kubenswrapper[4860]: I1211 08:29:39.401901 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e208d74e-f50f-443c-aa2d-eccdb584413d-internal-tls-certs\") pod \"keystone-748d5bd978-4sz6v\" (UID: \"e208d74e-f50f-443c-aa2d-eccdb584413d\") " pod="openstack/keystone-748d5bd978-4sz6v" Dec 11 08:29:39 crc kubenswrapper[4860]: I1211 08:29:39.401974 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e208d74e-f50f-443c-aa2d-eccdb584413d-fernet-keys\") pod \"keystone-748d5bd978-4sz6v\" (UID: \"e208d74e-f50f-443c-aa2d-eccdb584413d\") " pod="openstack/keystone-748d5bd978-4sz6v" Dec 11 08:29:39 crc kubenswrapper[4860]: I1211 08:29:39.402000 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e208d74e-f50f-443c-aa2d-eccdb584413d-scripts\") pod \"keystone-748d5bd978-4sz6v\" (UID: \"e208d74e-f50f-443c-aa2d-eccdb584413d\") " pod="openstack/keystone-748d5bd978-4sz6v" Dec 11 08:29:39 crc kubenswrapper[4860]: I1211 08:29:39.402027 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/e208d74e-f50f-443c-aa2d-eccdb584413d-credential-keys\") pod \"keystone-748d5bd978-4sz6v\" (UID: \"e208d74e-f50f-443c-aa2d-eccdb584413d\") " pod="openstack/keystone-748d5bd978-4sz6v" Dec 11 08:29:39 crc kubenswrapper[4860]: I1211 08:29:39.402054 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e208d74e-f50f-443c-aa2d-eccdb584413d-public-tls-certs\") pod \"keystone-748d5bd978-4sz6v\" (UID: \"e208d74e-f50f-443c-aa2d-eccdb584413d\") " pod="openstack/keystone-748d5bd978-4sz6v" Dec 11 08:29:39 crc kubenswrapper[4860]: I1211 08:29:39.402140 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e208d74e-f50f-443c-aa2d-eccdb584413d-combined-ca-bundle\") pod \"keystone-748d5bd978-4sz6v\" (UID: \"e208d74e-f50f-443c-aa2d-eccdb584413d\") " pod="openstack/keystone-748d5bd978-4sz6v" Dec 11 08:29:39 crc kubenswrapper[4860]: I1211 08:29:39.402166 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lvmrl\" (UniqueName: \"kubernetes.io/projected/e208d74e-f50f-443c-aa2d-eccdb584413d-kube-api-access-lvmrl\") pod \"keystone-748d5bd978-4sz6v\" (UID: \"e208d74e-f50f-443c-aa2d-eccdb584413d\") " pod="openstack/keystone-748d5bd978-4sz6v" Dec 11 08:29:39 crc kubenswrapper[4860]: I1211 08:29:39.402272 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e208d74e-f50f-443c-aa2d-eccdb584413d-config-data\") pod \"keystone-748d5bd978-4sz6v\" (UID: \"e208d74e-f50f-443c-aa2d-eccdb584413d\") " pod="openstack/keystone-748d5bd978-4sz6v" Dec 11 08:29:39 crc kubenswrapper[4860]: I1211 08:29:39.417405 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e208d74e-f50f-443c-aa2d-eccdb584413d-internal-tls-certs\") pod \"keystone-748d5bd978-4sz6v\" (UID: \"e208d74e-f50f-443c-aa2d-eccdb584413d\") " pod="openstack/keystone-748d5bd978-4sz6v" Dec 11 08:29:39 crc kubenswrapper[4860]: I1211 08:29:39.422667 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e208d74e-f50f-443c-aa2d-eccdb584413d-public-tls-certs\") pod \"keystone-748d5bd978-4sz6v\" (UID: \"e208d74e-f50f-443c-aa2d-eccdb584413d\") " pod="openstack/keystone-748d5bd978-4sz6v" Dec 11 08:29:39 crc kubenswrapper[4860]: I1211 08:29:39.423270 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/e208d74e-f50f-443c-aa2d-eccdb584413d-credential-keys\") pod \"keystone-748d5bd978-4sz6v\" (UID: \"e208d74e-f50f-443c-aa2d-eccdb584413d\") " pod="openstack/keystone-748d5bd978-4sz6v" Dec 11 08:29:39 crc kubenswrapper[4860]: I1211 08:29:39.433789 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e208d74e-f50f-443c-aa2d-eccdb584413d-fernet-keys\") pod \"keystone-748d5bd978-4sz6v\" (UID: \"e208d74e-f50f-443c-aa2d-eccdb584413d\") " pod="openstack/keystone-748d5bd978-4sz6v" Dec 11 08:29:39 crc kubenswrapper[4860]: I1211 08:29:39.434824 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e208d74e-f50f-443c-aa2d-eccdb584413d-config-data\") pod \"keystone-748d5bd978-4sz6v\" (UID: \"e208d74e-f50f-443c-aa2d-eccdb584413d\") " pod="openstack/keystone-748d5bd978-4sz6v" Dec 11 08:29:39 crc kubenswrapper[4860]: I1211 08:29:39.458399 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lvmrl\" (UniqueName: \"kubernetes.io/projected/e208d74e-f50f-443c-aa2d-eccdb584413d-kube-api-access-lvmrl\") pod \"keystone-748d5bd978-4sz6v\" (UID: \"e208d74e-f50f-443c-aa2d-eccdb584413d\") " pod="openstack/keystone-748d5bd978-4sz6v" Dec 11 08:29:39 crc kubenswrapper[4860]: I1211 08:29:39.467379 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e208d74e-f50f-443c-aa2d-eccdb584413d-scripts\") pod \"keystone-748d5bd978-4sz6v\" (UID: \"e208d74e-f50f-443c-aa2d-eccdb584413d\") " pod="openstack/keystone-748d5bd978-4sz6v" Dec 11 08:29:39 crc kubenswrapper[4860]: I1211 08:29:39.467980 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e208d74e-f50f-443c-aa2d-eccdb584413d-combined-ca-bundle\") pod \"keystone-748d5bd978-4sz6v\" (UID: \"e208d74e-f50f-443c-aa2d-eccdb584413d\") " pod="openstack/keystone-748d5bd978-4sz6v" Dec 11 08:29:39 crc kubenswrapper[4860]: I1211 08:29:39.526385 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hg6tn\" (UniqueName: \"kubernetes.io/projected/fe0e81d5-8772-4047-9b03-fd8127be2b66-kube-api-access-hg6tn\") pod \"fe0e81d5-8772-4047-9b03-fd8127be2b66\" (UID: \"fe0e81d5-8772-4047-9b03-fd8127be2b66\") " Dec 11 08:29:39 crc kubenswrapper[4860]: I1211 08:29:39.526466 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fe0e81d5-8772-4047-9b03-fd8127be2b66-ovsdbserver-nb\") pod \"fe0e81d5-8772-4047-9b03-fd8127be2b66\" (UID: \"fe0e81d5-8772-4047-9b03-fd8127be2b66\") " Dec 11 08:29:39 crc kubenswrapper[4860]: I1211 08:29:39.526528 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fe0e81d5-8772-4047-9b03-fd8127be2b66-dns-svc\") pod \"fe0e81d5-8772-4047-9b03-fd8127be2b66\" (UID: \"fe0e81d5-8772-4047-9b03-fd8127be2b66\") " Dec 11 08:29:39 crc kubenswrapper[4860]: I1211 08:29:39.526556 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fe0e81d5-8772-4047-9b03-fd8127be2b66-config\") pod \"fe0e81d5-8772-4047-9b03-fd8127be2b66\" (UID: \"fe0e81d5-8772-4047-9b03-fd8127be2b66\") " Dec 11 08:29:39 crc kubenswrapper[4860]: I1211 08:29:39.526623 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fe0e81d5-8772-4047-9b03-fd8127be2b66-dns-swift-storage-0\") pod \"fe0e81d5-8772-4047-9b03-fd8127be2b66\" (UID: \"fe0e81d5-8772-4047-9b03-fd8127be2b66\") " Dec 11 08:29:39 crc kubenswrapper[4860]: I1211 08:29:39.526746 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fe0e81d5-8772-4047-9b03-fd8127be2b66-ovsdbserver-sb\") pod \"fe0e81d5-8772-4047-9b03-fd8127be2b66\" (UID: \"fe0e81d5-8772-4047-9b03-fd8127be2b66\") " Dec 11 08:29:39 crc kubenswrapper[4860]: I1211 08:29:39.592420 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-748d5bd978-4sz6v" Dec 11 08:29:39 crc kubenswrapper[4860]: I1211 08:29:39.599892 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fe0e81d5-8772-4047-9b03-fd8127be2b66-kube-api-access-hg6tn" (OuterVolumeSpecName: "kube-api-access-hg6tn") pod "fe0e81d5-8772-4047-9b03-fd8127be2b66" (UID: "fe0e81d5-8772-4047-9b03-fd8127be2b66"). InnerVolumeSpecName "kube-api-access-hg6tn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:29:39 crc kubenswrapper[4860]: I1211 08:29:39.629198 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hg6tn\" (UniqueName: \"kubernetes.io/projected/fe0e81d5-8772-4047-9b03-fd8127be2b66-kube-api-access-hg6tn\") on node \"crc\" DevicePath \"\"" Dec 11 08:29:39 crc kubenswrapper[4860]: I1211 08:29:39.647836 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="67ac03d0-1fbe-46ff-b61d-7d15bcf48f61" path="/var/lib/kubelet/pods/67ac03d0-1fbe-46ff-b61d-7d15bcf48f61/volumes" Dec 11 08:29:39 crc kubenswrapper[4860]: I1211 08:29:39.648995 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f153e5dc-0977-40b7-9ed6-ad88919f1081" path="/var/lib/kubelet/pods/f153e5dc-0977-40b7-9ed6-ad88919f1081/volumes" Dec 11 08:29:39 crc kubenswrapper[4860]: I1211 08:29:39.668819 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 11 08:29:39 crc kubenswrapper[4860]: I1211 08:29:39.982526 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fe0e81d5-8772-4047-9b03-fd8127be2b66-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "fe0e81d5-8772-4047-9b03-fd8127be2b66" (UID: "fe0e81d5-8772-4047-9b03-fd8127be2b66"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:29:39 crc kubenswrapper[4860]: I1211 08:29:39.984079 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fe0e81d5-8772-4047-9b03-fd8127be2b66-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "fe0e81d5-8772-4047-9b03-fd8127be2b66" (UID: "fe0e81d5-8772-4047-9b03-fd8127be2b66"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:29:39 crc kubenswrapper[4860]: I1211 08:29:39.991885 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fe0e81d5-8772-4047-9b03-fd8127be2b66-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "fe0e81d5-8772-4047-9b03-fd8127be2b66" (UID: "fe0e81d5-8772-4047-9b03-fd8127be2b66"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:29:40 crc kubenswrapper[4860]: I1211 08:29:40.015966 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fe0e81d5-8772-4047-9b03-fd8127be2b66-config" (OuterVolumeSpecName: "config") pod "fe0e81d5-8772-4047-9b03-fd8127be2b66" (UID: "fe0e81d5-8772-4047-9b03-fd8127be2b66"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:29:40 crc kubenswrapper[4860]: I1211 08:29:40.024722 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fe0e81d5-8772-4047-9b03-fd8127be2b66-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "fe0e81d5-8772-4047-9b03-fd8127be2b66" (UID: "fe0e81d5-8772-4047-9b03-fd8127be2b66"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:29:40 crc kubenswrapper[4860]: I1211 08:29:40.051892 4860 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fe0e81d5-8772-4047-9b03-fd8127be2b66-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 11 08:29:40 crc kubenswrapper[4860]: I1211 08:29:40.051952 4860 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fe0e81d5-8772-4047-9b03-fd8127be2b66-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 11 08:29:40 crc kubenswrapper[4860]: I1211 08:29:40.051963 4860 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fe0e81d5-8772-4047-9b03-fd8127be2b66-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 11 08:29:40 crc kubenswrapper[4860]: I1211 08:29:40.051972 4860 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fe0e81d5-8772-4047-9b03-fd8127be2b66-config\") on node \"crc\" DevicePath \"\"" Dec 11 08:29:40 crc kubenswrapper[4860]: I1211 08:29:40.051981 4860 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fe0e81d5-8772-4047-9b03-fd8127be2b66-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 11 08:29:40 crc kubenswrapper[4860]: I1211 08:29:40.248719 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc5c4795-65k7g" event={"ID":"fe0e81d5-8772-4047-9b03-fd8127be2b66","Type":"ContainerDied","Data":"b29d22420160c71e4e32523ec98bf724a176d783a853ef0b4d9dd854f65c61da"} Dec 11 08:29:40 crc kubenswrapper[4860]: I1211 08:29:40.248828 4860 scope.go:117] "RemoveContainer" containerID="5c1825a4e2c7afa36121cf72f71a0143cd39e74617e9246b561d6909c66cc58d" Dec 11 08:29:40 crc kubenswrapper[4860]: I1211 08:29:40.248984 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc5c4795-65k7g" Dec 11 08:29:40 crc kubenswrapper[4860]: I1211 08:29:40.263232 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-55ccdb6bbd-grdt6" event={"ID":"78683923-6a60-4a9a-bb81-51ae61f55376","Type":"ContainerStarted","Data":"419075c60fb3d8583e35dee45da2fa730355ba5e2f714956a8de05a232b6eefb"} Dec 11 08:29:40 crc kubenswrapper[4860]: I1211 08:29:40.263282 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-55ccdb6bbd-grdt6" event={"ID":"78683923-6a60-4a9a-bb81-51ae61f55376","Type":"ContainerStarted","Data":"230378f9d383be58191d6c592afb6af08a7d5400e0811a26356a2446ff2294d6"} Dec 11 08:29:40 crc kubenswrapper[4860]: I1211 08:29:40.263323 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-55ccdb6bbd-grdt6" Dec 11 08:29:40 crc kubenswrapper[4860]: I1211 08:29:40.263372 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-55ccdb6bbd-grdt6" Dec 11 08:29:40 crc kubenswrapper[4860]: I1211 08:29:40.332220 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-9bc765b68-htbxs"] Dec 11 08:29:40 crc kubenswrapper[4860]: I1211 08:29:40.360130 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-568d665c75-4vptt"] Dec 11 08:29:40 crc kubenswrapper[4860]: I1211 08:29:40.395809 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-684c9ddd98-t29bd"] Dec 11 08:29:40 crc kubenswrapper[4860]: I1211 08:29:40.410028 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-55ccdb6bbd-grdt6" podStartSLOduration=4.410000321 podStartE2EDuration="4.410000321s" podCreationTimestamp="2025-12-11 08:29:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:29:40.337607425 +0000 UTC m=+1113.066126480" watchObservedRunningTime="2025-12-11 08:29:40.410000321 +0000 UTC m=+1113.138519376" Dec 11 08:29:40 crc kubenswrapper[4860]: I1211 08:29:40.460735 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5ccc5c4795-65k7g"] Dec 11 08:29:40 crc kubenswrapper[4860]: I1211 08:29:40.485894 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5ccc5c4795-65k7g"] Dec 11 08:29:40 crc kubenswrapper[4860]: I1211 08:29:40.508188 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-688c87cc99-sbps7"] Dec 11 08:29:40 crc kubenswrapper[4860]: I1211 08:29:40.534025 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 11 08:29:41 crc kubenswrapper[4860]: I1211 08:29:41.596809 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fe0e81d5-8772-4047-9b03-fd8127be2b66" path="/var/lib/kubelet/pods/fe0e81d5-8772-4047-9b03-fd8127be2b66/volumes" Dec 11 08:29:42 crc kubenswrapper[4860]: I1211 08:29:42.275129 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-6bbf49f946-sw5n4"] Dec 11 08:29:42 crc kubenswrapper[4860]: E1211 08:29:42.275899 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe0e81d5-8772-4047-9b03-fd8127be2b66" containerName="dnsmasq-dns" Dec 11 08:29:42 crc kubenswrapper[4860]: I1211 08:29:42.275921 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe0e81d5-8772-4047-9b03-fd8127be2b66" containerName="dnsmasq-dns" Dec 11 08:29:42 crc kubenswrapper[4860]: E1211 08:29:42.275935 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe0e81d5-8772-4047-9b03-fd8127be2b66" containerName="init" Dec 11 08:29:42 crc kubenswrapper[4860]: I1211 08:29:42.275944 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe0e81d5-8772-4047-9b03-fd8127be2b66" containerName="init" Dec 11 08:29:42 crc kubenswrapper[4860]: I1211 08:29:42.276179 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="fe0e81d5-8772-4047-9b03-fd8127be2b66" containerName="dnsmasq-dns" Dec 11 08:29:42 crc kubenswrapper[4860]: I1211 08:29:42.278477 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6bbf49f946-sw5n4" Dec 11 08:29:42 crc kubenswrapper[4860]: I1211 08:29:42.281367 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Dec 11 08:29:42 crc kubenswrapper[4860]: I1211 08:29:42.283023 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Dec 11 08:29:42 crc kubenswrapper[4860]: I1211 08:29:42.290569 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6bbf49f946-sw5n4"] Dec 11 08:29:42 crc kubenswrapper[4860]: I1211 08:29:42.420691 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4a941157-652e-4efa-9165-ba55161bfb51-config-data-custom\") pod \"barbican-api-6bbf49f946-sw5n4\" (UID: \"4a941157-652e-4efa-9165-ba55161bfb51\") " pod="openstack/barbican-api-6bbf49f946-sw5n4" Dec 11 08:29:42 crc kubenswrapper[4860]: I1211 08:29:42.420758 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4a941157-652e-4efa-9165-ba55161bfb51-logs\") pod \"barbican-api-6bbf49f946-sw5n4\" (UID: \"4a941157-652e-4efa-9165-ba55161bfb51\") " pod="openstack/barbican-api-6bbf49f946-sw5n4" Dec 11 08:29:42 crc kubenswrapper[4860]: I1211 08:29:42.420826 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r4ddw\" (UniqueName: \"kubernetes.io/projected/4a941157-652e-4efa-9165-ba55161bfb51-kube-api-access-r4ddw\") pod \"barbican-api-6bbf49f946-sw5n4\" (UID: \"4a941157-652e-4efa-9165-ba55161bfb51\") " pod="openstack/barbican-api-6bbf49f946-sw5n4" Dec 11 08:29:42 crc kubenswrapper[4860]: I1211 08:29:42.420877 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4a941157-652e-4efa-9165-ba55161bfb51-public-tls-certs\") pod \"barbican-api-6bbf49f946-sw5n4\" (UID: \"4a941157-652e-4efa-9165-ba55161bfb51\") " pod="openstack/barbican-api-6bbf49f946-sw5n4" Dec 11 08:29:42 crc kubenswrapper[4860]: I1211 08:29:42.420953 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a941157-652e-4efa-9165-ba55161bfb51-combined-ca-bundle\") pod \"barbican-api-6bbf49f946-sw5n4\" (UID: \"4a941157-652e-4efa-9165-ba55161bfb51\") " pod="openstack/barbican-api-6bbf49f946-sw5n4" Dec 11 08:29:42 crc kubenswrapper[4860]: I1211 08:29:42.421045 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4a941157-652e-4efa-9165-ba55161bfb51-internal-tls-certs\") pod \"barbican-api-6bbf49f946-sw5n4\" (UID: \"4a941157-652e-4efa-9165-ba55161bfb51\") " pod="openstack/barbican-api-6bbf49f946-sw5n4" Dec 11 08:29:42 crc kubenswrapper[4860]: I1211 08:29:42.421074 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a941157-652e-4efa-9165-ba55161bfb51-config-data\") pod \"barbican-api-6bbf49f946-sw5n4\" (UID: \"4a941157-652e-4efa-9165-ba55161bfb51\") " pod="openstack/barbican-api-6bbf49f946-sw5n4" Dec 11 08:29:42 crc kubenswrapper[4860]: I1211 08:29:42.523000 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4a941157-652e-4efa-9165-ba55161bfb51-internal-tls-certs\") pod \"barbican-api-6bbf49f946-sw5n4\" (UID: \"4a941157-652e-4efa-9165-ba55161bfb51\") " pod="openstack/barbican-api-6bbf49f946-sw5n4" Dec 11 08:29:42 crc kubenswrapper[4860]: I1211 08:29:42.523323 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a941157-652e-4efa-9165-ba55161bfb51-config-data\") pod \"barbican-api-6bbf49f946-sw5n4\" (UID: \"4a941157-652e-4efa-9165-ba55161bfb51\") " pod="openstack/barbican-api-6bbf49f946-sw5n4" Dec 11 08:29:42 crc kubenswrapper[4860]: I1211 08:29:42.523379 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4a941157-652e-4efa-9165-ba55161bfb51-config-data-custom\") pod \"barbican-api-6bbf49f946-sw5n4\" (UID: \"4a941157-652e-4efa-9165-ba55161bfb51\") " pod="openstack/barbican-api-6bbf49f946-sw5n4" Dec 11 08:29:42 crc kubenswrapper[4860]: I1211 08:29:42.523407 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4a941157-652e-4efa-9165-ba55161bfb51-logs\") pod \"barbican-api-6bbf49f946-sw5n4\" (UID: \"4a941157-652e-4efa-9165-ba55161bfb51\") " pod="openstack/barbican-api-6bbf49f946-sw5n4" Dec 11 08:29:42 crc kubenswrapper[4860]: I1211 08:29:42.523439 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r4ddw\" (UniqueName: \"kubernetes.io/projected/4a941157-652e-4efa-9165-ba55161bfb51-kube-api-access-r4ddw\") pod \"barbican-api-6bbf49f946-sw5n4\" (UID: \"4a941157-652e-4efa-9165-ba55161bfb51\") " pod="openstack/barbican-api-6bbf49f946-sw5n4" Dec 11 08:29:42 crc kubenswrapper[4860]: I1211 08:29:42.523476 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4a941157-652e-4efa-9165-ba55161bfb51-public-tls-certs\") pod \"barbican-api-6bbf49f946-sw5n4\" (UID: \"4a941157-652e-4efa-9165-ba55161bfb51\") " pod="openstack/barbican-api-6bbf49f946-sw5n4" Dec 11 08:29:42 crc kubenswrapper[4860]: I1211 08:29:42.523530 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a941157-652e-4efa-9165-ba55161bfb51-combined-ca-bundle\") pod \"barbican-api-6bbf49f946-sw5n4\" (UID: \"4a941157-652e-4efa-9165-ba55161bfb51\") " pod="openstack/barbican-api-6bbf49f946-sw5n4" Dec 11 08:29:42 crc kubenswrapper[4860]: I1211 08:29:42.523848 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4a941157-652e-4efa-9165-ba55161bfb51-logs\") pod \"barbican-api-6bbf49f946-sw5n4\" (UID: \"4a941157-652e-4efa-9165-ba55161bfb51\") " pod="openstack/barbican-api-6bbf49f946-sw5n4" Dec 11 08:29:42 crc kubenswrapper[4860]: I1211 08:29:42.534852 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a941157-652e-4efa-9165-ba55161bfb51-combined-ca-bundle\") pod \"barbican-api-6bbf49f946-sw5n4\" (UID: \"4a941157-652e-4efa-9165-ba55161bfb51\") " pod="openstack/barbican-api-6bbf49f946-sw5n4" Dec 11 08:29:42 crc kubenswrapper[4860]: I1211 08:29:42.548418 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4a941157-652e-4efa-9165-ba55161bfb51-public-tls-certs\") pod \"barbican-api-6bbf49f946-sw5n4\" (UID: \"4a941157-652e-4efa-9165-ba55161bfb51\") " pod="openstack/barbican-api-6bbf49f946-sw5n4" Dec 11 08:29:42 crc kubenswrapper[4860]: I1211 08:29:42.548860 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4a941157-652e-4efa-9165-ba55161bfb51-internal-tls-certs\") pod \"barbican-api-6bbf49f946-sw5n4\" (UID: \"4a941157-652e-4efa-9165-ba55161bfb51\") " pod="openstack/barbican-api-6bbf49f946-sw5n4" Dec 11 08:29:42 crc kubenswrapper[4860]: I1211 08:29:42.560294 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4a941157-652e-4efa-9165-ba55161bfb51-config-data-custom\") pod \"barbican-api-6bbf49f946-sw5n4\" (UID: \"4a941157-652e-4efa-9165-ba55161bfb51\") " pod="openstack/barbican-api-6bbf49f946-sw5n4" Dec 11 08:29:42 crc kubenswrapper[4860]: I1211 08:29:42.561019 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a941157-652e-4efa-9165-ba55161bfb51-config-data\") pod \"barbican-api-6bbf49f946-sw5n4\" (UID: \"4a941157-652e-4efa-9165-ba55161bfb51\") " pod="openstack/barbican-api-6bbf49f946-sw5n4" Dec 11 08:29:42 crc kubenswrapper[4860]: I1211 08:29:42.561153 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r4ddw\" (UniqueName: \"kubernetes.io/projected/4a941157-652e-4efa-9165-ba55161bfb51-kube-api-access-r4ddw\") pod \"barbican-api-6bbf49f946-sw5n4\" (UID: \"4a941157-652e-4efa-9165-ba55161bfb51\") " pod="openstack/barbican-api-6bbf49f946-sw5n4" Dec 11 08:29:42 crc kubenswrapper[4860]: I1211 08:29:42.642875 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6bbf49f946-sw5n4" Dec 11 08:29:43 crc kubenswrapper[4860]: W1211 08:29:43.130596 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5bc875db_b1de_4250_b9e7_bdd3a54bb08f.slice/crio-5f23cfec3d4c07fa6689cf43fd028636386475064e2c2335cbaf1bf64442da5b WatchSource:0}: Error finding container 5f23cfec3d4c07fa6689cf43fd028636386475064e2c2335cbaf1bf64442da5b: Status 404 returned error can't find the container with id 5f23cfec3d4c07fa6689cf43fd028636386475064e2c2335cbaf1bf64442da5b Dec 11 08:29:43 crc kubenswrapper[4860]: W1211 08:29:43.133586 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda7e1def2_f417_40f5_a7dc_c63148858073.slice/crio-06bc6e04e503e5e21255f2314d8c517ee1b8f94ad036a2189842f5e534ab106f WatchSource:0}: Error finding container 06bc6e04e503e5e21255f2314d8c517ee1b8f94ad036a2189842f5e534ab106f: Status 404 returned error can't find the container with id 06bc6e04e503e5e21255f2314d8c517ee1b8f94ad036a2189842f5e534ab106f Dec 11 08:29:43 crc kubenswrapper[4860]: W1211 08:29:43.144283 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd8954df2_1f06_4bbe_bb33_a5ac8fdf9675.slice/crio-bf0b1b404abce746a221aa7f00542b6d9dd8436e98ab6c04102892609660fb0a WatchSource:0}: Error finding container bf0b1b404abce746a221aa7f00542b6d9dd8436e98ab6c04102892609660fb0a: Status 404 returned error can't find the container with id bf0b1b404abce746a221aa7f00542b6d9dd8436e98ab6c04102892609660fb0a Dec 11 08:29:43 crc kubenswrapper[4860]: I1211 08:29:43.298292 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-684c9ddd98-t29bd" event={"ID":"d8954df2-1f06-4bbe-bb33-a5ac8fdf9675","Type":"ContainerStarted","Data":"bf0b1b404abce746a221aa7f00542b6d9dd8436e98ab6c04102892609660fb0a"} Dec 11 08:29:43 crc kubenswrapper[4860]: I1211 08:29:43.300981 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"83082bff-37e1-4ac1-9074-ae9d69ed66b0","Type":"ContainerStarted","Data":"df498fa320b29fcbeb2fb09668b03f3a7564a2703477e78ff4f432d27746d0ba"} Dec 11 08:29:43 crc kubenswrapper[4860]: I1211 08:29:43.306474 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-9bc765b68-htbxs" event={"ID":"0e28c0dc-a169-423c-ae52-4392a9a27a3e","Type":"ContainerStarted","Data":"8c6029898ce57a7af7dbf7baf1cbf99727cf17fc76410859e7ea8d16235c4e0f"} Dec 11 08:29:43 crc kubenswrapper[4860]: I1211 08:29:43.307939 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-688c87cc99-sbps7" event={"ID":"5bc875db-b1de-4250-b9e7-bdd3a54bb08f","Type":"ContainerStarted","Data":"5f23cfec3d4c07fa6689cf43fd028636386475064e2c2335cbaf1bf64442da5b"} Dec 11 08:29:43 crc kubenswrapper[4860]: I1211 08:29:43.309581 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-568d665c75-4vptt" event={"ID":"a7e1def2-f417-40f5-a7dc-c63148858073","Type":"ContainerStarted","Data":"06bc6e04e503e5e21255f2314d8c517ee1b8f94ad036a2189842f5e534ab106f"} Dec 11 08:29:43 crc kubenswrapper[4860]: I1211 08:29:43.311260 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"a3d65995-e403-4e92-b7aa-b39b8084d1bc","Type":"ContainerStarted","Data":"099601ebce1a3974f9e908e2166615be7d1eb2fba71bc7455e8b36301a7886da"} Dec 11 08:29:45 crc kubenswrapper[4860]: I1211 08:29:45.300799 4860 scope.go:117] "RemoveContainer" containerID="e600d5682ae00beaea5e17f2097e7aaf07fbca21b4d3422f88dd24fd87141681" Dec 11 08:29:46 crc kubenswrapper[4860]: I1211 08:29:46.090260 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6bbf49f946-sw5n4"] Dec 11 08:29:46 crc kubenswrapper[4860]: W1211 08:29:46.118450 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4a941157_652e_4efa_9165_ba55161bfb51.slice/crio-b4fd585f8e98bcc9bd64ffc322a768312300642cc7775244da2451ed3b32d0c0 WatchSource:0}: Error finding container b4fd585f8e98bcc9bd64ffc322a768312300642cc7775244da2451ed3b32d0c0: Status 404 returned error can't find the container with id b4fd585f8e98bcc9bd64ffc322a768312300642cc7775244da2451ed3b32d0c0 Dec 11 08:29:46 crc kubenswrapper[4860]: I1211 08:29:46.243856 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-748d5bd978-4sz6v"] Dec 11 08:29:46 crc kubenswrapper[4860]: W1211 08:29:46.270541 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode208d74e_f50f_443c_aa2d_eccdb584413d.slice/crio-173b98519d25e79a34656b665c5f42f3b262f07e07490350c964adf0144729d8 WatchSource:0}: Error finding container 173b98519d25e79a34656b665c5f42f3b262f07e07490350c964adf0144729d8: Status 404 returned error can't find the container with id 173b98519d25e79a34656b665c5f42f3b262f07e07490350c964adf0144729d8 Dec 11 08:29:46 crc kubenswrapper[4860]: I1211 08:29:46.370763 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6bbf49f946-sw5n4" event={"ID":"4a941157-652e-4efa-9165-ba55161bfb51","Type":"ContainerStarted","Data":"b4fd585f8e98bcc9bd64ffc322a768312300642cc7775244da2451ed3b32d0c0"} Dec 11 08:29:46 crc kubenswrapper[4860]: I1211 08:29:46.389055 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-748d5bd978-4sz6v" event={"ID":"e208d74e-f50f-443c-aa2d-eccdb584413d","Type":"ContainerStarted","Data":"173b98519d25e79a34656b665c5f42f3b262f07e07490350c964adf0144729d8"} Dec 11 08:29:46 crc kubenswrapper[4860]: I1211 08:29:46.697017 4860 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-776755b658-9fwr9" podUID="8f38e033-5243-4e65-a7e0-29f1282b063f" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.142:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.142:8443: connect: connection refused" Dec 11 08:29:46 crc kubenswrapper[4860]: I1211 08:29:46.840753 4860 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-6b4697fc68-vnd26" podUID="78b0f75e-d447-4541-be11-9afe4549d2e1" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.143:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.143:8443: connect: connection refused" Dec 11 08:29:47 crc kubenswrapper[4860]: I1211 08:29:47.414121 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-748d5bd978-4sz6v" event={"ID":"e208d74e-f50f-443c-aa2d-eccdb584413d","Type":"ContainerStarted","Data":"91e69082b542f13de68a7fcef270278602d7e8c59998e02a2aaa89fb9e379bf0"} Dec 11 08:29:47 crc kubenswrapper[4860]: I1211 08:29:47.414805 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-748d5bd978-4sz6v" Dec 11 08:29:47 crc kubenswrapper[4860]: I1211 08:29:47.417929 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"83082bff-37e1-4ac1-9074-ae9d69ed66b0","Type":"ContainerStarted","Data":"80eab5ca76a36476cefed22689137813096f4b92f167de2ab1d721595040852f"} Dec 11 08:29:47 crc kubenswrapper[4860]: I1211 08:29:47.421589 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-9bc765b68-htbxs" event={"ID":"0e28c0dc-a169-423c-ae52-4392a9a27a3e","Type":"ContainerStarted","Data":"c8f91956bd174855f2e8c77b9ae2a5ed3b795d6f0f347182c612a951754995ed"} Dec 11 08:29:47 crc kubenswrapper[4860]: I1211 08:29:47.421798 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-9bc765b68-htbxs" event={"ID":"0e28c0dc-a169-423c-ae52-4392a9a27a3e","Type":"ContainerStarted","Data":"1a39faaae155e7004b855c4ae10bd43b2d7f19545e167115add4bd8275302ead"} Dec 11 08:29:47 crc kubenswrapper[4860]: I1211 08:29:47.423005 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-9bc765b68-htbxs" Dec 11 08:29:47 crc kubenswrapper[4860]: I1211 08:29:47.423077 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-9bc765b68-htbxs" Dec 11 08:29:47 crc kubenswrapper[4860]: I1211 08:29:47.427422 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6bbf49f946-sw5n4" event={"ID":"4a941157-652e-4efa-9165-ba55161bfb51","Type":"ContainerStarted","Data":"8369aaff634ba5f0d5e951fd7246c7174413e05ca6cd4e907767cb3a30d825f5"} Dec 11 08:29:47 crc kubenswrapper[4860]: I1211 08:29:47.427521 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6bbf49f946-sw5n4" Dec 11 08:29:47 crc kubenswrapper[4860]: I1211 08:29:47.427575 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6bbf49f946-sw5n4" Dec 11 08:29:47 crc kubenswrapper[4860]: I1211 08:29:47.442665 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-748d5bd978-4sz6v" podStartSLOduration=8.442629072999999 podStartE2EDuration="8.442629073s" podCreationTimestamp="2025-12-11 08:29:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:29:47.435755616 +0000 UTC m=+1120.164274711" watchObservedRunningTime="2025-12-11 08:29:47.442629073 +0000 UTC m=+1120.171148128" Dec 11 08:29:47 crc kubenswrapper[4860]: I1211 08:29:47.461457 4860 generic.go:334] "Generic (PLEG): container finished" podID="5bc875db-b1de-4250-b9e7-bdd3a54bb08f" containerID="69e0e84cf252adb8a847241eca4fd03574e73a2fe440b02276821f9fab9eecd4" exitCode=0 Dec 11 08:29:47 crc kubenswrapper[4860]: I1211 08:29:47.461784 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-688c87cc99-sbps7" event={"ID":"5bc875db-b1de-4250-b9e7-bdd3a54bb08f","Type":"ContainerDied","Data":"69e0e84cf252adb8a847241eca4fd03574e73a2fe440b02276821f9fab9eecd4"} Dec 11 08:29:47 crc kubenswrapper[4860]: I1211 08:29:47.464347 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-6bbf49f946-sw5n4" podStartSLOduration=5.464307464 podStartE2EDuration="5.464307464s" podCreationTimestamp="2025-12-11 08:29:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:29:47.460747332 +0000 UTC m=+1120.189266387" watchObservedRunningTime="2025-12-11 08:29:47.464307464 +0000 UTC m=+1120.192826519" Dec 11 08:29:47 crc kubenswrapper[4860]: I1211 08:29:47.493346 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"a3d65995-e403-4e92-b7aa-b39b8084d1bc","Type":"ContainerStarted","Data":"e0910e8fa92a5f01a5e9fcf10245b4f567cd1f91da1a21ce559bfaf3ce103f95"} Dec 11 08:29:47 crc kubenswrapper[4860]: I1211 08:29:47.554716 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-9bc765b68-htbxs" podStartSLOduration=9.554694395 podStartE2EDuration="9.554694395s" podCreationTimestamp="2025-12-11 08:29:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:29:47.480687239 +0000 UTC m=+1120.209206314" watchObservedRunningTime="2025-12-11 08:29:47.554694395 +0000 UTC m=+1120.283213450" Dec 11 08:29:48 crc kubenswrapper[4860]: I1211 08:29:48.507386 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-688c87cc99-sbps7" event={"ID":"5bc875db-b1de-4250-b9e7-bdd3a54bb08f","Type":"ContainerStarted","Data":"fb75e210972ef3010591ee0d3883d834f38a29b13cf4e9a71a0291a1ac174852"} Dec 11 08:29:48 crc kubenswrapper[4860]: I1211 08:29:48.508829 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-688c87cc99-sbps7" Dec 11 08:29:48 crc kubenswrapper[4860]: I1211 08:29:48.512348 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"a3d65995-e403-4e92-b7aa-b39b8084d1bc","Type":"ContainerStarted","Data":"b2623116284268f161ab7f24e48bff42f3f1d8de1c286d66200076c14429dc28"} Dec 11 08:29:48 crc kubenswrapper[4860]: I1211 08:29:48.514297 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"040a0e8c-efd3-4e3b-be51-59e311ae4406","Type":"ContainerStarted","Data":"6462b2d96ccced3f64b359817e1219e8bb94873c821a2372070dd0c60f426b68"} Dec 11 08:29:48 crc kubenswrapper[4860]: I1211 08:29:48.516751 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"83082bff-37e1-4ac1-9074-ae9d69ed66b0","Type":"ContainerStarted","Data":"239173e5da0a533df7671655d48736e1bc38b5098483280871fd5ef02669df4c"} Dec 11 08:29:48 crc kubenswrapper[4860]: I1211 08:29:48.518870 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-hccts" event={"ID":"d472d27e-9bc9-4d99-8fd9-697f7f44f57d","Type":"ContainerStarted","Data":"3e0c4ee5f6bdd93a3d5e3c150e68872d9d1ed659b0878f63a767498c0c729c7c"} Dec 11 08:29:48 crc kubenswrapper[4860]: I1211 08:29:48.523719 4860 generic.go:334] "Generic (PLEG): container finished" podID="0e28c0dc-a169-423c-ae52-4392a9a27a3e" containerID="c8f91956bd174855f2e8c77b9ae2a5ed3b795d6f0f347182c612a951754995ed" exitCode=1 Dec 11 08:29:48 crc kubenswrapper[4860]: I1211 08:29:48.523793 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-9bc765b68-htbxs" event={"ID":"0e28c0dc-a169-423c-ae52-4392a9a27a3e","Type":"ContainerDied","Data":"c8f91956bd174855f2e8c77b9ae2a5ed3b795d6f0f347182c612a951754995ed"} Dec 11 08:29:48 crc kubenswrapper[4860]: I1211 08:29:48.524357 4860 scope.go:117] "RemoveContainer" containerID="c8f91956bd174855f2e8c77b9ae2a5ed3b795d6f0f347182c612a951754995ed" Dec 11 08:29:48 crc kubenswrapper[4860]: I1211 08:29:48.534254 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6bbf49f946-sw5n4" event={"ID":"4a941157-652e-4efa-9165-ba55161bfb51","Type":"ContainerStarted","Data":"5139c2cc09c8608ec866a334fff36d4f39f5165c3a9620fd814a1fceffba3751"} Dec 11 08:29:48 crc kubenswrapper[4860]: I1211 08:29:48.582073 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=11.582044486000001 podStartE2EDuration="11.582044486s" podCreationTimestamp="2025-12-11 08:29:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:29:48.548907213 +0000 UTC m=+1121.277426278" watchObservedRunningTime="2025-12-11 08:29:48.582044486 +0000 UTC m=+1121.310563531" Dec 11 08:29:48 crc kubenswrapper[4860]: I1211 08:29:48.597318 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-688c87cc99-sbps7" podStartSLOduration=10.597277934 podStartE2EDuration="10.597277934s" podCreationTimestamp="2025-12-11 08:29:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:29:48.527945795 +0000 UTC m=+1121.256464860" watchObservedRunningTime="2025-12-11 08:29:48.597277934 +0000 UTC m=+1121.325797029" Dec 11 08:29:48 crc kubenswrapper[4860]: I1211 08:29:48.607705 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-hccts" podStartSLOduration=3.174544536 podStartE2EDuration="50.607675661s" podCreationTimestamp="2025-12-11 08:28:58 +0000 UTC" firstStartedPulling="2025-12-11 08:28:59.745872142 +0000 UTC m=+1072.474391197" lastFinishedPulling="2025-12-11 08:29:47.179003267 +0000 UTC m=+1119.907522322" observedRunningTime="2025-12-11 08:29:48.564307028 +0000 UTC m=+1121.292826083" watchObservedRunningTime="2025-12-11 08:29:48.607675661 +0000 UTC m=+1121.336194716" Dec 11 08:29:48 crc kubenswrapper[4860]: I1211 08:29:48.613718 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=11.613229765 podStartE2EDuration="11.613229765s" podCreationTimestamp="2025-12-11 08:29:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:29:48.585733191 +0000 UTC m=+1121.314252246" watchObservedRunningTime="2025-12-11 08:29:48.613229765 +0000 UTC m=+1121.341748820" Dec 11 08:29:48 crc kubenswrapper[4860]: I1211 08:29:48.697284 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 11 08:29:48 crc kubenswrapper[4860]: I1211 08:29:48.697333 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 11 08:29:48 crc kubenswrapper[4860]: I1211 08:29:48.750912 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 11 08:29:48 crc kubenswrapper[4860]: I1211 08:29:48.762823 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 11 08:29:49 crc kubenswrapper[4860]: I1211 08:29:49.541011 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 11 08:29:49 crc kubenswrapper[4860]: I1211 08:29:49.541323 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 11 08:29:50 crc kubenswrapper[4860]: I1211 08:29:50.552974 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-684c9ddd98-t29bd" event={"ID":"d8954df2-1f06-4bbe-bb33-a5ac8fdf9675","Type":"ContainerStarted","Data":"2ae9f42750f40a0a16a34a4ea3e24c9e2bd92958176c1a7247206b79992ac0ae"} Dec 11 08:29:50 crc kubenswrapper[4860]: I1211 08:29:50.553342 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-684c9ddd98-t29bd" event={"ID":"d8954df2-1f06-4bbe-bb33-a5ac8fdf9675","Type":"ContainerStarted","Data":"162c739c93edca4939129f26c9eab872f94a6e043a19e5a3217196336ebfcb31"} Dec 11 08:29:50 crc kubenswrapper[4860]: I1211 08:29:50.559947 4860 generic.go:334] "Generic (PLEG): container finished" podID="0e28c0dc-a169-423c-ae52-4392a9a27a3e" containerID="9fd2099322d500e85f729d00e2769ccfc0be26e059a57bfba2f1a45d2245d434" exitCode=1 Dec 11 08:29:50 crc kubenswrapper[4860]: I1211 08:29:50.560059 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-9bc765b68-htbxs" event={"ID":"0e28c0dc-a169-423c-ae52-4392a9a27a3e","Type":"ContainerDied","Data":"9fd2099322d500e85f729d00e2769ccfc0be26e059a57bfba2f1a45d2245d434"} Dec 11 08:29:50 crc kubenswrapper[4860]: I1211 08:29:50.560142 4860 scope.go:117] "RemoveContainer" containerID="c8f91956bd174855f2e8c77b9ae2a5ed3b795d6f0f347182c612a951754995ed" Dec 11 08:29:50 crc kubenswrapper[4860]: I1211 08:29:50.560751 4860 scope.go:117] "RemoveContainer" containerID="9fd2099322d500e85f729d00e2769ccfc0be26e059a57bfba2f1a45d2245d434" Dec 11 08:29:50 crc kubenswrapper[4860]: E1211 08:29:50.561052 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=barbican-api pod=barbican-api-9bc765b68-htbxs_openstack(0e28c0dc-a169-423c-ae52-4392a9a27a3e)\"" pod="openstack/barbican-api-9bc765b68-htbxs" podUID="0e28c0dc-a169-423c-ae52-4392a9a27a3e" Dec 11 08:29:50 crc kubenswrapper[4860]: I1211 08:29:50.564991 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-568d665c75-4vptt" event={"ID":"a7e1def2-f417-40f5-a7dc-c63148858073","Type":"ContainerStarted","Data":"38623051a719529b891c203b59a549a94b3c4d84433bfbe35562179779c4470e"} Dec 11 08:29:50 crc kubenswrapper[4860]: I1211 08:29:50.565034 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-568d665c75-4vptt" event={"ID":"a7e1def2-f417-40f5-a7dc-c63148858073","Type":"ContainerStarted","Data":"a589a482cf3db13d8622e4fce05fd94ceede45692610b5d2c592e37723f35155"} Dec 11 08:29:50 crc kubenswrapper[4860]: I1211 08:29:50.580290 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-684c9ddd98-t29bd" podStartSLOduration=6.47921848 podStartE2EDuration="12.580263791s" podCreationTimestamp="2025-12-11 08:29:38 +0000 UTC" firstStartedPulling="2025-12-11 08:29:43.153515052 +0000 UTC m=+1115.882034107" lastFinishedPulling="2025-12-11 08:29:49.254560363 +0000 UTC m=+1121.983079418" observedRunningTime="2025-12-11 08:29:50.575904614 +0000 UTC m=+1123.304423689" watchObservedRunningTime="2025-12-11 08:29:50.580263791 +0000 UTC m=+1123.308782846" Dec 11 08:29:50 crc kubenswrapper[4860]: I1211 08:29:50.612260 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-568d665c75-4vptt" podStartSLOduration=6.490920528 podStartE2EDuration="12.612235166s" podCreationTimestamp="2025-12-11 08:29:38 +0000 UTC" firstStartedPulling="2025-12-11 08:29:43.142491776 +0000 UTC m=+1115.871010831" lastFinishedPulling="2025-12-11 08:29:49.263806414 +0000 UTC m=+1121.992325469" observedRunningTime="2025-12-11 08:29:50.600195838 +0000 UTC m=+1123.328714893" watchObservedRunningTime="2025-12-11 08:29:50.612235166 +0000 UTC m=+1123.340754221" Dec 11 08:29:51 crc kubenswrapper[4860]: I1211 08:29:51.073537 4860 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/barbican-api-9bc765b68-htbxs" Dec 11 08:29:51 crc kubenswrapper[4860]: I1211 08:29:51.075215 4860 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-9bc765b68-htbxs" podUID="0e28c0dc-a169-423c-ae52-4392a9a27a3e" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.157:9311/healthcheck\": dial tcp 10.217.0.157:9311: connect: connection refused" Dec 11 08:29:51 crc kubenswrapper[4860]: I1211 08:29:51.599629 4860 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-9bc765b68-htbxs" podUID="0e28c0dc-a169-423c-ae52-4392a9a27a3e" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.157:9311/healthcheck\": dial tcp 10.217.0.157:9311: connect: connection refused" Dec 11 08:29:51 crc kubenswrapper[4860]: I1211 08:29:51.601078 4860 scope.go:117] "RemoveContainer" containerID="9fd2099322d500e85f729d00e2769ccfc0be26e059a57bfba2f1a45d2245d434" Dec 11 08:29:51 crc kubenswrapper[4860]: E1211 08:29:51.601396 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=barbican-api pod=barbican-api-9bc765b68-htbxs_openstack(0e28c0dc-a169-423c-ae52-4392a9a27a3e)\"" pod="openstack/barbican-api-9bc765b68-htbxs" podUID="0e28c0dc-a169-423c-ae52-4392a9a27a3e" Dec 11 08:29:52 crc kubenswrapper[4860]: I1211 08:29:52.607582 4860 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-9bc765b68-htbxs" podUID="0e28c0dc-a169-423c-ae52-4392a9a27a3e" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.157:9311/healthcheck\": dial tcp 10.217.0.157:9311: connect: connection refused" Dec 11 08:29:52 crc kubenswrapper[4860]: I1211 08:29:52.608303 4860 scope.go:117] "RemoveContainer" containerID="9fd2099322d500e85f729d00e2769ccfc0be26e059a57bfba2f1a45d2245d434" Dec 11 08:29:52 crc kubenswrapper[4860]: E1211 08:29:52.608572 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=barbican-api pod=barbican-api-9bc765b68-htbxs_openstack(0e28c0dc-a169-423c-ae52-4392a9a27a3e)\"" pod="openstack/barbican-api-9bc765b68-htbxs" podUID="0e28c0dc-a169-423c-ae52-4392a9a27a3e" Dec 11 08:29:53 crc kubenswrapper[4860]: I1211 08:29:53.619278 4860 generic.go:334] "Generic (PLEG): container finished" podID="d472d27e-9bc9-4d99-8fd9-697f7f44f57d" containerID="3e0c4ee5f6bdd93a3d5e3c150e68872d9d1ed659b0878f63a767498c0c729c7c" exitCode=0 Dec 11 08:29:53 crc kubenswrapper[4860]: I1211 08:29:53.619474 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-hccts" event={"ID":"d472d27e-9bc9-4d99-8fd9-697f7f44f57d","Type":"ContainerDied","Data":"3e0c4ee5f6bdd93a3d5e3c150e68872d9d1ed659b0878f63a767498c0c729c7c"} Dec 11 08:29:54 crc kubenswrapper[4860]: I1211 08:29:54.065979 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-688c87cc99-sbps7" Dec 11 08:29:54 crc kubenswrapper[4860]: I1211 08:29:54.077410 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-9bc765b68-htbxs" Dec 11 08:29:54 crc kubenswrapper[4860]: I1211 08:29:54.077905 4860 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-9bc765b68-htbxs" podUID="0e28c0dc-a169-423c-ae52-4392a9a27a3e" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.157:9311/healthcheck\": dial tcp 10.217.0.157:9311: connect: connection refused" Dec 11 08:29:54 crc kubenswrapper[4860]: I1211 08:29:54.078385 4860 scope.go:117] "RemoveContainer" containerID="9fd2099322d500e85f729d00e2769ccfc0be26e059a57bfba2f1a45d2245d434" Dec 11 08:29:54 crc kubenswrapper[4860]: I1211 08:29:54.078449 4860 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-9bc765b68-htbxs" podUID="0e28c0dc-a169-423c-ae52-4392a9a27a3e" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.157:9311/healthcheck\": dial tcp 10.217.0.157:9311: connect: connection refused" Dec 11 08:29:54 crc kubenswrapper[4860]: E1211 08:29:54.078628 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=barbican-api pod=barbican-api-9bc765b68-htbxs_openstack(0e28c0dc-a169-423c-ae52-4392a9a27a3e)\"" pod="openstack/barbican-api-9bc765b68-htbxs" podUID="0e28c0dc-a169-423c-ae52-4392a9a27a3e" Dec 11 08:29:54 crc kubenswrapper[4860]: I1211 08:29:54.079296 4860 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-9bc765b68-htbxs" podUID="0e28c0dc-a169-423c-ae52-4392a9a27a3e" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.157:9311/healthcheck\": dial tcp 10.217.0.157:9311: connect: connection refused" Dec 11 08:29:54 crc kubenswrapper[4860]: I1211 08:29:54.099001 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6bbf49f946-sw5n4" Dec 11 08:29:54 crc kubenswrapper[4860]: I1211 08:29:54.138529 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-fcfdd6f9f-9vj6f"] Dec 11 08:29:54 crc kubenswrapper[4860]: I1211 08:29:54.138877 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-fcfdd6f9f-9vj6f" podUID="b469f4ba-a16a-4425-b2d6-6d95be8dcc0d" containerName="dnsmasq-dns" containerID="cri-o://8e4b1bc15563056317722c20ddb943a307f432b3ab351a7aa0200777cd3ea1c5" gracePeriod=10 Dec 11 08:29:54 crc kubenswrapper[4860]: I1211 08:29:54.299373 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6bbf49f946-sw5n4" Dec 11 08:29:54 crc kubenswrapper[4860]: I1211 08:29:54.373193 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-9bc765b68-htbxs"] Dec 11 08:29:54 crc kubenswrapper[4860]: I1211 08:29:54.634123 4860 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-9bc765b68-htbxs" podUID="0e28c0dc-a169-423c-ae52-4392a9a27a3e" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.157:9311/healthcheck\": dial tcp 10.217.0.157:9311: connect: connection refused" Dec 11 08:29:54 crc kubenswrapper[4860]: I1211 08:29:54.634943 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-9bc765b68-htbxs" podUID="0e28c0dc-a169-423c-ae52-4392a9a27a3e" containerName="barbican-api-log" containerID="cri-o://1a39faaae155e7004b855c4ae10bd43b2d7f19545e167115add4bd8275302ead" gracePeriod=30 Dec 11 08:29:55 crc kubenswrapper[4860]: E1211 08:29:55.133472 4860 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb469f4ba_a16a_4425_b2d6_6d95be8dcc0d.slice/crio-conmon-8e4b1bc15563056317722c20ddb943a307f432b3ab351a7aa0200777cd3ea1c5.scope\": RecentStats: unable to find data in memory cache]" Dec 11 08:29:55 crc kubenswrapper[4860]: I1211 08:29:55.659910 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-hccts" Dec 11 08:29:55 crc kubenswrapper[4860]: I1211 08:29:55.697915 4860 generic.go:334] "Generic (PLEG): container finished" podID="0e28c0dc-a169-423c-ae52-4392a9a27a3e" containerID="1a39faaae155e7004b855c4ae10bd43b2d7f19545e167115add4bd8275302ead" exitCode=143 Dec 11 08:29:55 crc kubenswrapper[4860]: I1211 08:29:55.697979 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-9bc765b68-htbxs" event={"ID":"0e28c0dc-a169-423c-ae52-4392a9a27a3e","Type":"ContainerDied","Data":"1a39faaae155e7004b855c4ae10bd43b2d7f19545e167115add4bd8275302ead"} Dec 11 08:29:55 crc kubenswrapper[4860]: I1211 08:29:55.706308 4860 generic.go:334] "Generic (PLEG): container finished" podID="b469f4ba-a16a-4425-b2d6-6d95be8dcc0d" containerID="8e4b1bc15563056317722c20ddb943a307f432b3ab351a7aa0200777cd3ea1c5" exitCode=0 Dec 11 08:29:55 crc kubenswrapper[4860]: I1211 08:29:55.706481 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-fcfdd6f9f-9vj6f" event={"ID":"b469f4ba-a16a-4425-b2d6-6d95be8dcc0d","Type":"ContainerDied","Data":"8e4b1bc15563056317722c20ddb943a307f432b3ab351a7aa0200777cd3ea1c5"} Dec 11 08:29:55 crc kubenswrapper[4860]: I1211 08:29:55.711160 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d472d27e-9bc9-4d99-8fd9-697f7f44f57d-etc-machine-id\") pod \"d472d27e-9bc9-4d99-8fd9-697f7f44f57d\" (UID: \"d472d27e-9bc9-4d99-8fd9-697f7f44f57d\") " Dec 11 08:29:55 crc kubenswrapper[4860]: I1211 08:29:55.711324 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d472d27e-9bc9-4d99-8fd9-697f7f44f57d-combined-ca-bundle\") pod \"d472d27e-9bc9-4d99-8fd9-697f7f44f57d\" (UID: \"d472d27e-9bc9-4d99-8fd9-697f7f44f57d\") " Dec 11 08:29:55 crc kubenswrapper[4860]: I1211 08:29:55.711520 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d472d27e-9bc9-4d99-8fd9-697f7f44f57d-scripts\") pod \"d472d27e-9bc9-4d99-8fd9-697f7f44f57d\" (UID: \"d472d27e-9bc9-4d99-8fd9-697f7f44f57d\") " Dec 11 08:29:55 crc kubenswrapper[4860]: I1211 08:29:55.711909 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d472d27e-9bc9-4d99-8fd9-697f7f44f57d-db-sync-config-data\") pod \"d472d27e-9bc9-4d99-8fd9-697f7f44f57d\" (UID: \"d472d27e-9bc9-4d99-8fd9-697f7f44f57d\") " Dec 11 08:29:55 crc kubenswrapper[4860]: I1211 08:29:55.712155 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-84jp8\" (UniqueName: \"kubernetes.io/projected/d472d27e-9bc9-4d99-8fd9-697f7f44f57d-kube-api-access-84jp8\") pod \"d472d27e-9bc9-4d99-8fd9-697f7f44f57d\" (UID: \"d472d27e-9bc9-4d99-8fd9-697f7f44f57d\") " Dec 11 08:29:55 crc kubenswrapper[4860]: I1211 08:29:55.712484 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d472d27e-9bc9-4d99-8fd9-697f7f44f57d-config-data\") pod \"d472d27e-9bc9-4d99-8fd9-697f7f44f57d\" (UID: \"d472d27e-9bc9-4d99-8fd9-697f7f44f57d\") " Dec 11 08:29:55 crc kubenswrapper[4860]: I1211 08:29:55.716609 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d472d27e-9bc9-4d99-8fd9-697f7f44f57d-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "d472d27e-9bc9-4d99-8fd9-697f7f44f57d" (UID: "d472d27e-9bc9-4d99-8fd9-697f7f44f57d"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 08:29:55 crc kubenswrapper[4860]: I1211 08:29:55.719616 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-hccts" event={"ID":"d472d27e-9bc9-4d99-8fd9-697f7f44f57d","Type":"ContainerDied","Data":"b3e71423675304467fc90781a8bf9efd1d88d3d2db17dc44e313cfabdf925503"} Dec 11 08:29:55 crc kubenswrapper[4860]: I1211 08:29:55.719680 4860 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b3e71423675304467fc90781a8bf9efd1d88d3d2db17dc44e313cfabdf925503" Dec 11 08:29:55 crc kubenswrapper[4860]: I1211 08:29:55.719764 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-hccts" Dec 11 08:29:55 crc kubenswrapper[4860]: I1211 08:29:55.724883 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d472d27e-9bc9-4d99-8fd9-697f7f44f57d-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "d472d27e-9bc9-4d99-8fd9-697f7f44f57d" (UID: "d472d27e-9bc9-4d99-8fd9-697f7f44f57d"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:29:55 crc kubenswrapper[4860]: I1211 08:29:55.744563 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d472d27e-9bc9-4d99-8fd9-697f7f44f57d-scripts" (OuterVolumeSpecName: "scripts") pod "d472d27e-9bc9-4d99-8fd9-697f7f44f57d" (UID: "d472d27e-9bc9-4d99-8fd9-697f7f44f57d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:29:55 crc kubenswrapper[4860]: I1211 08:29:55.757958 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d472d27e-9bc9-4d99-8fd9-697f7f44f57d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d472d27e-9bc9-4d99-8fd9-697f7f44f57d" (UID: "d472d27e-9bc9-4d99-8fd9-697f7f44f57d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:29:55 crc kubenswrapper[4860]: I1211 08:29:55.785670 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d472d27e-9bc9-4d99-8fd9-697f7f44f57d-kube-api-access-84jp8" (OuterVolumeSpecName: "kube-api-access-84jp8") pod "d472d27e-9bc9-4d99-8fd9-697f7f44f57d" (UID: "d472d27e-9bc9-4d99-8fd9-697f7f44f57d"). InnerVolumeSpecName "kube-api-access-84jp8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:29:55 crc kubenswrapper[4860]: I1211 08:29:55.815542 4860 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d472d27e-9bc9-4d99-8fd9-697f7f44f57d-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 08:29:55 crc kubenswrapper[4860]: I1211 08:29:55.815573 4860 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d472d27e-9bc9-4d99-8fd9-697f7f44f57d-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 08:29:55 crc kubenswrapper[4860]: I1211 08:29:55.815583 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-84jp8\" (UniqueName: \"kubernetes.io/projected/d472d27e-9bc9-4d99-8fd9-697f7f44f57d-kube-api-access-84jp8\") on node \"crc\" DevicePath \"\"" Dec 11 08:29:55 crc kubenswrapper[4860]: I1211 08:29:55.815595 4860 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d472d27e-9bc9-4d99-8fd9-697f7f44f57d-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 11 08:29:55 crc kubenswrapper[4860]: I1211 08:29:55.815604 4860 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d472d27e-9bc9-4d99-8fd9-697f7f44f57d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 08:29:55 crc kubenswrapper[4860]: I1211 08:29:55.822235 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d472d27e-9bc9-4d99-8fd9-697f7f44f57d-config-data" (OuterVolumeSpecName: "config-data") pod "d472d27e-9bc9-4d99-8fd9-697f7f44f57d" (UID: "d472d27e-9bc9-4d99-8fd9-697f7f44f57d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:29:55 crc kubenswrapper[4860]: I1211 08:29:55.917915 4860 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d472d27e-9bc9-4d99-8fd9-697f7f44f57d-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 08:29:56 crc kubenswrapper[4860]: I1211 08:29:56.696844 4860 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-776755b658-9fwr9" podUID="8f38e033-5243-4e65-a7e0-29f1282b063f" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.142:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.142:8443: connect: connection refused" Dec 11 08:29:56 crc kubenswrapper[4860]: I1211 08:29:56.834519 4860 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-6b4697fc68-vnd26" podUID="78b0f75e-d447-4541-be11-9afe4549d2e1" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.143:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.143:8443: connect: connection refused" Dec 11 08:29:57 crc kubenswrapper[4860]: I1211 08:29:57.145564 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 11 08:29:57 crc kubenswrapper[4860]: E1211 08:29:57.146634 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d472d27e-9bc9-4d99-8fd9-697f7f44f57d" containerName="cinder-db-sync" Dec 11 08:29:57 crc kubenswrapper[4860]: I1211 08:29:57.146669 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="d472d27e-9bc9-4d99-8fd9-697f7f44f57d" containerName="cinder-db-sync" Dec 11 08:29:57 crc kubenswrapper[4860]: I1211 08:29:57.146890 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="d472d27e-9bc9-4d99-8fd9-697f7f44f57d" containerName="cinder-db-sync" Dec 11 08:29:57 crc kubenswrapper[4860]: I1211 08:29:57.148272 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 11 08:29:57 crc kubenswrapper[4860]: I1211 08:29:57.156177 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-8nqnh" Dec 11 08:29:57 crc kubenswrapper[4860]: I1211 08:29:57.156445 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 11 08:29:57 crc kubenswrapper[4860]: I1211 08:29:57.156557 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 11 08:29:57 crc kubenswrapper[4860]: I1211 08:29:57.156757 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 11 08:29:57 crc kubenswrapper[4860]: I1211 08:29:57.176953 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 11 08:29:57 crc kubenswrapper[4860]: I1211 08:29:57.212422 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6bb4fc677f-mcn2v"] Dec 11 08:29:57 crc kubenswrapper[4860]: I1211 08:29:57.215744 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bb4fc677f-mcn2v" Dec 11 08:29:57 crc kubenswrapper[4860]: I1211 08:29:57.219910 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6bb4fc677f-mcn2v"] Dec 11 08:29:57 crc kubenswrapper[4860]: I1211 08:29:57.256900 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6075b482-4926-4cf5-b632-0a731d2686dc-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"6075b482-4926-4cf5-b632-0a731d2686dc\") " pod="openstack/cinder-scheduler-0" Dec 11 08:29:57 crc kubenswrapper[4860]: I1211 08:29:57.256962 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6075b482-4926-4cf5-b632-0a731d2686dc-config-data\") pod \"cinder-scheduler-0\" (UID: \"6075b482-4926-4cf5-b632-0a731d2686dc\") " pod="openstack/cinder-scheduler-0" Dec 11 08:29:57 crc kubenswrapper[4860]: I1211 08:29:57.257042 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bsbt6\" (UniqueName: \"kubernetes.io/projected/6075b482-4926-4cf5-b632-0a731d2686dc-kube-api-access-bsbt6\") pod \"cinder-scheduler-0\" (UID: \"6075b482-4926-4cf5-b632-0a731d2686dc\") " pod="openstack/cinder-scheduler-0" Dec 11 08:29:57 crc kubenswrapper[4860]: I1211 08:29:57.257073 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6075b482-4926-4cf5-b632-0a731d2686dc-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"6075b482-4926-4cf5-b632-0a731d2686dc\") " pod="openstack/cinder-scheduler-0" Dec 11 08:29:57 crc kubenswrapper[4860]: I1211 08:29:57.257192 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6075b482-4926-4cf5-b632-0a731d2686dc-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"6075b482-4926-4cf5-b632-0a731d2686dc\") " pod="openstack/cinder-scheduler-0" Dec 11 08:29:57 crc kubenswrapper[4860]: I1211 08:29:57.257254 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6075b482-4926-4cf5-b632-0a731d2686dc-scripts\") pod \"cinder-scheduler-0\" (UID: \"6075b482-4926-4cf5-b632-0a731d2686dc\") " pod="openstack/cinder-scheduler-0" Dec 11 08:29:57 crc kubenswrapper[4860]: I1211 08:29:57.352550 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 11 08:29:57 crc kubenswrapper[4860]: I1211 08:29:57.354332 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 11 08:29:57 crc kubenswrapper[4860]: I1211 08:29:57.358885 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bsbt6\" (UniqueName: \"kubernetes.io/projected/6075b482-4926-4cf5-b632-0a731d2686dc-kube-api-access-bsbt6\") pod \"cinder-scheduler-0\" (UID: \"6075b482-4926-4cf5-b632-0a731d2686dc\") " pod="openstack/cinder-scheduler-0" Dec 11 08:29:57 crc kubenswrapper[4860]: I1211 08:29:57.358946 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6075b482-4926-4cf5-b632-0a731d2686dc-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"6075b482-4926-4cf5-b632-0a731d2686dc\") " pod="openstack/cinder-scheduler-0" Dec 11 08:29:57 crc kubenswrapper[4860]: I1211 08:29:57.358985 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4qtwf\" (UniqueName: \"kubernetes.io/projected/8ad63b0c-4ee1-4a57-9116-a8fed2dae193-kube-api-access-4qtwf\") pod \"dnsmasq-dns-6bb4fc677f-mcn2v\" (UID: \"8ad63b0c-4ee1-4a57-9116-a8fed2dae193\") " pod="openstack/dnsmasq-dns-6bb4fc677f-mcn2v" Dec 11 08:29:57 crc kubenswrapper[4860]: I1211 08:29:57.359012 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8ad63b0c-4ee1-4a57-9116-a8fed2dae193-dns-svc\") pod \"dnsmasq-dns-6bb4fc677f-mcn2v\" (UID: \"8ad63b0c-4ee1-4a57-9116-a8fed2dae193\") " pod="openstack/dnsmasq-dns-6bb4fc677f-mcn2v" Dec 11 08:29:57 crc kubenswrapper[4860]: I1211 08:29:57.359030 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8ad63b0c-4ee1-4a57-9116-a8fed2dae193-config\") pod \"dnsmasq-dns-6bb4fc677f-mcn2v\" (UID: \"8ad63b0c-4ee1-4a57-9116-a8fed2dae193\") " pod="openstack/dnsmasq-dns-6bb4fc677f-mcn2v" Dec 11 08:29:57 crc kubenswrapper[4860]: I1211 08:29:57.359056 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8ad63b0c-4ee1-4a57-9116-a8fed2dae193-ovsdbserver-sb\") pod \"dnsmasq-dns-6bb4fc677f-mcn2v\" (UID: \"8ad63b0c-4ee1-4a57-9116-a8fed2dae193\") " pod="openstack/dnsmasq-dns-6bb4fc677f-mcn2v" Dec 11 08:29:57 crc kubenswrapper[4860]: I1211 08:29:57.359086 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6075b482-4926-4cf5-b632-0a731d2686dc-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"6075b482-4926-4cf5-b632-0a731d2686dc\") " pod="openstack/cinder-scheduler-0" Dec 11 08:29:57 crc kubenswrapper[4860]: I1211 08:29:57.359130 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6075b482-4926-4cf5-b632-0a731d2686dc-scripts\") pod \"cinder-scheduler-0\" (UID: \"6075b482-4926-4cf5-b632-0a731d2686dc\") " pod="openstack/cinder-scheduler-0" Dec 11 08:29:57 crc kubenswrapper[4860]: I1211 08:29:57.359178 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6075b482-4926-4cf5-b632-0a731d2686dc-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"6075b482-4926-4cf5-b632-0a731d2686dc\") " pod="openstack/cinder-scheduler-0" Dec 11 08:29:57 crc kubenswrapper[4860]: I1211 08:29:57.359206 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8ad63b0c-4ee1-4a57-9116-a8fed2dae193-dns-swift-storage-0\") pod \"dnsmasq-dns-6bb4fc677f-mcn2v\" (UID: \"8ad63b0c-4ee1-4a57-9116-a8fed2dae193\") " pod="openstack/dnsmasq-dns-6bb4fc677f-mcn2v" Dec 11 08:29:57 crc kubenswrapper[4860]: I1211 08:29:57.359239 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6075b482-4926-4cf5-b632-0a731d2686dc-config-data\") pod \"cinder-scheduler-0\" (UID: \"6075b482-4926-4cf5-b632-0a731d2686dc\") " pod="openstack/cinder-scheduler-0" Dec 11 08:29:57 crc kubenswrapper[4860]: I1211 08:29:57.359272 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8ad63b0c-4ee1-4a57-9116-a8fed2dae193-ovsdbserver-nb\") pod \"dnsmasq-dns-6bb4fc677f-mcn2v\" (UID: \"8ad63b0c-4ee1-4a57-9116-a8fed2dae193\") " pod="openstack/dnsmasq-dns-6bb4fc677f-mcn2v" Dec 11 08:29:57 crc kubenswrapper[4860]: I1211 08:29:57.365050 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6075b482-4926-4cf5-b632-0a731d2686dc-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"6075b482-4926-4cf5-b632-0a731d2686dc\") " pod="openstack/cinder-scheduler-0" Dec 11 08:29:57 crc kubenswrapper[4860]: I1211 08:29:57.370817 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6075b482-4926-4cf5-b632-0a731d2686dc-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"6075b482-4926-4cf5-b632-0a731d2686dc\") " pod="openstack/cinder-scheduler-0" Dec 11 08:29:57 crc kubenswrapper[4860]: I1211 08:29:57.380711 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6075b482-4926-4cf5-b632-0a731d2686dc-config-data\") pod \"cinder-scheduler-0\" (UID: \"6075b482-4926-4cf5-b632-0a731d2686dc\") " pod="openstack/cinder-scheduler-0" Dec 11 08:29:57 crc kubenswrapper[4860]: I1211 08:29:57.386592 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 11 08:29:57 crc kubenswrapper[4860]: I1211 08:29:57.391232 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6075b482-4926-4cf5-b632-0a731d2686dc-scripts\") pod \"cinder-scheduler-0\" (UID: \"6075b482-4926-4cf5-b632-0a731d2686dc\") " pod="openstack/cinder-scheduler-0" Dec 11 08:29:57 crc kubenswrapper[4860]: I1211 08:29:57.391604 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 11 08:29:57 crc kubenswrapper[4860]: I1211 08:29:57.399713 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6075b482-4926-4cf5-b632-0a731d2686dc-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"6075b482-4926-4cf5-b632-0a731d2686dc\") " pod="openstack/cinder-scheduler-0" Dec 11 08:29:57 crc kubenswrapper[4860]: I1211 08:29:57.411043 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bsbt6\" (UniqueName: \"kubernetes.io/projected/6075b482-4926-4cf5-b632-0a731d2686dc-kube-api-access-bsbt6\") pod \"cinder-scheduler-0\" (UID: \"6075b482-4926-4cf5-b632-0a731d2686dc\") " pod="openstack/cinder-scheduler-0" Dec 11 08:29:57 crc kubenswrapper[4860]: I1211 08:29:57.506704 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/206264ab-1727-44fe-a196-c59d47b424ca-config-data-custom\") pod \"cinder-api-0\" (UID: \"206264ab-1727-44fe-a196-c59d47b424ca\") " pod="openstack/cinder-api-0" Dec 11 08:29:57 crc kubenswrapper[4860]: I1211 08:29:57.506767 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8ad63b0c-4ee1-4a57-9116-a8fed2dae193-dns-swift-storage-0\") pod \"dnsmasq-dns-6bb4fc677f-mcn2v\" (UID: \"8ad63b0c-4ee1-4a57-9116-a8fed2dae193\") " pod="openstack/dnsmasq-dns-6bb4fc677f-mcn2v" Dec 11 08:29:57 crc kubenswrapper[4860]: I1211 08:29:57.506808 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8ad63b0c-4ee1-4a57-9116-a8fed2dae193-ovsdbserver-nb\") pod \"dnsmasq-dns-6bb4fc677f-mcn2v\" (UID: \"8ad63b0c-4ee1-4a57-9116-a8fed2dae193\") " pod="openstack/dnsmasq-dns-6bb4fc677f-mcn2v" Dec 11 08:29:57 crc kubenswrapper[4860]: I1211 08:29:57.506829 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/206264ab-1727-44fe-a196-c59d47b424ca-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"206264ab-1727-44fe-a196-c59d47b424ca\") " pod="openstack/cinder-api-0" Dec 11 08:29:57 crc kubenswrapper[4860]: I1211 08:29:57.506869 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/206264ab-1727-44fe-a196-c59d47b424ca-config-data\") pod \"cinder-api-0\" (UID: \"206264ab-1727-44fe-a196-c59d47b424ca\") " pod="openstack/cinder-api-0" Dec 11 08:29:57 crc kubenswrapper[4860]: I1211 08:29:57.506898 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/206264ab-1727-44fe-a196-c59d47b424ca-scripts\") pod \"cinder-api-0\" (UID: \"206264ab-1727-44fe-a196-c59d47b424ca\") " pod="openstack/cinder-api-0" Dec 11 08:29:57 crc kubenswrapper[4860]: I1211 08:29:57.506958 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/206264ab-1727-44fe-a196-c59d47b424ca-logs\") pod \"cinder-api-0\" (UID: \"206264ab-1727-44fe-a196-c59d47b424ca\") " pod="openstack/cinder-api-0" Dec 11 08:29:57 crc kubenswrapper[4860]: I1211 08:29:57.507032 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4qtwf\" (UniqueName: \"kubernetes.io/projected/8ad63b0c-4ee1-4a57-9116-a8fed2dae193-kube-api-access-4qtwf\") pod \"dnsmasq-dns-6bb4fc677f-mcn2v\" (UID: \"8ad63b0c-4ee1-4a57-9116-a8fed2dae193\") " pod="openstack/dnsmasq-dns-6bb4fc677f-mcn2v" Dec 11 08:29:57 crc kubenswrapper[4860]: I1211 08:29:57.507061 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8ad63b0c-4ee1-4a57-9116-a8fed2dae193-dns-svc\") pod \"dnsmasq-dns-6bb4fc677f-mcn2v\" (UID: \"8ad63b0c-4ee1-4a57-9116-a8fed2dae193\") " pod="openstack/dnsmasq-dns-6bb4fc677f-mcn2v" Dec 11 08:29:57 crc kubenswrapper[4860]: I1211 08:29:57.507082 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8ad63b0c-4ee1-4a57-9116-a8fed2dae193-config\") pod \"dnsmasq-dns-6bb4fc677f-mcn2v\" (UID: \"8ad63b0c-4ee1-4a57-9116-a8fed2dae193\") " pod="openstack/dnsmasq-dns-6bb4fc677f-mcn2v" Dec 11 08:29:57 crc kubenswrapper[4860]: I1211 08:29:57.507107 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fg8qf\" (UniqueName: \"kubernetes.io/projected/206264ab-1727-44fe-a196-c59d47b424ca-kube-api-access-fg8qf\") pod \"cinder-api-0\" (UID: \"206264ab-1727-44fe-a196-c59d47b424ca\") " pod="openstack/cinder-api-0" Dec 11 08:29:57 crc kubenswrapper[4860]: I1211 08:29:57.507144 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8ad63b0c-4ee1-4a57-9116-a8fed2dae193-ovsdbserver-sb\") pod \"dnsmasq-dns-6bb4fc677f-mcn2v\" (UID: \"8ad63b0c-4ee1-4a57-9116-a8fed2dae193\") " pod="openstack/dnsmasq-dns-6bb4fc677f-mcn2v" Dec 11 08:29:57 crc kubenswrapper[4860]: I1211 08:29:57.507218 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/206264ab-1727-44fe-a196-c59d47b424ca-etc-machine-id\") pod \"cinder-api-0\" (UID: \"206264ab-1727-44fe-a196-c59d47b424ca\") " pod="openstack/cinder-api-0" Dec 11 08:29:57 crc kubenswrapper[4860]: I1211 08:29:57.521459 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8ad63b0c-4ee1-4a57-9116-a8fed2dae193-ovsdbserver-sb\") pod \"dnsmasq-dns-6bb4fc677f-mcn2v\" (UID: \"8ad63b0c-4ee1-4a57-9116-a8fed2dae193\") " pod="openstack/dnsmasq-dns-6bb4fc677f-mcn2v" Dec 11 08:29:57 crc kubenswrapper[4860]: I1211 08:29:57.522037 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8ad63b0c-4ee1-4a57-9116-a8fed2dae193-dns-swift-storage-0\") pod \"dnsmasq-dns-6bb4fc677f-mcn2v\" (UID: \"8ad63b0c-4ee1-4a57-9116-a8fed2dae193\") " pod="openstack/dnsmasq-dns-6bb4fc677f-mcn2v" Dec 11 08:29:57 crc kubenswrapper[4860]: I1211 08:29:57.524530 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8ad63b0c-4ee1-4a57-9116-a8fed2dae193-dns-svc\") pod \"dnsmasq-dns-6bb4fc677f-mcn2v\" (UID: \"8ad63b0c-4ee1-4a57-9116-a8fed2dae193\") " pod="openstack/dnsmasq-dns-6bb4fc677f-mcn2v" Dec 11 08:29:57 crc kubenswrapper[4860]: I1211 08:29:57.541065 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8ad63b0c-4ee1-4a57-9116-a8fed2dae193-config\") pod \"dnsmasq-dns-6bb4fc677f-mcn2v\" (UID: \"8ad63b0c-4ee1-4a57-9116-a8fed2dae193\") " pod="openstack/dnsmasq-dns-6bb4fc677f-mcn2v" Dec 11 08:29:57 crc kubenswrapper[4860]: I1211 08:29:57.541177 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 11 08:29:57 crc kubenswrapper[4860]: I1211 08:29:57.551183 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8ad63b0c-4ee1-4a57-9116-a8fed2dae193-ovsdbserver-nb\") pod \"dnsmasq-dns-6bb4fc677f-mcn2v\" (UID: \"8ad63b0c-4ee1-4a57-9116-a8fed2dae193\") " pod="openstack/dnsmasq-dns-6bb4fc677f-mcn2v" Dec 11 08:29:57 crc kubenswrapper[4860]: I1211 08:29:57.591671 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4qtwf\" (UniqueName: \"kubernetes.io/projected/8ad63b0c-4ee1-4a57-9116-a8fed2dae193-kube-api-access-4qtwf\") pod \"dnsmasq-dns-6bb4fc677f-mcn2v\" (UID: \"8ad63b0c-4ee1-4a57-9116-a8fed2dae193\") " pod="openstack/dnsmasq-dns-6bb4fc677f-mcn2v" Dec 11 08:29:57 crc kubenswrapper[4860]: I1211 08:29:57.622041 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/206264ab-1727-44fe-a196-c59d47b424ca-logs\") pod \"cinder-api-0\" (UID: \"206264ab-1727-44fe-a196-c59d47b424ca\") " pod="openstack/cinder-api-0" Dec 11 08:29:57 crc kubenswrapper[4860]: I1211 08:29:57.622109 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fg8qf\" (UniqueName: \"kubernetes.io/projected/206264ab-1727-44fe-a196-c59d47b424ca-kube-api-access-fg8qf\") pod \"cinder-api-0\" (UID: \"206264ab-1727-44fe-a196-c59d47b424ca\") " pod="openstack/cinder-api-0" Dec 11 08:29:57 crc kubenswrapper[4860]: I1211 08:29:57.622167 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/206264ab-1727-44fe-a196-c59d47b424ca-etc-machine-id\") pod \"cinder-api-0\" (UID: \"206264ab-1727-44fe-a196-c59d47b424ca\") " pod="openstack/cinder-api-0" Dec 11 08:29:57 crc kubenswrapper[4860]: I1211 08:29:57.622221 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/206264ab-1727-44fe-a196-c59d47b424ca-config-data-custom\") pod \"cinder-api-0\" (UID: \"206264ab-1727-44fe-a196-c59d47b424ca\") " pod="openstack/cinder-api-0" Dec 11 08:29:57 crc kubenswrapper[4860]: I1211 08:29:57.622248 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/206264ab-1727-44fe-a196-c59d47b424ca-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"206264ab-1727-44fe-a196-c59d47b424ca\") " pod="openstack/cinder-api-0" Dec 11 08:29:57 crc kubenswrapper[4860]: I1211 08:29:57.622268 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/206264ab-1727-44fe-a196-c59d47b424ca-config-data\") pod \"cinder-api-0\" (UID: \"206264ab-1727-44fe-a196-c59d47b424ca\") " pod="openstack/cinder-api-0" Dec 11 08:29:57 crc kubenswrapper[4860]: I1211 08:29:57.622288 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/206264ab-1727-44fe-a196-c59d47b424ca-scripts\") pod \"cinder-api-0\" (UID: \"206264ab-1727-44fe-a196-c59d47b424ca\") " pod="openstack/cinder-api-0" Dec 11 08:29:57 crc kubenswrapper[4860]: I1211 08:29:57.622760 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/206264ab-1727-44fe-a196-c59d47b424ca-etc-machine-id\") pod \"cinder-api-0\" (UID: \"206264ab-1727-44fe-a196-c59d47b424ca\") " pod="openstack/cinder-api-0" Dec 11 08:29:57 crc kubenswrapper[4860]: I1211 08:29:57.623067 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/206264ab-1727-44fe-a196-c59d47b424ca-logs\") pod \"cinder-api-0\" (UID: \"206264ab-1727-44fe-a196-c59d47b424ca\") " pod="openstack/cinder-api-0" Dec 11 08:29:57 crc kubenswrapper[4860]: I1211 08:29:57.632360 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/206264ab-1727-44fe-a196-c59d47b424ca-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"206264ab-1727-44fe-a196-c59d47b424ca\") " pod="openstack/cinder-api-0" Dec 11 08:29:57 crc kubenswrapper[4860]: I1211 08:29:57.649749 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/206264ab-1727-44fe-a196-c59d47b424ca-config-data\") pod \"cinder-api-0\" (UID: \"206264ab-1727-44fe-a196-c59d47b424ca\") " pod="openstack/cinder-api-0" Dec 11 08:29:57 crc kubenswrapper[4860]: I1211 08:29:57.650550 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/206264ab-1727-44fe-a196-c59d47b424ca-scripts\") pod \"cinder-api-0\" (UID: \"206264ab-1727-44fe-a196-c59d47b424ca\") " pod="openstack/cinder-api-0" Dec 11 08:29:57 crc kubenswrapper[4860]: I1211 08:29:57.673693 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fg8qf\" (UniqueName: \"kubernetes.io/projected/206264ab-1727-44fe-a196-c59d47b424ca-kube-api-access-fg8qf\") pod \"cinder-api-0\" (UID: \"206264ab-1727-44fe-a196-c59d47b424ca\") " pod="openstack/cinder-api-0" Dec 11 08:29:57 crc kubenswrapper[4860]: I1211 08:29:57.698382 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/206264ab-1727-44fe-a196-c59d47b424ca-config-data-custom\") pod \"cinder-api-0\" (UID: \"206264ab-1727-44fe-a196-c59d47b424ca\") " pod="openstack/cinder-api-0" Dec 11 08:29:57 crc kubenswrapper[4860]: I1211 08:29:57.870839 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bb4fc677f-mcn2v" Dec 11 08:29:57 crc kubenswrapper[4860]: I1211 08:29:57.872443 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 11 08:29:58 crc kubenswrapper[4860]: I1211 08:29:58.447876 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 11 08:29:58 crc kubenswrapper[4860]: I1211 08:29:58.447936 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 11 08:29:58 crc kubenswrapper[4860]: I1211 08:29:58.485755 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 11 08:29:58 crc kubenswrapper[4860]: I1211 08:29:58.539511 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 11 08:29:58 crc kubenswrapper[4860]: I1211 08:29:58.659748 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-fcfdd6f9f-9vj6f" Dec 11 08:29:58 crc kubenswrapper[4860]: I1211 08:29:58.692590 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-9bc765b68-htbxs" Dec 11 08:29:58 crc kubenswrapper[4860]: I1211 08:29:58.749083 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b469f4ba-a16a-4425-b2d6-6d95be8dcc0d-dns-svc\") pod \"b469f4ba-a16a-4425-b2d6-6d95be8dcc0d\" (UID: \"b469f4ba-a16a-4425-b2d6-6d95be8dcc0d\") " Dec 11 08:29:58 crc kubenswrapper[4860]: I1211 08:29:58.749119 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0e28c0dc-a169-423c-ae52-4392a9a27a3e-logs\") pod \"0e28c0dc-a169-423c-ae52-4392a9a27a3e\" (UID: \"0e28c0dc-a169-423c-ae52-4392a9a27a3e\") " Dec 11 08:29:58 crc kubenswrapper[4860]: I1211 08:29:58.749190 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b469f4ba-a16a-4425-b2d6-6d95be8dcc0d-ovsdbserver-sb\") pod \"b469f4ba-a16a-4425-b2d6-6d95be8dcc0d\" (UID: \"b469f4ba-a16a-4425-b2d6-6d95be8dcc0d\") " Dec 11 08:29:58 crc kubenswrapper[4860]: I1211 08:29:58.749270 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mxrqv\" (UniqueName: \"kubernetes.io/projected/0e28c0dc-a169-423c-ae52-4392a9a27a3e-kube-api-access-mxrqv\") pod \"0e28c0dc-a169-423c-ae52-4392a9a27a3e\" (UID: \"0e28c0dc-a169-423c-ae52-4392a9a27a3e\") " Dec 11 08:29:58 crc kubenswrapper[4860]: I1211 08:29:58.749297 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e28c0dc-a169-423c-ae52-4392a9a27a3e-config-data\") pod \"0e28c0dc-a169-423c-ae52-4392a9a27a3e\" (UID: \"0e28c0dc-a169-423c-ae52-4392a9a27a3e\") " Dec 11 08:29:58 crc kubenswrapper[4860]: I1211 08:29:58.749317 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b469f4ba-a16a-4425-b2d6-6d95be8dcc0d-config\") pod \"b469f4ba-a16a-4425-b2d6-6d95be8dcc0d\" (UID: \"b469f4ba-a16a-4425-b2d6-6d95be8dcc0d\") " Dec 11 08:29:58 crc kubenswrapper[4860]: I1211 08:29:58.749399 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w86nj\" (UniqueName: \"kubernetes.io/projected/b469f4ba-a16a-4425-b2d6-6d95be8dcc0d-kube-api-access-w86nj\") pod \"b469f4ba-a16a-4425-b2d6-6d95be8dcc0d\" (UID: \"b469f4ba-a16a-4425-b2d6-6d95be8dcc0d\") " Dec 11 08:29:58 crc kubenswrapper[4860]: I1211 08:29:58.749429 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e28c0dc-a169-423c-ae52-4392a9a27a3e-combined-ca-bundle\") pod \"0e28c0dc-a169-423c-ae52-4392a9a27a3e\" (UID: \"0e28c0dc-a169-423c-ae52-4392a9a27a3e\") " Dec 11 08:29:58 crc kubenswrapper[4860]: I1211 08:29:58.749452 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0e28c0dc-a169-423c-ae52-4392a9a27a3e-config-data-custom\") pod \"0e28c0dc-a169-423c-ae52-4392a9a27a3e\" (UID: \"0e28c0dc-a169-423c-ae52-4392a9a27a3e\") " Dec 11 08:29:58 crc kubenswrapper[4860]: I1211 08:29:58.749490 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b469f4ba-a16a-4425-b2d6-6d95be8dcc0d-ovsdbserver-nb\") pod \"b469f4ba-a16a-4425-b2d6-6d95be8dcc0d\" (UID: \"b469f4ba-a16a-4425-b2d6-6d95be8dcc0d\") " Dec 11 08:29:58 crc kubenswrapper[4860]: I1211 08:29:58.749504 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b469f4ba-a16a-4425-b2d6-6d95be8dcc0d-dns-swift-storage-0\") pod \"b469f4ba-a16a-4425-b2d6-6d95be8dcc0d\" (UID: \"b469f4ba-a16a-4425-b2d6-6d95be8dcc0d\") " Dec 11 08:29:58 crc kubenswrapper[4860]: I1211 08:29:58.756227 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0e28c0dc-a169-423c-ae52-4392a9a27a3e-logs" (OuterVolumeSpecName: "logs") pod "0e28c0dc-a169-423c-ae52-4392a9a27a3e" (UID: "0e28c0dc-a169-423c-ae52-4392a9a27a3e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:29:58 crc kubenswrapper[4860]: I1211 08:29:58.773904 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b469f4ba-a16a-4425-b2d6-6d95be8dcc0d-kube-api-access-w86nj" (OuterVolumeSpecName: "kube-api-access-w86nj") pod "b469f4ba-a16a-4425-b2d6-6d95be8dcc0d" (UID: "b469f4ba-a16a-4425-b2d6-6d95be8dcc0d"). InnerVolumeSpecName "kube-api-access-w86nj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:29:58 crc kubenswrapper[4860]: I1211 08:29:58.784239 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e28c0dc-a169-423c-ae52-4392a9a27a3e-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "0e28c0dc-a169-423c-ae52-4392a9a27a3e" (UID: "0e28c0dc-a169-423c-ae52-4392a9a27a3e"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:29:58 crc kubenswrapper[4860]: I1211 08:29:58.795276 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0e28c0dc-a169-423c-ae52-4392a9a27a3e-kube-api-access-mxrqv" (OuterVolumeSpecName: "kube-api-access-mxrqv") pod "0e28c0dc-a169-423c-ae52-4392a9a27a3e" (UID: "0e28c0dc-a169-423c-ae52-4392a9a27a3e"). InnerVolumeSpecName "kube-api-access-mxrqv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:29:58 crc kubenswrapper[4860]: I1211 08:29:58.842298 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-9bc765b68-htbxs" event={"ID":"0e28c0dc-a169-423c-ae52-4392a9a27a3e","Type":"ContainerDied","Data":"8c6029898ce57a7af7dbf7baf1cbf99727cf17fc76410859e7ea8d16235c4e0f"} Dec 11 08:29:58 crc kubenswrapper[4860]: I1211 08:29:58.842351 4860 scope.go:117] "RemoveContainer" containerID="9fd2099322d500e85f729d00e2769ccfc0be26e059a57bfba2f1a45d2245d434" Dec 11 08:29:58 crc kubenswrapper[4860]: I1211 08:29:58.842473 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-9bc765b68-htbxs" Dec 11 08:29:58 crc kubenswrapper[4860]: I1211 08:29:58.852969 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mxrqv\" (UniqueName: \"kubernetes.io/projected/0e28c0dc-a169-423c-ae52-4392a9a27a3e-kube-api-access-mxrqv\") on node \"crc\" DevicePath \"\"" Dec 11 08:29:58 crc kubenswrapper[4860]: I1211 08:29:58.853006 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w86nj\" (UniqueName: \"kubernetes.io/projected/b469f4ba-a16a-4425-b2d6-6d95be8dcc0d-kube-api-access-w86nj\") on node \"crc\" DevicePath \"\"" Dec 11 08:29:58 crc kubenswrapper[4860]: I1211 08:29:58.853020 4860 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0e28c0dc-a169-423c-ae52-4392a9a27a3e-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 11 08:29:58 crc kubenswrapper[4860]: I1211 08:29:58.853031 4860 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0e28c0dc-a169-423c-ae52-4392a9a27a3e-logs\") on node \"crc\" DevicePath \"\"" Dec 11 08:29:58 crc kubenswrapper[4860]: I1211 08:29:58.862120 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-fcfdd6f9f-9vj6f" event={"ID":"b469f4ba-a16a-4425-b2d6-6d95be8dcc0d","Type":"ContainerDied","Data":"c3e348dd7e58f1be91412928b9c09c6522ff3d83d2b6500bdd0733b2cc707285"} Dec 11 08:29:58 crc kubenswrapper[4860]: I1211 08:29:58.862168 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-fcfdd6f9f-9vj6f" Dec 11 08:29:58 crc kubenswrapper[4860]: I1211 08:29:58.862718 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 11 08:29:58 crc kubenswrapper[4860]: I1211 08:29:58.862743 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 11 08:29:58 crc kubenswrapper[4860]: I1211 08:29:58.878593 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b469f4ba-a16a-4425-b2d6-6d95be8dcc0d-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "b469f4ba-a16a-4425-b2d6-6d95be8dcc0d" (UID: "b469f4ba-a16a-4425-b2d6-6d95be8dcc0d"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:29:58 crc kubenswrapper[4860]: I1211 08:29:58.879254 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b469f4ba-a16a-4425-b2d6-6d95be8dcc0d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b469f4ba-a16a-4425-b2d6-6d95be8dcc0d" (UID: "b469f4ba-a16a-4425-b2d6-6d95be8dcc0d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:29:58 crc kubenswrapper[4860]: I1211 08:29:58.893825 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e28c0dc-a169-423c-ae52-4392a9a27a3e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0e28c0dc-a169-423c-ae52-4392a9a27a3e" (UID: "0e28c0dc-a169-423c-ae52-4392a9a27a3e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:29:58 crc kubenswrapper[4860]: I1211 08:29:58.935986 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b469f4ba-a16a-4425-b2d6-6d95be8dcc0d-config" (OuterVolumeSpecName: "config") pod "b469f4ba-a16a-4425-b2d6-6d95be8dcc0d" (UID: "b469f4ba-a16a-4425-b2d6-6d95be8dcc0d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:29:58 crc kubenswrapper[4860]: I1211 08:29:58.943051 4860 scope.go:117] "RemoveContainer" containerID="1a39faaae155e7004b855c4ae10bd43b2d7f19545e167115add4bd8275302ead" Dec 11 08:29:58 crc kubenswrapper[4860]: I1211 08:29:58.955182 4860 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b469f4ba-a16a-4425-b2d6-6d95be8dcc0d-config\") on node \"crc\" DevicePath \"\"" Dec 11 08:29:58 crc kubenswrapper[4860]: I1211 08:29:58.957263 4860 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e28c0dc-a169-423c-ae52-4392a9a27a3e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 08:29:58 crc kubenswrapper[4860]: I1211 08:29:58.957354 4860 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b469f4ba-a16a-4425-b2d6-6d95be8dcc0d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 11 08:29:58 crc kubenswrapper[4860]: I1211 08:29:58.957581 4860 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b469f4ba-a16a-4425-b2d6-6d95be8dcc0d-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 11 08:29:58 crc kubenswrapper[4860]: I1211 08:29:58.973208 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b469f4ba-a16a-4425-b2d6-6d95be8dcc0d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b469f4ba-a16a-4425-b2d6-6d95be8dcc0d" (UID: "b469f4ba-a16a-4425-b2d6-6d95be8dcc0d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:29:58 crc kubenswrapper[4860]: I1211 08:29:58.978896 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b469f4ba-a16a-4425-b2d6-6d95be8dcc0d-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b469f4ba-a16a-4425-b2d6-6d95be8dcc0d" (UID: "b469f4ba-a16a-4425-b2d6-6d95be8dcc0d"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:29:58 crc kubenswrapper[4860]: I1211 08:29:58.982073 4860 scope.go:117] "RemoveContainer" containerID="8e4b1bc15563056317722c20ddb943a307f432b3ab351a7aa0200777cd3ea1c5" Dec 11 08:29:58 crc kubenswrapper[4860]: I1211 08:29:58.995150 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e28c0dc-a169-423c-ae52-4392a9a27a3e-config-data" (OuterVolumeSpecName: "config-data") pod "0e28c0dc-a169-423c-ae52-4392a9a27a3e" (UID: "0e28c0dc-a169-423c-ae52-4392a9a27a3e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:29:59 crc kubenswrapper[4860]: E1211 08:29:59.010167 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ceilometer-0" podUID="040a0e8c-efd3-4e3b-be51-59e311ae4406" Dec 11 08:29:59 crc kubenswrapper[4860]: I1211 08:29:59.059044 4860 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e28c0dc-a169-423c-ae52-4392a9a27a3e-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 08:29:59 crc kubenswrapper[4860]: I1211 08:29:59.059084 4860 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b469f4ba-a16a-4425-b2d6-6d95be8dcc0d-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 11 08:29:59 crc kubenswrapper[4860]: I1211 08:29:59.059131 4860 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b469f4ba-a16a-4425-b2d6-6d95be8dcc0d-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 11 08:29:59 crc kubenswrapper[4860]: I1211 08:29:59.102766 4860 scope.go:117] "RemoveContainer" containerID="2812c3cdf32d7175891d726a1e9edd0bf65a84cf2104bcd2d114a5b27c8145ed" Dec 11 08:29:59 crc kubenswrapper[4860]: I1211 08:29:59.231286 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-9bc765b68-htbxs"] Dec 11 08:29:59 crc kubenswrapper[4860]: I1211 08:29:59.244001 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-9bc765b68-htbxs"] Dec 11 08:29:59 crc kubenswrapper[4860]: I1211 08:29:59.294459 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 11 08:29:59 crc kubenswrapper[4860]: I1211 08:29:59.316848 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 11 08:29:59 crc kubenswrapper[4860]: I1211 08:29:59.333802 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-fcfdd6f9f-9vj6f"] Dec 11 08:29:59 crc kubenswrapper[4860]: I1211 08:29:59.365796 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-fcfdd6f9f-9vj6f"] Dec 11 08:29:59 crc kubenswrapper[4860]: I1211 08:29:59.378246 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 11 08:29:59 crc kubenswrapper[4860]: I1211 08:29:59.511387 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6bb4fc677f-mcn2v"] Dec 11 08:29:59 crc kubenswrapper[4860]: W1211 08:29:59.512251 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8ad63b0c_4ee1_4a57_9116_a8fed2dae193.slice/crio-441f8a1c9c0169f1a8f3ea5b7181e356c2b01e36ab18ac1e1249a121676b54ce WatchSource:0}: Error finding container 441f8a1c9c0169f1a8f3ea5b7181e356c2b01e36ab18ac1e1249a121676b54ce: Status 404 returned error can't find the container with id 441f8a1c9c0169f1a8f3ea5b7181e356c2b01e36ab18ac1e1249a121676b54ce Dec 11 08:29:59 crc kubenswrapper[4860]: I1211 08:29:59.598977 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0e28c0dc-a169-423c-ae52-4392a9a27a3e" path="/var/lib/kubelet/pods/0e28c0dc-a169-423c-ae52-4392a9a27a3e/volumes" Dec 11 08:29:59 crc kubenswrapper[4860]: I1211 08:29:59.599746 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b469f4ba-a16a-4425-b2d6-6d95be8dcc0d" path="/var/lib/kubelet/pods/b469f4ba-a16a-4425-b2d6-6d95be8dcc0d/volumes" Dec 11 08:29:59 crc kubenswrapper[4860]: I1211 08:29:59.876115 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"040a0e8c-efd3-4e3b-be51-59e311ae4406","Type":"ContainerStarted","Data":"934878d249a946b8d22b5cd89ec6fac0d221ad3772ada43066a6d86fd0a35514"} Dec 11 08:29:59 crc kubenswrapper[4860]: I1211 08:29:59.876215 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="040a0e8c-efd3-4e3b-be51-59e311ae4406" containerName="ceilometer-notification-agent" containerID="cri-o://2c85cf7db795596f54f14c6ed32043967f06c05f5cab51172fbd00666aa45279" gracePeriod=30 Dec 11 08:29:59 crc kubenswrapper[4860]: I1211 08:29:59.876287 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="040a0e8c-efd3-4e3b-be51-59e311ae4406" containerName="proxy-httpd" containerID="cri-o://934878d249a946b8d22b5cd89ec6fac0d221ad3772ada43066a6d86fd0a35514" gracePeriod=30 Dec 11 08:29:59 crc kubenswrapper[4860]: I1211 08:29:59.876372 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="040a0e8c-efd3-4e3b-be51-59e311ae4406" containerName="sg-core" containerID="cri-o://6462b2d96ccced3f64b359817e1219e8bb94873c821a2372070dd0c60f426b68" gracePeriod=30 Dec 11 08:29:59 crc kubenswrapper[4860]: I1211 08:29:59.876900 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 11 08:29:59 crc kubenswrapper[4860]: I1211 08:29:59.885583 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"6075b482-4926-4cf5-b632-0a731d2686dc","Type":"ContainerStarted","Data":"7032faeaef7a9f5f18ba1e486ffed6c02c2057f931f2ef0af05bf26658f60558"} Dec 11 08:29:59 crc kubenswrapper[4860]: I1211 08:29:59.891268 4860 generic.go:334] "Generic (PLEG): container finished" podID="8ad63b0c-4ee1-4a57-9116-a8fed2dae193" containerID="98e328ef2d1f54bfc06f425726808d1fccb7c13163c8851db70e87e836ba8ec5" exitCode=0 Dec 11 08:29:59 crc kubenswrapper[4860]: I1211 08:29:59.891355 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bb4fc677f-mcn2v" event={"ID":"8ad63b0c-4ee1-4a57-9116-a8fed2dae193","Type":"ContainerDied","Data":"98e328ef2d1f54bfc06f425726808d1fccb7c13163c8851db70e87e836ba8ec5"} Dec 11 08:29:59 crc kubenswrapper[4860]: I1211 08:29:59.891614 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bb4fc677f-mcn2v" event={"ID":"8ad63b0c-4ee1-4a57-9116-a8fed2dae193","Type":"ContainerStarted","Data":"441f8a1c9c0169f1a8f3ea5b7181e356c2b01e36ab18ac1e1249a121676b54ce"} Dec 11 08:29:59 crc kubenswrapper[4860]: I1211 08:29:59.897449 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"206264ab-1727-44fe-a196-c59d47b424ca","Type":"ContainerStarted","Data":"47478a04c990911deca9a4a68e051799dd5cdcf0e5b4d144e8b4043fccf47688"} Dec 11 08:30:00 crc kubenswrapper[4860]: I1211 08:30:00.159388 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29424030-m6tp2"] Dec 11 08:30:00 crc kubenswrapper[4860]: E1211 08:30:00.160082 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e28c0dc-a169-423c-ae52-4392a9a27a3e" containerName="barbican-api-log" Dec 11 08:30:00 crc kubenswrapper[4860]: I1211 08:30:00.160108 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e28c0dc-a169-423c-ae52-4392a9a27a3e" containerName="barbican-api-log" Dec 11 08:30:00 crc kubenswrapper[4860]: E1211 08:30:00.160127 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e28c0dc-a169-423c-ae52-4392a9a27a3e" containerName="barbican-api" Dec 11 08:30:00 crc kubenswrapper[4860]: I1211 08:30:00.160138 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e28c0dc-a169-423c-ae52-4392a9a27a3e" containerName="barbican-api" Dec 11 08:30:00 crc kubenswrapper[4860]: E1211 08:30:00.160158 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b469f4ba-a16a-4425-b2d6-6d95be8dcc0d" containerName="init" Dec 11 08:30:00 crc kubenswrapper[4860]: I1211 08:30:00.160165 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="b469f4ba-a16a-4425-b2d6-6d95be8dcc0d" containerName="init" Dec 11 08:30:00 crc kubenswrapper[4860]: E1211 08:30:00.160195 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b469f4ba-a16a-4425-b2d6-6d95be8dcc0d" containerName="dnsmasq-dns" Dec 11 08:30:00 crc kubenswrapper[4860]: I1211 08:30:00.160202 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="b469f4ba-a16a-4425-b2d6-6d95be8dcc0d" containerName="dnsmasq-dns" Dec 11 08:30:00 crc kubenswrapper[4860]: I1211 08:30:00.160431 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e28c0dc-a169-423c-ae52-4392a9a27a3e" containerName="barbican-api-log" Dec 11 08:30:00 crc kubenswrapper[4860]: I1211 08:30:00.160462 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="b469f4ba-a16a-4425-b2d6-6d95be8dcc0d" containerName="dnsmasq-dns" Dec 11 08:30:00 crc kubenswrapper[4860]: I1211 08:30:00.160475 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e28c0dc-a169-423c-ae52-4392a9a27a3e" containerName="barbican-api" Dec 11 08:30:00 crc kubenswrapper[4860]: I1211 08:30:00.161261 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29424030-m6tp2" Dec 11 08:30:00 crc kubenswrapper[4860]: I1211 08:30:00.164765 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 11 08:30:00 crc kubenswrapper[4860]: I1211 08:30:00.164974 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 11 08:30:00 crc kubenswrapper[4860]: I1211 08:30:00.190324 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29424030-m6tp2"] Dec 11 08:30:00 crc kubenswrapper[4860]: I1211 08:30:00.204708 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5330e7d4-9a3d-41cc-9f94-e17ad44f2080-secret-volume\") pod \"collect-profiles-29424030-m6tp2\" (UID: \"5330e7d4-9a3d-41cc-9f94-e17ad44f2080\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424030-m6tp2" Dec 11 08:30:00 crc kubenswrapper[4860]: I1211 08:30:00.205220 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5330e7d4-9a3d-41cc-9f94-e17ad44f2080-config-volume\") pod \"collect-profiles-29424030-m6tp2\" (UID: \"5330e7d4-9a3d-41cc-9f94-e17ad44f2080\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424030-m6tp2" Dec 11 08:30:00 crc kubenswrapper[4860]: I1211 08:30:00.205342 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nlx6c\" (UniqueName: \"kubernetes.io/projected/5330e7d4-9a3d-41cc-9f94-e17ad44f2080-kube-api-access-nlx6c\") pod \"collect-profiles-29424030-m6tp2\" (UID: \"5330e7d4-9a3d-41cc-9f94-e17ad44f2080\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424030-m6tp2" Dec 11 08:30:00 crc kubenswrapper[4860]: I1211 08:30:00.307397 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5330e7d4-9a3d-41cc-9f94-e17ad44f2080-secret-volume\") pod \"collect-profiles-29424030-m6tp2\" (UID: \"5330e7d4-9a3d-41cc-9f94-e17ad44f2080\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424030-m6tp2" Dec 11 08:30:00 crc kubenswrapper[4860]: I1211 08:30:00.307759 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5330e7d4-9a3d-41cc-9f94-e17ad44f2080-config-volume\") pod \"collect-profiles-29424030-m6tp2\" (UID: \"5330e7d4-9a3d-41cc-9f94-e17ad44f2080\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424030-m6tp2" Dec 11 08:30:00 crc kubenswrapper[4860]: I1211 08:30:00.307782 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nlx6c\" (UniqueName: \"kubernetes.io/projected/5330e7d4-9a3d-41cc-9f94-e17ad44f2080-kube-api-access-nlx6c\") pod \"collect-profiles-29424030-m6tp2\" (UID: \"5330e7d4-9a3d-41cc-9f94-e17ad44f2080\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424030-m6tp2" Dec 11 08:30:00 crc kubenswrapper[4860]: I1211 08:30:00.308886 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5330e7d4-9a3d-41cc-9f94-e17ad44f2080-config-volume\") pod \"collect-profiles-29424030-m6tp2\" (UID: \"5330e7d4-9a3d-41cc-9f94-e17ad44f2080\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424030-m6tp2" Dec 11 08:30:00 crc kubenswrapper[4860]: I1211 08:30:00.313904 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5330e7d4-9a3d-41cc-9f94-e17ad44f2080-secret-volume\") pod \"collect-profiles-29424030-m6tp2\" (UID: \"5330e7d4-9a3d-41cc-9f94-e17ad44f2080\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424030-m6tp2" Dec 11 08:30:00 crc kubenswrapper[4860]: I1211 08:30:00.327001 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nlx6c\" (UniqueName: \"kubernetes.io/projected/5330e7d4-9a3d-41cc-9f94-e17ad44f2080-kube-api-access-nlx6c\") pod \"collect-profiles-29424030-m6tp2\" (UID: \"5330e7d4-9a3d-41cc-9f94-e17ad44f2080\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424030-m6tp2" Dec 11 08:30:00 crc kubenswrapper[4860]: I1211 08:30:00.534051 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29424030-m6tp2" Dec 11 08:30:00 crc kubenswrapper[4860]: I1211 08:30:00.954535 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"206264ab-1727-44fe-a196-c59d47b424ca","Type":"ContainerStarted","Data":"1ba4aaf56f0fad18428713f7066c8d0e7c53383b5031c3d6ba63ce17a88224ea"} Dec 11 08:30:00 crc kubenswrapper[4860]: I1211 08:30:00.969696 4860 generic.go:334] "Generic (PLEG): container finished" podID="040a0e8c-efd3-4e3b-be51-59e311ae4406" containerID="934878d249a946b8d22b5cd89ec6fac0d221ad3772ada43066a6d86fd0a35514" exitCode=0 Dec 11 08:30:00 crc kubenswrapper[4860]: I1211 08:30:00.969734 4860 generic.go:334] "Generic (PLEG): container finished" podID="040a0e8c-efd3-4e3b-be51-59e311ae4406" containerID="6462b2d96ccced3f64b359817e1219e8bb94873c821a2372070dd0c60f426b68" exitCode=2 Dec 11 08:30:00 crc kubenswrapper[4860]: I1211 08:30:00.969815 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"040a0e8c-efd3-4e3b-be51-59e311ae4406","Type":"ContainerDied","Data":"934878d249a946b8d22b5cd89ec6fac0d221ad3772ada43066a6d86fd0a35514"} Dec 11 08:30:00 crc kubenswrapper[4860]: I1211 08:30:00.969862 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"040a0e8c-efd3-4e3b-be51-59e311ae4406","Type":"ContainerDied","Data":"6462b2d96ccced3f64b359817e1219e8bb94873c821a2372070dd0c60f426b68"} Dec 11 08:30:00 crc kubenswrapper[4860]: I1211 08:30:00.974580 4860 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 11 08:30:00 crc kubenswrapper[4860]: I1211 08:30:00.974599 4860 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 11 08:30:00 crc kubenswrapper[4860]: I1211 08:30:00.976172 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bb4fc677f-mcn2v" event={"ID":"8ad63b0c-4ee1-4a57-9116-a8fed2dae193","Type":"ContainerStarted","Data":"e0a65b2d44cd87deeec4155c7f61b065474241e76538ddbd2f6cb854d27d25d3"} Dec 11 08:30:00 crc kubenswrapper[4860]: I1211 08:30:00.976229 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6bb4fc677f-mcn2v" Dec 11 08:30:00 crc kubenswrapper[4860]: I1211 08:30:00.998577 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6bb4fc677f-mcn2v" podStartSLOduration=3.998558718 podStartE2EDuration="3.998558718s" podCreationTimestamp="2025-12-11 08:29:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:30:00.997363251 +0000 UTC m=+1133.725882326" watchObservedRunningTime="2025-12-11 08:30:00.998558718 +0000 UTC m=+1133.727077773" Dec 11 08:30:01 crc kubenswrapper[4860]: I1211 08:30:01.131514 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29424030-m6tp2"] Dec 11 08:30:01 crc kubenswrapper[4860]: W1211 08:30:01.142453 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5330e7d4_9a3d_41cc_9f94_e17ad44f2080.slice/crio-1e328701a9be52af56c0e87d5e786f770139206deb72c880fd8c0a4a5ec36b99 WatchSource:0}: Error finding container 1e328701a9be52af56c0e87d5e786f770139206deb72c880fd8c0a4a5ec36b99: Status 404 returned error can't find the container with id 1e328701a9be52af56c0e87d5e786f770139206deb72c880fd8c0a4a5ec36b99 Dec 11 08:30:01 crc kubenswrapper[4860]: I1211 08:30:01.864743 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 11 08:30:01 crc kubenswrapper[4860]: I1211 08:30:01.867635 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 11 08:30:01 crc kubenswrapper[4860]: I1211 08:30:01.987410 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"6075b482-4926-4cf5-b632-0a731d2686dc","Type":"ContainerStarted","Data":"cfac87528b0888c0e951b979a1fd4da21776995c318a81737d10b48d2d6886c6"} Dec 11 08:30:01 crc kubenswrapper[4860]: I1211 08:30:01.987774 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"6075b482-4926-4cf5-b632-0a731d2686dc","Type":"ContainerStarted","Data":"5c83bda1a7e1ac8d4f912f91a01f263592a538d5209e99c9a8bf509f9d11da72"} Dec 11 08:30:01 crc kubenswrapper[4860]: I1211 08:30:01.990255 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"206264ab-1727-44fe-a196-c59d47b424ca","Type":"ContainerStarted","Data":"049c4cdb859199817bd492464385cf21cba4affa02d61d9606f1b1a98a152336"} Dec 11 08:30:01 crc kubenswrapper[4860]: I1211 08:30:01.990377 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="206264ab-1727-44fe-a196-c59d47b424ca" containerName="cinder-api-log" containerID="cri-o://1ba4aaf56f0fad18428713f7066c8d0e7c53383b5031c3d6ba63ce17a88224ea" gracePeriod=30 Dec 11 08:30:01 crc kubenswrapper[4860]: I1211 08:30:01.990412 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="206264ab-1727-44fe-a196-c59d47b424ca" containerName="cinder-api" containerID="cri-o://049c4cdb859199817bd492464385cf21cba4affa02d61d9606f1b1a98a152336" gracePeriod=30 Dec 11 08:30:01 crc kubenswrapper[4860]: I1211 08:30:01.990405 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 11 08:30:01 crc kubenswrapper[4860]: I1211 08:30:01.997439 4860 generic.go:334] "Generic (PLEG): container finished" podID="5330e7d4-9a3d-41cc-9f94-e17ad44f2080" containerID="25be5020a4cf38879be0a9e4ba9401a639ce81df3089ca082995ed05ab6b1bd7" exitCode=0 Dec 11 08:30:01 crc kubenswrapper[4860]: I1211 08:30:01.998387 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29424030-m6tp2" event={"ID":"5330e7d4-9a3d-41cc-9f94-e17ad44f2080","Type":"ContainerDied","Data":"25be5020a4cf38879be0a9e4ba9401a639ce81df3089ca082995ed05ab6b1bd7"} Dec 11 08:30:01 crc kubenswrapper[4860]: I1211 08:30:01.998421 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29424030-m6tp2" event={"ID":"5330e7d4-9a3d-41cc-9f94-e17ad44f2080","Type":"ContainerStarted","Data":"1e328701a9be52af56c0e87d5e786f770139206deb72c880fd8c0a4a5ec36b99"} Dec 11 08:30:02 crc kubenswrapper[4860]: I1211 08:30:02.017166 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.912824603 podStartE2EDuration="5.017136233s" podCreationTimestamp="2025-12-11 08:29:57 +0000 UTC" firstStartedPulling="2025-12-11 08:29:59.263989979 +0000 UTC m=+1131.992509034" lastFinishedPulling="2025-12-11 08:30:00.368301599 +0000 UTC m=+1133.096820664" observedRunningTime="2025-12-11 08:30:02.012874989 +0000 UTC m=+1134.741394054" watchObservedRunningTime="2025-12-11 08:30:02.017136233 +0000 UTC m=+1134.745655288" Dec 11 08:30:02 crc kubenswrapper[4860]: I1211 08:30:02.112095 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-ccf7c6598-jcbgb" Dec 11 08:30:02 crc kubenswrapper[4860]: I1211 08:30:02.134940 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=5.134922924 podStartE2EDuration="5.134922924s" podCreationTimestamp="2025-12-11 08:29:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:30:02.06828099 +0000 UTC m=+1134.796800045" watchObservedRunningTime="2025-12-11 08:30:02.134922924 +0000 UTC m=+1134.863441969" Dec 11 08:30:02 crc kubenswrapper[4860]: I1211 08:30:02.542219 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 11 08:30:02 crc kubenswrapper[4860]: I1211 08:30:02.751478 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 11 08:30:02 crc kubenswrapper[4860]: I1211 08:30:02.783497 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fg8qf\" (UniqueName: \"kubernetes.io/projected/206264ab-1727-44fe-a196-c59d47b424ca-kube-api-access-fg8qf\") pod \"206264ab-1727-44fe-a196-c59d47b424ca\" (UID: \"206264ab-1727-44fe-a196-c59d47b424ca\") " Dec 11 08:30:02 crc kubenswrapper[4860]: I1211 08:30:02.783628 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/206264ab-1727-44fe-a196-c59d47b424ca-combined-ca-bundle\") pod \"206264ab-1727-44fe-a196-c59d47b424ca\" (UID: \"206264ab-1727-44fe-a196-c59d47b424ca\") " Dec 11 08:30:02 crc kubenswrapper[4860]: I1211 08:30:02.784139 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/206264ab-1727-44fe-a196-c59d47b424ca-logs\") pod \"206264ab-1727-44fe-a196-c59d47b424ca\" (UID: \"206264ab-1727-44fe-a196-c59d47b424ca\") " Dec 11 08:30:02 crc kubenswrapper[4860]: I1211 08:30:02.784173 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/206264ab-1727-44fe-a196-c59d47b424ca-etc-machine-id\") pod \"206264ab-1727-44fe-a196-c59d47b424ca\" (UID: \"206264ab-1727-44fe-a196-c59d47b424ca\") " Dec 11 08:30:02 crc kubenswrapper[4860]: I1211 08:30:02.784264 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/206264ab-1727-44fe-a196-c59d47b424ca-config-data\") pod \"206264ab-1727-44fe-a196-c59d47b424ca\" (UID: \"206264ab-1727-44fe-a196-c59d47b424ca\") " Dec 11 08:30:02 crc kubenswrapper[4860]: I1211 08:30:02.784317 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/206264ab-1727-44fe-a196-c59d47b424ca-config-data-custom\") pod \"206264ab-1727-44fe-a196-c59d47b424ca\" (UID: \"206264ab-1727-44fe-a196-c59d47b424ca\") " Dec 11 08:30:02 crc kubenswrapper[4860]: I1211 08:30:02.784361 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/206264ab-1727-44fe-a196-c59d47b424ca-scripts\") pod \"206264ab-1727-44fe-a196-c59d47b424ca\" (UID: \"206264ab-1727-44fe-a196-c59d47b424ca\") " Dec 11 08:30:02 crc kubenswrapper[4860]: I1211 08:30:02.784415 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/206264ab-1727-44fe-a196-c59d47b424ca-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "206264ab-1727-44fe-a196-c59d47b424ca" (UID: "206264ab-1727-44fe-a196-c59d47b424ca"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 08:30:02 crc kubenswrapper[4860]: I1211 08:30:02.784794 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/206264ab-1727-44fe-a196-c59d47b424ca-logs" (OuterVolumeSpecName: "logs") pod "206264ab-1727-44fe-a196-c59d47b424ca" (UID: "206264ab-1727-44fe-a196-c59d47b424ca"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:30:02 crc kubenswrapper[4860]: I1211 08:30:02.785740 4860 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/206264ab-1727-44fe-a196-c59d47b424ca-logs\") on node \"crc\" DevicePath \"\"" Dec 11 08:30:02 crc kubenswrapper[4860]: I1211 08:30:02.785767 4860 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/206264ab-1727-44fe-a196-c59d47b424ca-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 11 08:30:02 crc kubenswrapper[4860]: I1211 08:30:02.790750 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/206264ab-1727-44fe-a196-c59d47b424ca-kube-api-access-fg8qf" (OuterVolumeSpecName: "kube-api-access-fg8qf") pod "206264ab-1727-44fe-a196-c59d47b424ca" (UID: "206264ab-1727-44fe-a196-c59d47b424ca"). InnerVolumeSpecName "kube-api-access-fg8qf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:30:02 crc kubenswrapper[4860]: I1211 08:30:02.801963 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/206264ab-1727-44fe-a196-c59d47b424ca-scripts" (OuterVolumeSpecName: "scripts") pod "206264ab-1727-44fe-a196-c59d47b424ca" (UID: "206264ab-1727-44fe-a196-c59d47b424ca"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:30:02 crc kubenswrapper[4860]: I1211 08:30:02.803797 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/206264ab-1727-44fe-a196-c59d47b424ca-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "206264ab-1727-44fe-a196-c59d47b424ca" (UID: "206264ab-1727-44fe-a196-c59d47b424ca"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:30:02 crc kubenswrapper[4860]: I1211 08:30:02.823807 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/206264ab-1727-44fe-a196-c59d47b424ca-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "206264ab-1727-44fe-a196-c59d47b424ca" (UID: "206264ab-1727-44fe-a196-c59d47b424ca"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:30:02 crc kubenswrapper[4860]: I1211 08:30:02.845202 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/206264ab-1727-44fe-a196-c59d47b424ca-config-data" (OuterVolumeSpecName: "config-data") pod "206264ab-1727-44fe-a196-c59d47b424ca" (UID: "206264ab-1727-44fe-a196-c59d47b424ca"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:30:02 crc kubenswrapper[4860]: I1211 08:30:02.887727 4860 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/206264ab-1727-44fe-a196-c59d47b424ca-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 08:30:02 crc kubenswrapper[4860]: I1211 08:30:02.887761 4860 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/206264ab-1727-44fe-a196-c59d47b424ca-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 11 08:30:02 crc kubenswrapper[4860]: I1211 08:30:02.887774 4860 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/206264ab-1727-44fe-a196-c59d47b424ca-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 08:30:02 crc kubenswrapper[4860]: I1211 08:30:02.887784 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fg8qf\" (UniqueName: \"kubernetes.io/projected/206264ab-1727-44fe-a196-c59d47b424ca-kube-api-access-fg8qf\") on node \"crc\" DevicePath \"\"" Dec 11 08:30:02 crc kubenswrapper[4860]: I1211 08:30:02.887794 4860 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/206264ab-1727-44fe-a196-c59d47b424ca-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 08:30:03 crc kubenswrapper[4860]: I1211 08:30:03.012036 4860 generic.go:334] "Generic (PLEG): container finished" podID="206264ab-1727-44fe-a196-c59d47b424ca" containerID="049c4cdb859199817bd492464385cf21cba4affa02d61d9606f1b1a98a152336" exitCode=0 Dec 11 08:30:03 crc kubenswrapper[4860]: I1211 08:30:03.012072 4860 generic.go:334] "Generic (PLEG): container finished" podID="206264ab-1727-44fe-a196-c59d47b424ca" containerID="1ba4aaf56f0fad18428713f7066c8d0e7c53383b5031c3d6ba63ce17a88224ea" exitCode=143 Dec 11 08:30:03 crc kubenswrapper[4860]: I1211 08:30:03.012100 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 11 08:30:03 crc kubenswrapper[4860]: I1211 08:30:03.012197 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"206264ab-1727-44fe-a196-c59d47b424ca","Type":"ContainerDied","Data":"049c4cdb859199817bd492464385cf21cba4affa02d61d9606f1b1a98a152336"} Dec 11 08:30:03 crc kubenswrapper[4860]: I1211 08:30:03.012225 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"206264ab-1727-44fe-a196-c59d47b424ca","Type":"ContainerDied","Data":"1ba4aaf56f0fad18428713f7066c8d0e7c53383b5031c3d6ba63ce17a88224ea"} Dec 11 08:30:03 crc kubenswrapper[4860]: I1211 08:30:03.012237 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"206264ab-1727-44fe-a196-c59d47b424ca","Type":"ContainerDied","Data":"47478a04c990911deca9a4a68e051799dd5cdcf0e5b4d144e8b4043fccf47688"} Dec 11 08:30:03 crc kubenswrapper[4860]: I1211 08:30:03.012252 4860 scope.go:117] "RemoveContainer" containerID="049c4cdb859199817bd492464385cf21cba4affa02d61d9606f1b1a98a152336" Dec 11 08:30:03 crc kubenswrapper[4860]: I1211 08:30:03.060824 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 11 08:30:03 crc kubenswrapper[4860]: I1211 08:30:03.061341 4860 scope.go:117] "RemoveContainer" containerID="1ba4aaf56f0fad18428713f7066c8d0e7c53383b5031c3d6ba63ce17a88224ea" Dec 11 08:30:03 crc kubenswrapper[4860]: I1211 08:30:03.075396 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Dec 11 08:30:03 crc kubenswrapper[4860]: I1211 08:30:03.091566 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 11 08:30:03 crc kubenswrapper[4860]: E1211 08:30:03.092206 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="206264ab-1727-44fe-a196-c59d47b424ca" containerName="cinder-api-log" Dec 11 08:30:03 crc kubenswrapper[4860]: I1211 08:30:03.092233 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="206264ab-1727-44fe-a196-c59d47b424ca" containerName="cinder-api-log" Dec 11 08:30:03 crc kubenswrapper[4860]: E1211 08:30:03.092253 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="206264ab-1727-44fe-a196-c59d47b424ca" containerName="cinder-api" Dec 11 08:30:03 crc kubenswrapper[4860]: I1211 08:30:03.092260 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="206264ab-1727-44fe-a196-c59d47b424ca" containerName="cinder-api" Dec 11 08:30:03 crc kubenswrapper[4860]: E1211 08:30:03.092274 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e28c0dc-a169-423c-ae52-4392a9a27a3e" containerName="barbican-api" Dec 11 08:30:03 crc kubenswrapper[4860]: I1211 08:30:03.092280 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e28c0dc-a169-423c-ae52-4392a9a27a3e" containerName="barbican-api" Dec 11 08:30:03 crc kubenswrapper[4860]: I1211 08:30:03.092482 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e28c0dc-a169-423c-ae52-4392a9a27a3e" containerName="barbican-api" Dec 11 08:30:03 crc kubenswrapper[4860]: I1211 08:30:03.092506 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="206264ab-1727-44fe-a196-c59d47b424ca" containerName="cinder-api" Dec 11 08:30:03 crc kubenswrapper[4860]: I1211 08:30:03.092523 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="206264ab-1727-44fe-a196-c59d47b424ca" containerName="cinder-api-log" Dec 11 08:30:03 crc kubenswrapper[4860]: I1211 08:30:03.093833 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 11 08:30:03 crc kubenswrapper[4860]: I1211 08:30:03.102045 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 11 08:30:03 crc kubenswrapper[4860]: I1211 08:30:03.102582 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Dec 11 08:30:03 crc kubenswrapper[4860]: I1211 08:30:03.102812 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 11 08:30:03 crc kubenswrapper[4860]: I1211 08:30:03.102981 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Dec 11 08:30:03 crc kubenswrapper[4860]: I1211 08:30:03.239045 4860 scope.go:117] "RemoveContainer" containerID="049c4cdb859199817bd492464385cf21cba4affa02d61d9606f1b1a98a152336" Dec 11 08:30:03 crc kubenswrapper[4860]: E1211 08:30:03.239739 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"049c4cdb859199817bd492464385cf21cba4affa02d61d9606f1b1a98a152336\": container with ID starting with 049c4cdb859199817bd492464385cf21cba4affa02d61d9606f1b1a98a152336 not found: ID does not exist" containerID="049c4cdb859199817bd492464385cf21cba4affa02d61d9606f1b1a98a152336" Dec 11 08:30:03 crc kubenswrapper[4860]: I1211 08:30:03.239792 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"049c4cdb859199817bd492464385cf21cba4affa02d61d9606f1b1a98a152336"} err="failed to get container status \"049c4cdb859199817bd492464385cf21cba4affa02d61d9606f1b1a98a152336\": rpc error: code = NotFound desc = could not find container \"049c4cdb859199817bd492464385cf21cba4affa02d61d9606f1b1a98a152336\": container with ID starting with 049c4cdb859199817bd492464385cf21cba4affa02d61d9606f1b1a98a152336 not found: ID does not exist" Dec 11 08:30:03 crc kubenswrapper[4860]: I1211 08:30:03.239841 4860 scope.go:117] "RemoveContainer" containerID="1ba4aaf56f0fad18428713f7066c8d0e7c53383b5031c3d6ba63ce17a88224ea" Dec 11 08:30:03 crc kubenswrapper[4860]: E1211 08:30:03.240152 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1ba4aaf56f0fad18428713f7066c8d0e7c53383b5031c3d6ba63ce17a88224ea\": container with ID starting with 1ba4aaf56f0fad18428713f7066c8d0e7c53383b5031c3d6ba63ce17a88224ea not found: ID does not exist" containerID="1ba4aaf56f0fad18428713f7066c8d0e7c53383b5031c3d6ba63ce17a88224ea" Dec 11 08:30:03 crc kubenswrapper[4860]: I1211 08:30:03.240177 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ba4aaf56f0fad18428713f7066c8d0e7c53383b5031c3d6ba63ce17a88224ea"} err="failed to get container status \"1ba4aaf56f0fad18428713f7066c8d0e7c53383b5031c3d6ba63ce17a88224ea\": rpc error: code = NotFound desc = could not find container \"1ba4aaf56f0fad18428713f7066c8d0e7c53383b5031c3d6ba63ce17a88224ea\": container with ID starting with 1ba4aaf56f0fad18428713f7066c8d0e7c53383b5031c3d6ba63ce17a88224ea not found: ID does not exist" Dec 11 08:30:03 crc kubenswrapper[4860]: I1211 08:30:03.240208 4860 scope.go:117] "RemoveContainer" containerID="049c4cdb859199817bd492464385cf21cba4affa02d61d9606f1b1a98a152336" Dec 11 08:30:03 crc kubenswrapper[4860]: I1211 08:30:03.240420 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"049c4cdb859199817bd492464385cf21cba4affa02d61d9606f1b1a98a152336"} err="failed to get container status \"049c4cdb859199817bd492464385cf21cba4affa02d61d9606f1b1a98a152336\": rpc error: code = NotFound desc = could not find container \"049c4cdb859199817bd492464385cf21cba4affa02d61d9606f1b1a98a152336\": container with ID starting with 049c4cdb859199817bd492464385cf21cba4affa02d61d9606f1b1a98a152336 not found: ID does not exist" Dec 11 08:30:03 crc kubenswrapper[4860]: I1211 08:30:03.240456 4860 scope.go:117] "RemoveContainer" containerID="1ba4aaf56f0fad18428713f7066c8d0e7c53383b5031c3d6ba63ce17a88224ea" Dec 11 08:30:03 crc kubenswrapper[4860]: I1211 08:30:03.240634 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ba4aaf56f0fad18428713f7066c8d0e7c53383b5031c3d6ba63ce17a88224ea"} err="failed to get container status \"1ba4aaf56f0fad18428713f7066c8d0e7c53383b5031c3d6ba63ce17a88224ea\": rpc error: code = NotFound desc = could not find container \"1ba4aaf56f0fad18428713f7066c8d0e7c53383b5031c3d6ba63ce17a88224ea\": container with ID starting with 1ba4aaf56f0fad18428713f7066c8d0e7c53383b5031c3d6ba63ce17a88224ea not found: ID does not exist" Dec 11 08:30:03 crc kubenswrapper[4860]: I1211 08:30:03.298071 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a6d65b6-5393-4edc-a9c3-3660c8860fd1-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"8a6d65b6-5393-4edc-a9c3-3660c8860fd1\") " pod="openstack/cinder-api-0" Dec 11 08:30:03 crc kubenswrapper[4860]: I1211 08:30:03.298225 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8a6d65b6-5393-4edc-a9c3-3660c8860fd1-etc-machine-id\") pod \"cinder-api-0\" (UID: \"8a6d65b6-5393-4edc-a9c3-3660c8860fd1\") " pod="openstack/cinder-api-0" Dec 11 08:30:03 crc kubenswrapper[4860]: I1211 08:30:03.298286 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8a6d65b6-5393-4edc-a9c3-3660c8860fd1-logs\") pod \"cinder-api-0\" (UID: \"8a6d65b6-5393-4edc-a9c3-3660c8860fd1\") " pod="openstack/cinder-api-0" Dec 11 08:30:03 crc kubenswrapper[4860]: I1211 08:30:03.298316 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a6d65b6-5393-4edc-a9c3-3660c8860fd1-public-tls-certs\") pod \"cinder-api-0\" (UID: \"8a6d65b6-5393-4edc-a9c3-3660c8860fd1\") " pod="openstack/cinder-api-0" Dec 11 08:30:03 crc kubenswrapper[4860]: I1211 08:30:03.298369 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a6d65b6-5393-4edc-a9c3-3660c8860fd1-config-data\") pod \"cinder-api-0\" (UID: \"8a6d65b6-5393-4edc-a9c3-3660c8860fd1\") " pod="openstack/cinder-api-0" Dec 11 08:30:03 crc kubenswrapper[4860]: I1211 08:30:03.298412 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a6d65b6-5393-4edc-a9c3-3660c8860fd1-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"8a6d65b6-5393-4edc-a9c3-3660c8860fd1\") " pod="openstack/cinder-api-0" Dec 11 08:30:03 crc kubenswrapper[4860]: I1211 08:30:03.298448 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8a6d65b6-5393-4edc-a9c3-3660c8860fd1-config-data-custom\") pod \"cinder-api-0\" (UID: \"8a6d65b6-5393-4edc-a9c3-3660c8860fd1\") " pod="openstack/cinder-api-0" Dec 11 08:30:03 crc kubenswrapper[4860]: I1211 08:30:03.298520 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a6d65b6-5393-4edc-a9c3-3660c8860fd1-scripts\") pod \"cinder-api-0\" (UID: \"8a6d65b6-5393-4edc-a9c3-3660c8860fd1\") " pod="openstack/cinder-api-0" Dec 11 08:30:03 crc kubenswrapper[4860]: I1211 08:30:03.298549 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dtp2g\" (UniqueName: \"kubernetes.io/projected/8a6d65b6-5393-4edc-a9c3-3660c8860fd1-kube-api-access-dtp2g\") pod \"cinder-api-0\" (UID: \"8a6d65b6-5393-4edc-a9c3-3660c8860fd1\") " pod="openstack/cinder-api-0" Dec 11 08:30:03 crc kubenswrapper[4860]: I1211 08:30:03.400493 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a6d65b6-5393-4edc-a9c3-3660c8860fd1-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"8a6d65b6-5393-4edc-a9c3-3660c8860fd1\") " pod="openstack/cinder-api-0" Dec 11 08:30:03 crc kubenswrapper[4860]: I1211 08:30:03.400572 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8a6d65b6-5393-4edc-a9c3-3660c8860fd1-etc-machine-id\") pod \"cinder-api-0\" (UID: \"8a6d65b6-5393-4edc-a9c3-3660c8860fd1\") " pod="openstack/cinder-api-0" Dec 11 08:30:03 crc kubenswrapper[4860]: I1211 08:30:03.400607 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8a6d65b6-5393-4edc-a9c3-3660c8860fd1-logs\") pod \"cinder-api-0\" (UID: \"8a6d65b6-5393-4edc-a9c3-3660c8860fd1\") " pod="openstack/cinder-api-0" Dec 11 08:30:03 crc kubenswrapper[4860]: I1211 08:30:03.400654 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a6d65b6-5393-4edc-a9c3-3660c8860fd1-public-tls-certs\") pod \"cinder-api-0\" (UID: \"8a6d65b6-5393-4edc-a9c3-3660c8860fd1\") " pod="openstack/cinder-api-0" Dec 11 08:30:03 crc kubenswrapper[4860]: I1211 08:30:03.400704 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a6d65b6-5393-4edc-a9c3-3660c8860fd1-config-data\") pod \"cinder-api-0\" (UID: \"8a6d65b6-5393-4edc-a9c3-3660c8860fd1\") " pod="openstack/cinder-api-0" Dec 11 08:30:03 crc kubenswrapper[4860]: I1211 08:30:03.400760 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a6d65b6-5393-4edc-a9c3-3660c8860fd1-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"8a6d65b6-5393-4edc-a9c3-3660c8860fd1\") " pod="openstack/cinder-api-0" Dec 11 08:30:03 crc kubenswrapper[4860]: I1211 08:30:03.400788 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8a6d65b6-5393-4edc-a9c3-3660c8860fd1-config-data-custom\") pod \"cinder-api-0\" (UID: \"8a6d65b6-5393-4edc-a9c3-3660c8860fd1\") " pod="openstack/cinder-api-0" Dec 11 08:30:03 crc kubenswrapper[4860]: I1211 08:30:03.400856 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a6d65b6-5393-4edc-a9c3-3660c8860fd1-scripts\") pod \"cinder-api-0\" (UID: \"8a6d65b6-5393-4edc-a9c3-3660c8860fd1\") " pod="openstack/cinder-api-0" Dec 11 08:30:03 crc kubenswrapper[4860]: I1211 08:30:03.400904 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dtp2g\" (UniqueName: \"kubernetes.io/projected/8a6d65b6-5393-4edc-a9c3-3660c8860fd1-kube-api-access-dtp2g\") pod \"cinder-api-0\" (UID: \"8a6d65b6-5393-4edc-a9c3-3660c8860fd1\") " pod="openstack/cinder-api-0" Dec 11 08:30:03 crc kubenswrapper[4860]: I1211 08:30:03.401405 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8a6d65b6-5393-4edc-a9c3-3660c8860fd1-etc-machine-id\") pod \"cinder-api-0\" (UID: \"8a6d65b6-5393-4edc-a9c3-3660c8860fd1\") " pod="openstack/cinder-api-0" Dec 11 08:30:03 crc kubenswrapper[4860]: I1211 08:30:03.401844 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8a6d65b6-5393-4edc-a9c3-3660c8860fd1-logs\") pod \"cinder-api-0\" (UID: \"8a6d65b6-5393-4edc-a9c3-3660c8860fd1\") " pod="openstack/cinder-api-0" Dec 11 08:30:03 crc kubenswrapper[4860]: I1211 08:30:03.416275 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a6d65b6-5393-4edc-a9c3-3660c8860fd1-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"8a6d65b6-5393-4edc-a9c3-3660c8860fd1\") " pod="openstack/cinder-api-0" Dec 11 08:30:03 crc kubenswrapper[4860]: I1211 08:30:03.432515 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a6d65b6-5393-4edc-a9c3-3660c8860fd1-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"8a6d65b6-5393-4edc-a9c3-3660c8860fd1\") " pod="openstack/cinder-api-0" Dec 11 08:30:03 crc kubenswrapper[4860]: I1211 08:30:03.434986 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a6d65b6-5393-4edc-a9c3-3660c8860fd1-config-data\") pod \"cinder-api-0\" (UID: \"8a6d65b6-5393-4edc-a9c3-3660c8860fd1\") " pod="openstack/cinder-api-0" Dec 11 08:30:03 crc kubenswrapper[4860]: I1211 08:30:03.435202 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a6d65b6-5393-4edc-a9c3-3660c8860fd1-scripts\") pod \"cinder-api-0\" (UID: \"8a6d65b6-5393-4edc-a9c3-3660c8860fd1\") " pod="openstack/cinder-api-0" Dec 11 08:30:03 crc kubenswrapper[4860]: I1211 08:30:03.435575 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8a6d65b6-5393-4edc-a9c3-3660c8860fd1-config-data-custom\") pod \"cinder-api-0\" (UID: \"8a6d65b6-5393-4edc-a9c3-3660c8860fd1\") " pod="openstack/cinder-api-0" Dec 11 08:30:03 crc kubenswrapper[4860]: I1211 08:30:03.450185 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dtp2g\" (UniqueName: \"kubernetes.io/projected/8a6d65b6-5393-4edc-a9c3-3660c8860fd1-kube-api-access-dtp2g\") pod \"cinder-api-0\" (UID: \"8a6d65b6-5393-4edc-a9c3-3660c8860fd1\") " pod="openstack/cinder-api-0" Dec 11 08:30:03 crc kubenswrapper[4860]: I1211 08:30:03.451063 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a6d65b6-5393-4edc-a9c3-3660c8860fd1-public-tls-certs\") pod \"cinder-api-0\" (UID: \"8a6d65b6-5393-4edc-a9c3-3660c8860fd1\") " pod="openstack/cinder-api-0" Dec 11 08:30:03 crc kubenswrapper[4860]: I1211 08:30:03.520393 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 11 08:30:03 crc kubenswrapper[4860]: I1211 08:30:03.597526 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="206264ab-1727-44fe-a196-c59d47b424ca" path="/var/lib/kubelet/pods/206264ab-1727-44fe-a196-c59d47b424ca/volumes" Dec 11 08:30:03 crc kubenswrapper[4860]: I1211 08:30:03.688789 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29424030-m6tp2" Dec 11 08:30:03 crc kubenswrapper[4860]: I1211 08:30:03.696528 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 11 08:30:03 crc kubenswrapper[4860]: I1211 08:30:03.826140 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/040a0e8c-efd3-4e3b-be51-59e311ae4406-config-data\") pod \"040a0e8c-efd3-4e3b-be51-59e311ae4406\" (UID: \"040a0e8c-efd3-4e3b-be51-59e311ae4406\") " Dec 11 08:30:03 crc kubenswrapper[4860]: I1211 08:30:03.826251 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vdbx8\" (UniqueName: \"kubernetes.io/projected/040a0e8c-efd3-4e3b-be51-59e311ae4406-kube-api-access-vdbx8\") pod \"040a0e8c-efd3-4e3b-be51-59e311ae4406\" (UID: \"040a0e8c-efd3-4e3b-be51-59e311ae4406\") " Dec 11 08:30:03 crc kubenswrapper[4860]: I1211 08:30:03.826300 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/040a0e8c-efd3-4e3b-be51-59e311ae4406-log-httpd\") pod \"040a0e8c-efd3-4e3b-be51-59e311ae4406\" (UID: \"040a0e8c-efd3-4e3b-be51-59e311ae4406\") " Dec 11 08:30:03 crc kubenswrapper[4860]: I1211 08:30:03.826326 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nlx6c\" (UniqueName: \"kubernetes.io/projected/5330e7d4-9a3d-41cc-9f94-e17ad44f2080-kube-api-access-nlx6c\") pod \"5330e7d4-9a3d-41cc-9f94-e17ad44f2080\" (UID: \"5330e7d4-9a3d-41cc-9f94-e17ad44f2080\") " Dec 11 08:30:03 crc kubenswrapper[4860]: I1211 08:30:03.826375 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/040a0e8c-efd3-4e3b-be51-59e311ae4406-run-httpd\") pod \"040a0e8c-efd3-4e3b-be51-59e311ae4406\" (UID: \"040a0e8c-efd3-4e3b-be51-59e311ae4406\") " Dec 11 08:30:03 crc kubenswrapper[4860]: I1211 08:30:03.826396 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/040a0e8c-efd3-4e3b-be51-59e311ae4406-combined-ca-bundle\") pod \"040a0e8c-efd3-4e3b-be51-59e311ae4406\" (UID: \"040a0e8c-efd3-4e3b-be51-59e311ae4406\") " Dec 11 08:30:03 crc kubenswrapper[4860]: I1211 08:30:03.826431 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5330e7d4-9a3d-41cc-9f94-e17ad44f2080-secret-volume\") pod \"5330e7d4-9a3d-41cc-9f94-e17ad44f2080\" (UID: \"5330e7d4-9a3d-41cc-9f94-e17ad44f2080\") " Dec 11 08:30:03 crc kubenswrapper[4860]: I1211 08:30:03.826452 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5330e7d4-9a3d-41cc-9f94-e17ad44f2080-config-volume\") pod \"5330e7d4-9a3d-41cc-9f94-e17ad44f2080\" (UID: \"5330e7d4-9a3d-41cc-9f94-e17ad44f2080\") " Dec 11 08:30:03 crc kubenswrapper[4860]: I1211 08:30:03.826553 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/040a0e8c-efd3-4e3b-be51-59e311ae4406-scripts\") pod \"040a0e8c-efd3-4e3b-be51-59e311ae4406\" (UID: \"040a0e8c-efd3-4e3b-be51-59e311ae4406\") " Dec 11 08:30:03 crc kubenswrapper[4860]: I1211 08:30:03.826624 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/040a0e8c-efd3-4e3b-be51-59e311ae4406-sg-core-conf-yaml\") pod \"040a0e8c-efd3-4e3b-be51-59e311ae4406\" (UID: \"040a0e8c-efd3-4e3b-be51-59e311ae4406\") " Dec 11 08:30:03 crc kubenswrapper[4860]: I1211 08:30:03.827195 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/040a0e8c-efd3-4e3b-be51-59e311ae4406-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "040a0e8c-efd3-4e3b-be51-59e311ae4406" (UID: "040a0e8c-efd3-4e3b-be51-59e311ae4406"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:30:03 crc kubenswrapper[4860]: I1211 08:30:03.827262 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/040a0e8c-efd3-4e3b-be51-59e311ae4406-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "040a0e8c-efd3-4e3b-be51-59e311ae4406" (UID: "040a0e8c-efd3-4e3b-be51-59e311ae4406"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:30:03 crc kubenswrapper[4860]: I1211 08:30:03.827752 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5330e7d4-9a3d-41cc-9f94-e17ad44f2080-config-volume" (OuterVolumeSpecName: "config-volume") pod "5330e7d4-9a3d-41cc-9f94-e17ad44f2080" (UID: "5330e7d4-9a3d-41cc-9f94-e17ad44f2080"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:30:03 crc kubenswrapper[4860]: I1211 08:30:03.833272 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5330e7d4-9a3d-41cc-9f94-e17ad44f2080-kube-api-access-nlx6c" (OuterVolumeSpecName: "kube-api-access-nlx6c") pod "5330e7d4-9a3d-41cc-9f94-e17ad44f2080" (UID: "5330e7d4-9a3d-41cc-9f94-e17ad44f2080"). InnerVolumeSpecName "kube-api-access-nlx6c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:30:03 crc kubenswrapper[4860]: I1211 08:30:03.833826 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5330e7d4-9a3d-41cc-9f94-e17ad44f2080-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "5330e7d4-9a3d-41cc-9f94-e17ad44f2080" (UID: "5330e7d4-9a3d-41cc-9f94-e17ad44f2080"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:30:03 crc kubenswrapper[4860]: I1211 08:30:03.835786 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/040a0e8c-efd3-4e3b-be51-59e311ae4406-scripts" (OuterVolumeSpecName: "scripts") pod "040a0e8c-efd3-4e3b-be51-59e311ae4406" (UID: "040a0e8c-efd3-4e3b-be51-59e311ae4406"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:30:03 crc kubenswrapper[4860]: I1211 08:30:03.836369 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/040a0e8c-efd3-4e3b-be51-59e311ae4406-kube-api-access-vdbx8" (OuterVolumeSpecName: "kube-api-access-vdbx8") pod "040a0e8c-efd3-4e3b-be51-59e311ae4406" (UID: "040a0e8c-efd3-4e3b-be51-59e311ae4406"). InnerVolumeSpecName "kube-api-access-vdbx8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:30:03 crc kubenswrapper[4860]: I1211 08:30:03.860533 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/040a0e8c-efd3-4e3b-be51-59e311ae4406-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "040a0e8c-efd3-4e3b-be51-59e311ae4406" (UID: "040a0e8c-efd3-4e3b-be51-59e311ae4406"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:30:03 crc kubenswrapper[4860]: I1211 08:30:03.881674 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/040a0e8c-efd3-4e3b-be51-59e311ae4406-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "040a0e8c-efd3-4e3b-be51-59e311ae4406" (UID: "040a0e8c-efd3-4e3b-be51-59e311ae4406"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:30:03 crc kubenswrapper[4860]: I1211 08:30:03.906452 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/040a0e8c-efd3-4e3b-be51-59e311ae4406-config-data" (OuterVolumeSpecName: "config-data") pod "040a0e8c-efd3-4e3b-be51-59e311ae4406" (UID: "040a0e8c-efd3-4e3b-be51-59e311ae4406"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:30:03 crc kubenswrapper[4860]: I1211 08:30:03.928885 4860 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/040a0e8c-efd3-4e3b-be51-59e311ae4406-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 11 08:30:03 crc kubenswrapper[4860]: I1211 08:30:03.928929 4860 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/040a0e8c-efd3-4e3b-be51-59e311ae4406-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 08:30:03 crc kubenswrapper[4860]: I1211 08:30:03.928940 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vdbx8\" (UniqueName: \"kubernetes.io/projected/040a0e8c-efd3-4e3b-be51-59e311ae4406-kube-api-access-vdbx8\") on node \"crc\" DevicePath \"\"" Dec 11 08:30:03 crc kubenswrapper[4860]: I1211 08:30:03.928953 4860 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/040a0e8c-efd3-4e3b-be51-59e311ae4406-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 11 08:30:03 crc kubenswrapper[4860]: I1211 08:30:03.928963 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nlx6c\" (UniqueName: \"kubernetes.io/projected/5330e7d4-9a3d-41cc-9f94-e17ad44f2080-kube-api-access-nlx6c\") on node \"crc\" DevicePath \"\"" Dec 11 08:30:03 crc kubenswrapper[4860]: I1211 08:30:03.928971 4860 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/040a0e8c-efd3-4e3b-be51-59e311ae4406-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 11 08:30:03 crc kubenswrapper[4860]: I1211 08:30:03.928979 4860 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/040a0e8c-efd3-4e3b-be51-59e311ae4406-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 08:30:03 crc kubenswrapper[4860]: I1211 08:30:03.928989 4860 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5330e7d4-9a3d-41cc-9f94-e17ad44f2080-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 11 08:30:03 crc kubenswrapper[4860]: I1211 08:30:03.928998 4860 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5330e7d4-9a3d-41cc-9f94-e17ad44f2080-config-volume\") on node \"crc\" DevicePath \"\"" Dec 11 08:30:03 crc kubenswrapper[4860]: I1211 08:30:03.929008 4860 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/040a0e8c-efd3-4e3b-be51-59e311ae4406-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 08:30:04 crc kubenswrapper[4860]: I1211 08:30:04.026979 4860 generic.go:334] "Generic (PLEG): container finished" podID="040a0e8c-efd3-4e3b-be51-59e311ae4406" containerID="2c85cf7db795596f54f14c6ed32043967f06c05f5cab51172fbd00666aa45279" exitCode=0 Dec 11 08:30:04 crc kubenswrapper[4860]: I1211 08:30:04.027045 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"040a0e8c-efd3-4e3b-be51-59e311ae4406","Type":"ContainerDied","Data":"2c85cf7db795596f54f14c6ed32043967f06c05f5cab51172fbd00666aa45279"} Dec 11 08:30:04 crc kubenswrapper[4860]: I1211 08:30:04.027073 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"040a0e8c-efd3-4e3b-be51-59e311ae4406","Type":"ContainerDied","Data":"445eca66a37a53e77187da28ab3dd3c55b55729539d52878dfbe12fabc64bdd3"} Dec 11 08:30:04 crc kubenswrapper[4860]: I1211 08:30:04.027093 4860 scope.go:117] "RemoveContainer" containerID="934878d249a946b8d22b5cd89ec6fac0d221ad3772ada43066a6d86fd0a35514" Dec 11 08:30:04 crc kubenswrapper[4860]: I1211 08:30:04.027214 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 11 08:30:04 crc kubenswrapper[4860]: I1211 08:30:04.032775 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 11 08:30:04 crc kubenswrapper[4860]: I1211 08:30:04.036825 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29424030-m6tp2" event={"ID":"5330e7d4-9a3d-41cc-9f94-e17ad44f2080","Type":"ContainerDied","Data":"1e328701a9be52af56c0e87d5e786f770139206deb72c880fd8c0a4a5ec36b99"} Dec 11 08:30:04 crc kubenswrapper[4860]: I1211 08:30:04.036874 4860 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1e328701a9be52af56c0e87d5e786f770139206deb72c880fd8c0a4a5ec36b99" Dec 11 08:30:04 crc kubenswrapper[4860]: I1211 08:30:04.036849 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29424030-m6tp2" Dec 11 08:30:04 crc kubenswrapper[4860]: I1211 08:30:04.082978 4860 scope.go:117] "RemoveContainer" containerID="6462b2d96ccced3f64b359817e1219e8bb94873c821a2372070dd0c60f426b68" Dec 11 08:30:04 crc kubenswrapper[4860]: I1211 08:30:04.090427 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 11 08:30:04 crc kubenswrapper[4860]: I1211 08:30:04.111720 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 11 08:30:04 crc kubenswrapper[4860]: I1211 08:30:04.121500 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 11 08:30:04 crc kubenswrapper[4860]: E1211 08:30:04.122026 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="040a0e8c-efd3-4e3b-be51-59e311ae4406" containerName="sg-core" Dec 11 08:30:04 crc kubenswrapper[4860]: I1211 08:30:04.122046 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="040a0e8c-efd3-4e3b-be51-59e311ae4406" containerName="sg-core" Dec 11 08:30:04 crc kubenswrapper[4860]: E1211 08:30:04.125256 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="040a0e8c-efd3-4e3b-be51-59e311ae4406" containerName="ceilometer-notification-agent" Dec 11 08:30:04 crc kubenswrapper[4860]: I1211 08:30:04.125299 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="040a0e8c-efd3-4e3b-be51-59e311ae4406" containerName="ceilometer-notification-agent" Dec 11 08:30:04 crc kubenswrapper[4860]: E1211 08:30:04.125329 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="040a0e8c-efd3-4e3b-be51-59e311ae4406" containerName="proxy-httpd" Dec 11 08:30:04 crc kubenswrapper[4860]: I1211 08:30:04.125340 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="040a0e8c-efd3-4e3b-be51-59e311ae4406" containerName="proxy-httpd" Dec 11 08:30:04 crc kubenswrapper[4860]: E1211 08:30:04.125359 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5330e7d4-9a3d-41cc-9f94-e17ad44f2080" containerName="collect-profiles" Dec 11 08:30:04 crc kubenswrapper[4860]: I1211 08:30:04.125369 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="5330e7d4-9a3d-41cc-9f94-e17ad44f2080" containerName="collect-profiles" Dec 11 08:30:04 crc kubenswrapper[4860]: I1211 08:30:04.126497 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="5330e7d4-9a3d-41cc-9f94-e17ad44f2080" containerName="collect-profiles" Dec 11 08:30:04 crc kubenswrapper[4860]: I1211 08:30:04.126533 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="040a0e8c-efd3-4e3b-be51-59e311ae4406" containerName="ceilometer-notification-agent" Dec 11 08:30:04 crc kubenswrapper[4860]: I1211 08:30:04.126546 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="040a0e8c-efd3-4e3b-be51-59e311ae4406" containerName="proxy-httpd" Dec 11 08:30:04 crc kubenswrapper[4860]: I1211 08:30:04.126560 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="040a0e8c-efd3-4e3b-be51-59e311ae4406" containerName="sg-core" Dec 11 08:30:04 crc kubenswrapper[4860]: I1211 08:30:04.147319 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 11 08:30:04 crc kubenswrapper[4860]: I1211 08:30:04.147605 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 11 08:30:04 crc kubenswrapper[4860]: I1211 08:30:04.151229 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 11 08:30:04 crc kubenswrapper[4860]: I1211 08:30:04.153041 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 11 08:30:04 crc kubenswrapper[4860]: I1211 08:30:04.240603 4860 scope.go:117] "RemoveContainer" containerID="2c85cf7db795596f54f14c6ed32043967f06c05f5cab51172fbd00666aa45279" Dec 11 08:30:04 crc kubenswrapper[4860]: I1211 08:30:04.261963 4860 scope.go:117] "RemoveContainer" containerID="934878d249a946b8d22b5cd89ec6fac0d221ad3772ada43066a6d86fd0a35514" Dec 11 08:30:04 crc kubenswrapper[4860]: E1211 08:30:04.262999 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"934878d249a946b8d22b5cd89ec6fac0d221ad3772ada43066a6d86fd0a35514\": container with ID starting with 934878d249a946b8d22b5cd89ec6fac0d221ad3772ada43066a6d86fd0a35514 not found: ID does not exist" containerID="934878d249a946b8d22b5cd89ec6fac0d221ad3772ada43066a6d86fd0a35514" Dec 11 08:30:04 crc kubenswrapper[4860]: I1211 08:30:04.263062 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"934878d249a946b8d22b5cd89ec6fac0d221ad3772ada43066a6d86fd0a35514"} err="failed to get container status \"934878d249a946b8d22b5cd89ec6fac0d221ad3772ada43066a6d86fd0a35514\": rpc error: code = NotFound desc = could not find container \"934878d249a946b8d22b5cd89ec6fac0d221ad3772ada43066a6d86fd0a35514\": container with ID starting with 934878d249a946b8d22b5cd89ec6fac0d221ad3772ada43066a6d86fd0a35514 not found: ID does not exist" Dec 11 08:30:04 crc kubenswrapper[4860]: I1211 08:30:04.263096 4860 scope.go:117] "RemoveContainer" containerID="6462b2d96ccced3f64b359817e1219e8bb94873c821a2372070dd0c60f426b68" Dec 11 08:30:04 crc kubenswrapper[4860]: E1211 08:30:04.263980 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6462b2d96ccced3f64b359817e1219e8bb94873c821a2372070dd0c60f426b68\": container with ID starting with 6462b2d96ccced3f64b359817e1219e8bb94873c821a2372070dd0c60f426b68 not found: ID does not exist" containerID="6462b2d96ccced3f64b359817e1219e8bb94873c821a2372070dd0c60f426b68" Dec 11 08:30:04 crc kubenswrapper[4860]: I1211 08:30:04.264017 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6462b2d96ccced3f64b359817e1219e8bb94873c821a2372070dd0c60f426b68"} err="failed to get container status \"6462b2d96ccced3f64b359817e1219e8bb94873c821a2372070dd0c60f426b68\": rpc error: code = NotFound desc = could not find container \"6462b2d96ccced3f64b359817e1219e8bb94873c821a2372070dd0c60f426b68\": container with ID starting with 6462b2d96ccced3f64b359817e1219e8bb94873c821a2372070dd0c60f426b68 not found: ID does not exist" Dec 11 08:30:04 crc kubenswrapper[4860]: I1211 08:30:04.264243 4860 scope.go:117] "RemoveContainer" containerID="2c85cf7db795596f54f14c6ed32043967f06c05f5cab51172fbd00666aa45279" Dec 11 08:30:04 crc kubenswrapper[4860]: E1211 08:30:04.264524 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2c85cf7db795596f54f14c6ed32043967f06c05f5cab51172fbd00666aa45279\": container with ID starting with 2c85cf7db795596f54f14c6ed32043967f06c05f5cab51172fbd00666aa45279 not found: ID does not exist" containerID="2c85cf7db795596f54f14c6ed32043967f06c05f5cab51172fbd00666aa45279" Dec 11 08:30:04 crc kubenswrapper[4860]: I1211 08:30:04.264551 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2c85cf7db795596f54f14c6ed32043967f06c05f5cab51172fbd00666aa45279"} err="failed to get container status \"2c85cf7db795596f54f14c6ed32043967f06c05f5cab51172fbd00666aa45279\": rpc error: code = NotFound desc = could not find container \"2c85cf7db795596f54f14c6ed32043967f06c05f5cab51172fbd00666aa45279\": container with ID starting with 2c85cf7db795596f54f14c6ed32043967f06c05f5cab51172fbd00666aa45279 not found: ID does not exist" Dec 11 08:30:04 crc kubenswrapper[4860]: I1211 08:30:04.276610 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2e2c517e-81fe-4563-889e-af154d1f500a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2e2c517e-81fe-4563-889e-af154d1f500a\") " pod="openstack/ceilometer-0" Dec 11 08:30:04 crc kubenswrapper[4860]: I1211 08:30:04.276710 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e2c517e-81fe-4563-889e-af154d1f500a-scripts\") pod \"ceilometer-0\" (UID: \"2e2c517e-81fe-4563-889e-af154d1f500a\") " pod="openstack/ceilometer-0" Dec 11 08:30:04 crc kubenswrapper[4860]: I1211 08:30:04.276756 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e2c517e-81fe-4563-889e-af154d1f500a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2e2c517e-81fe-4563-889e-af154d1f500a\") " pod="openstack/ceilometer-0" Dec 11 08:30:04 crc kubenswrapper[4860]: I1211 08:30:04.276924 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e2c517e-81fe-4563-889e-af154d1f500a-log-httpd\") pod \"ceilometer-0\" (UID: \"2e2c517e-81fe-4563-889e-af154d1f500a\") " pod="openstack/ceilometer-0" Dec 11 08:30:04 crc kubenswrapper[4860]: I1211 08:30:04.276976 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e2c517e-81fe-4563-889e-af154d1f500a-config-data\") pod \"ceilometer-0\" (UID: \"2e2c517e-81fe-4563-889e-af154d1f500a\") " pod="openstack/ceilometer-0" Dec 11 08:30:04 crc kubenswrapper[4860]: I1211 08:30:04.277024 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5q92p\" (UniqueName: \"kubernetes.io/projected/2e2c517e-81fe-4563-889e-af154d1f500a-kube-api-access-5q92p\") pod \"ceilometer-0\" (UID: \"2e2c517e-81fe-4563-889e-af154d1f500a\") " pod="openstack/ceilometer-0" Dec 11 08:30:04 crc kubenswrapper[4860]: I1211 08:30:04.277107 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e2c517e-81fe-4563-889e-af154d1f500a-run-httpd\") pod \"ceilometer-0\" (UID: \"2e2c517e-81fe-4563-889e-af154d1f500a\") " pod="openstack/ceilometer-0" Dec 11 08:30:04 crc kubenswrapper[4860]: I1211 08:30:04.378825 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e2c517e-81fe-4563-889e-af154d1f500a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2e2c517e-81fe-4563-889e-af154d1f500a\") " pod="openstack/ceilometer-0" Dec 11 08:30:04 crc kubenswrapper[4860]: I1211 08:30:04.378902 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e2c517e-81fe-4563-889e-af154d1f500a-log-httpd\") pod \"ceilometer-0\" (UID: \"2e2c517e-81fe-4563-889e-af154d1f500a\") " pod="openstack/ceilometer-0" Dec 11 08:30:04 crc kubenswrapper[4860]: I1211 08:30:04.378933 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e2c517e-81fe-4563-889e-af154d1f500a-config-data\") pod \"ceilometer-0\" (UID: \"2e2c517e-81fe-4563-889e-af154d1f500a\") " pod="openstack/ceilometer-0" Dec 11 08:30:04 crc kubenswrapper[4860]: I1211 08:30:04.378967 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5q92p\" (UniqueName: \"kubernetes.io/projected/2e2c517e-81fe-4563-889e-af154d1f500a-kube-api-access-5q92p\") pod \"ceilometer-0\" (UID: \"2e2c517e-81fe-4563-889e-af154d1f500a\") " pod="openstack/ceilometer-0" Dec 11 08:30:04 crc kubenswrapper[4860]: I1211 08:30:04.379014 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e2c517e-81fe-4563-889e-af154d1f500a-run-httpd\") pod \"ceilometer-0\" (UID: \"2e2c517e-81fe-4563-889e-af154d1f500a\") " pod="openstack/ceilometer-0" Dec 11 08:30:04 crc kubenswrapper[4860]: I1211 08:30:04.379097 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2e2c517e-81fe-4563-889e-af154d1f500a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2e2c517e-81fe-4563-889e-af154d1f500a\") " pod="openstack/ceilometer-0" Dec 11 08:30:04 crc kubenswrapper[4860]: I1211 08:30:04.379161 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e2c517e-81fe-4563-889e-af154d1f500a-scripts\") pod \"ceilometer-0\" (UID: \"2e2c517e-81fe-4563-889e-af154d1f500a\") " pod="openstack/ceilometer-0" Dec 11 08:30:04 crc kubenswrapper[4860]: I1211 08:30:04.381695 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e2c517e-81fe-4563-889e-af154d1f500a-log-httpd\") pod \"ceilometer-0\" (UID: \"2e2c517e-81fe-4563-889e-af154d1f500a\") " pod="openstack/ceilometer-0" Dec 11 08:30:04 crc kubenswrapper[4860]: I1211 08:30:04.382834 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e2c517e-81fe-4563-889e-af154d1f500a-run-httpd\") pod \"ceilometer-0\" (UID: \"2e2c517e-81fe-4563-889e-af154d1f500a\") " pod="openstack/ceilometer-0" Dec 11 08:30:04 crc kubenswrapper[4860]: I1211 08:30:04.384050 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e2c517e-81fe-4563-889e-af154d1f500a-config-data\") pod \"ceilometer-0\" (UID: \"2e2c517e-81fe-4563-889e-af154d1f500a\") " pod="openstack/ceilometer-0" Dec 11 08:30:04 crc kubenswrapper[4860]: I1211 08:30:04.384899 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e2c517e-81fe-4563-889e-af154d1f500a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2e2c517e-81fe-4563-889e-af154d1f500a\") " pod="openstack/ceilometer-0" Dec 11 08:30:04 crc kubenswrapper[4860]: I1211 08:30:04.385939 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2e2c517e-81fe-4563-889e-af154d1f500a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2e2c517e-81fe-4563-889e-af154d1f500a\") " pod="openstack/ceilometer-0" Dec 11 08:30:04 crc kubenswrapper[4860]: I1211 08:30:04.390461 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e2c517e-81fe-4563-889e-af154d1f500a-scripts\") pod \"ceilometer-0\" (UID: \"2e2c517e-81fe-4563-889e-af154d1f500a\") " pod="openstack/ceilometer-0" Dec 11 08:30:04 crc kubenswrapper[4860]: I1211 08:30:04.405557 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5q92p\" (UniqueName: \"kubernetes.io/projected/2e2c517e-81fe-4563-889e-af154d1f500a-kube-api-access-5q92p\") pod \"ceilometer-0\" (UID: \"2e2c517e-81fe-4563-889e-af154d1f500a\") " pod="openstack/ceilometer-0" Dec 11 08:30:04 crc kubenswrapper[4860]: I1211 08:30:04.545699 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 11 08:30:04 crc kubenswrapper[4860]: W1211 08:30:04.998447 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2e2c517e_81fe_4563_889e_af154d1f500a.slice/crio-d716be20ff6c5eb0ac20517b3c44bd25c32944509c239d09b7b182d5dbc5e195 WatchSource:0}: Error finding container d716be20ff6c5eb0ac20517b3c44bd25c32944509c239d09b7b182d5dbc5e195: Status 404 returned error can't find the container with id d716be20ff6c5eb0ac20517b3c44bd25c32944509c239d09b7b182d5dbc5e195 Dec 11 08:30:05 crc kubenswrapper[4860]: I1211 08:30:05.006056 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 11 08:30:05 crc kubenswrapper[4860]: I1211 08:30:05.048401 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"8a6d65b6-5393-4edc-a9c3-3660c8860fd1","Type":"ContainerStarted","Data":"1fbc757080740f18026b4bfe2f7b75d1295a902ccf7c0ddac19db214b5d89a72"} Dec 11 08:30:05 crc kubenswrapper[4860]: I1211 08:30:05.048449 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"8a6d65b6-5393-4edc-a9c3-3660c8860fd1","Type":"ContainerStarted","Data":"b1fe4a08edf010d0b8b6437b890ead593daa5883bfcc03625b42ca5d835fa278"} Dec 11 08:30:05 crc kubenswrapper[4860]: I1211 08:30:05.049987 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2e2c517e-81fe-4563-889e-af154d1f500a","Type":"ContainerStarted","Data":"d716be20ff6c5eb0ac20517b3c44bd25c32944509c239d09b7b182d5dbc5e195"} Dec 11 08:30:05 crc kubenswrapper[4860]: I1211 08:30:05.608974 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="040a0e8c-efd3-4e3b-be51-59e311ae4406" path="/var/lib/kubelet/pods/040a0e8c-efd3-4e3b-be51-59e311ae4406/volumes" Dec 11 08:30:06 crc kubenswrapper[4860]: I1211 08:30:06.070946 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"8a6d65b6-5393-4edc-a9c3-3660c8860fd1","Type":"ContainerStarted","Data":"74f5fdd09e066565e14f25d1faec292644a839347abfcd5edb74d06e61ad3833"} Dec 11 08:30:06 crc kubenswrapper[4860]: I1211 08:30:06.071257 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 11 08:30:06 crc kubenswrapper[4860]: I1211 08:30:06.074670 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2e2c517e-81fe-4563-889e-af154d1f500a","Type":"ContainerStarted","Data":"938851d2fa08ce2e25fbee8b87015ca7ff04119efecbd755fcbb891a2b000533"} Dec 11 08:30:06 crc kubenswrapper[4860]: I1211 08:30:06.102617 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.102594342 podStartE2EDuration="3.102594342s" podCreationTimestamp="2025-12-11 08:30:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:30:06.095057305 +0000 UTC m=+1138.823576360" watchObservedRunningTime="2025-12-11 08:30:06.102594342 +0000 UTC m=+1138.831113397" Dec 11 08:30:06 crc kubenswrapper[4860]: I1211 08:30:06.749544 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-5fccb6555c-xwc9b" Dec 11 08:30:06 crc kubenswrapper[4860]: I1211 08:30:06.837522 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-ccf7c6598-jcbgb"] Dec 11 08:30:06 crc kubenswrapper[4860]: I1211 08:30:06.851239 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-ccf7c6598-jcbgb" podUID="e076681d-ddec-44a2-9ce4-39b87e5542c6" containerName="neutron-api" containerID="cri-o://fcaf6875741f0622368b408b61196853ecbfe68ddd9ec0f7cf84241df6b96926" gracePeriod=30 Dec 11 08:30:06 crc kubenswrapper[4860]: I1211 08:30:06.851446 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-ccf7c6598-jcbgb" podUID="e076681d-ddec-44a2-9ce4-39b87e5542c6" containerName="neutron-httpd" containerID="cri-o://fd59343d1602e30d80627c28c579dc28a98e01ed5bdc68b6006a2b053ba08373" gracePeriod=30 Dec 11 08:30:07 crc kubenswrapper[4860]: I1211 08:30:07.090816 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2e2c517e-81fe-4563-889e-af154d1f500a","Type":"ContainerStarted","Data":"2851fc988d20e4bf3cae08111c3baea890cba97eaace2938f550e3e6cb24f15b"} Dec 11 08:30:07 crc kubenswrapper[4860]: I1211 08:30:07.094209 4860 generic.go:334] "Generic (PLEG): container finished" podID="e076681d-ddec-44a2-9ce4-39b87e5542c6" containerID="fd59343d1602e30d80627c28c579dc28a98e01ed5bdc68b6006a2b053ba08373" exitCode=0 Dec 11 08:30:07 crc kubenswrapper[4860]: I1211 08:30:07.094408 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-ccf7c6598-jcbgb" event={"ID":"e076681d-ddec-44a2-9ce4-39b87e5542c6","Type":"ContainerDied","Data":"fd59343d1602e30d80627c28c579dc28a98e01ed5bdc68b6006a2b053ba08373"} Dec 11 08:30:07 crc kubenswrapper[4860]: I1211 08:30:07.871794 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6bb4fc677f-mcn2v" Dec 11 08:30:08 crc kubenswrapper[4860]: I1211 08:30:08.021076 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-688c87cc99-sbps7"] Dec 11 08:30:08 crc kubenswrapper[4860]: I1211 08:30:08.021662 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-688c87cc99-sbps7" podUID="5bc875db-b1de-4250-b9e7-bdd3a54bb08f" containerName="dnsmasq-dns" containerID="cri-o://fb75e210972ef3010591ee0d3883d834f38a29b13cf4e9a71a0291a1ac174852" gracePeriod=10 Dec 11 08:30:08 crc kubenswrapper[4860]: I1211 08:30:08.105925 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 11 08:30:08 crc kubenswrapper[4860]: I1211 08:30:08.150402 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 11 08:30:08 crc kubenswrapper[4860]: I1211 08:30:08.661597 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-688c87cc99-sbps7" Dec 11 08:30:08 crc kubenswrapper[4860]: I1211 08:30:08.714070 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5bc875db-b1de-4250-b9e7-bdd3a54bb08f-dns-swift-storage-0\") pod \"5bc875db-b1de-4250-b9e7-bdd3a54bb08f\" (UID: \"5bc875db-b1de-4250-b9e7-bdd3a54bb08f\") " Dec 11 08:30:08 crc kubenswrapper[4860]: I1211 08:30:08.714129 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5bc875db-b1de-4250-b9e7-bdd3a54bb08f-config\") pod \"5bc875db-b1de-4250-b9e7-bdd3a54bb08f\" (UID: \"5bc875db-b1de-4250-b9e7-bdd3a54bb08f\") " Dec 11 08:30:08 crc kubenswrapper[4860]: I1211 08:30:08.714253 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2bp67\" (UniqueName: \"kubernetes.io/projected/5bc875db-b1de-4250-b9e7-bdd3a54bb08f-kube-api-access-2bp67\") pod \"5bc875db-b1de-4250-b9e7-bdd3a54bb08f\" (UID: \"5bc875db-b1de-4250-b9e7-bdd3a54bb08f\") " Dec 11 08:30:08 crc kubenswrapper[4860]: I1211 08:30:08.714283 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5bc875db-b1de-4250-b9e7-bdd3a54bb08f-ovsdbserver-sb\") pod \"5bc875db-b1de-4250-b9e7-bdd3a54bb08f\" (UID: \"5bc875db-b1de-4250-b9e7-bdd3a54bb08f\") " Dec 11 08:30:08 crc kubenswrapper[4860]: I1211 08:30:08.714352 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5bc875db-b1de-4250-b9e7-bdd3a54bb08f-dns-svc\") pod \"5bc875db-b1de-4250-b9e7-bdd3a54bb08f\" (UID: \"5bc875db-b1de-4250-b9e7-bdd3a54bb08f\") " Dec 11 08:30:08 crc kubenswrapper[4860]: I1211 08:30:08.714462 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5bc875db-b1de-4250-b9e7-bdd3a54bb08f-ovsdbserver-nb\") pod \"5bc875db-b1de-4250-b9e7-bdd3a54bb08f\" (UID: \"5bc875db-b1de-4250-b9e7-bdd3a54bb08f\") " Dec 11 08:30:08 crc kubenswrapper[4860]: I1211 08:30:08.729869 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5bc875db-b1de-4250-b9e7-bdd3a54bb08f-kube-api-access-2bp67" (OuterVolumeSpecName: "kube-api-access-2bp67") pod "5bc875db-b1de-4250-b9e7-bdd3a54bb08f" (UID: "5bc875db-b1de-4250-b9e7-bdd3a54bb08f"). InnerVolumeSpecName "kube-api-access-2bp67". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:30:08 crc kubenswrapper[4860]: I1211 08:30:08.793243 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5bc875db-b1de-4250-b9e7-bdd3a54bb08f-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "5bc875db-b1de-4250-b9e7-bdd3a54bb08f" (UID: "5bc875db-b1de-4250-b9e7-bdd3a54bb08f"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:30:08 crc kubenswrapper[4860]: I1211 08:30:08.818527 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5bc875db-b1de-4250-b9e7-bdd3a54bb08f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5bc875db-b1de-4250-b9e7-bdd3a54bb08f" (UID: "5bc875db-b1de-4250-b9e7-bdd3a54bb08f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:30:08 crc kubenswrapper[4860]: I1211 08:30:08.819364 4860 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5bc875db-b1de-4250-b9e7-bdd3a54bb08f-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 11 08:30:08 crc kubenswrapper[4860]: I1211 08:30:08.819404 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2bp67\" (UniqueName: \"kubernetes.io/projected/5bc875db-b1de-4250-b9e7-bdd3a54bb08f-kube-api-access-2bp67\") on node \"crc\" DevicePath \"\"" Dec 11 08:30:08 crc kubenswrapper[4860]: I1211 08:30:08.819417 4860 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5bc875db-b1de-4250-b9e7-bdd3a54bb08f-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 11 08:30:08 crc kubenswrapper[4860]: I1211 08:30:08.819700 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5bc875db-b1de-4250-b9e7-bdd3a54bb08f-config" (OuterVolumeSpecName: "config") pod "5bc875db-b1de-4250-b9e7-bdd3a54bb08f" (UID: "5bc875db-b1de-4250-b9e7-bdd3a54bb08f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:30:08 crc kubenswrapper[4860]: I1211 08:30:08.826602 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5bc875db-b1de-4250-b9e7-bdd3a54bb08f-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "5bc875db-b1de-4250-b9e7-bdd3a54bb08f" (UID: "5bc875db-b1de-4250-b9e7-bdd3a54bb08f"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:30:08 crc kubenswrapper[4860]: I1211 08:30:08.838172 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5bc875db-b1de-4250-b9e7-bdd3a54bb08f-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "5bc875db-b1de-4250-b9e7-bdd3a54bb08f" (UID: "5bc875db-b1de-4250-b9e7-bdd3a54bb08f"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:30:08 crc kubenswrapper[4860]: I1211 08:30:08.921013 4860 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5bc875db-b1de-4250-b9e7-bdd3a54bb08f-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 11 08:30:08 crc kubenswrapper[4860]: I1211 08:30:08.921052 4860 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5bc875db-b1de-4250-b9e7-bdd3a54bb08f-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 11 08:30:08 crc kubenswrapper[4860]: I1211 08:30:08.921063 4860 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5bc875db-b1de-4250-b9e7-bdd3a54bb08f-config\") on node \"crc\" DevicePath \"\"" Dec 11 08:30:09 crc kubenswrapper[4860]: I1211 08:30:09.125361 4860 generic.go:334] "Generic (PLEG): container finished" podID="5bc875db-b1de-4250-b9e7-bdd3a54bb08f" containerID="fb75e210972ef3010591ee0d3883d834f38a29b13cf4e9a71a0291a1ac174852" exitCode=0 Dec 11 08:30:09 crc kubenswrapper[4860]: I1211 08:30:09.125431 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-688c87cc99-sbps7" Dec 11 08:30:09 crc kubenswrapper[4860]: I1211 08:30:09.125433 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-688c87cc99-sbps7" event={"ID":"5bc875db-b1de-4250-b9e7-bdd3a54bb08f","Type":"ContainerDied","Data":"fb75e210972ef3010591ee0d3883d834f38a29b13cf4e9a71a0291a1ac174852"} Dec 11 08:30:09 crc kubenswrapper[4860]: I1211 08:30:09.125468 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-688c87cc99-sbps7" event={"ID":"5bc875db-b1de-4250-b9e7-bdd3a54bb08f","Type":"ContainerDied","Data":"5f23cfec3d4c07fa6689cf43fd028636386475064e2c2335cbaf1bf64442da5b"} Dec 11 08:30:09 crc kubenswrapper[4860]: I1211 08:30:09.125502 4860 scope.go:117] "RemoveContainer" containerID="fb75e210972ef3010591ee0d3883d834f38a29b13cf4e9a71a0291a1ac174852" Dec 11 08:30:09 crc kubenswrapper[4860]: I1211 08:30:09.127987 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2e2c517e-81fe-4563-889e-af154d1f500a","Type":"ContainerStarted","Data":"0ebcc0b0cc22669fec7bffa7bda1858a957e5d7ac94dedf0cefa8a030cca3fe6"} Dec 11 08:30:09 crc kubenswrapper[4860]: I1211 08:30:09.128151 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="6075b482-4926-4cf5-b632-0a731d2686dc" containerName="cinder-scheduler" containerID="cri-o://5c83bda1a7e1ac8d4f912f91a01f263592a538d5209e99c9a8bf509f9d11da72" gracePeriod=30 Dec 11 08:30:09 crc kubenswrapper[4860]: I1211 08:30:09.128166 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="6075b482-4926-4cf5-b632-0a731d2686dc" containerName="probe" containerID="cri-o://cfac87528b0888c0e951b979a1fd4da21776995c318a81737d10b48d2d6886c6" gracePeriod=30 Dec 11 08:30:09 crc kubenswrapper[4860]: I1211 08:30:09.164203 4860 scope.go:117] "RemoveContainer" containerID="69e0e84cf252adb8a847241eca4fd03574e73a2fe440b02276821f9fab9eecd4" Dec 11 08:30:09 crc kubenswrapper[4860]: I1211 08:30:09.173904 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-688c87cc99-sbps7"] Dec 11 08:30:09 crc kubenswrapper[4860]: I1211 08:30:09.192277 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-688c87cc99-sbps7"] Dec 11 08:30:09 crc kubenswrapper[4860]: I1211 08:30:09.201744 4860 scope.go:117] "RemoveContainer" containerID="fb75e210972ef3010591ee0d3883d834f38a29b13cf4e9a71a0291a1ac174852" Dec 11 08:30:09 crc kubenswrapper[4860]: E1211 08:30:09.202324 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fb75e210972ef3010591ee0d3883d834f38a29b13cf4e9a71a0291a1ac174852\": container with ID starting with fb75e210972ef3010591ee0d3883d834f38a29b13cf4e9a71a0291a1ac174852 not found: ID does not exist" containerID="fb75e210972ef3010591ee0d3883d834f38a29b13cf4e9a71a0291a1ac174852" Dec 11 08:30:09 crc kubenswrapper[4860]: I1211 08:30:09.202374 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fb75e210972ef3010591ee0d3883d834f38a29b13cf4e9a71a0291a1ac174852"} err="failed to get container status \"fb75e210972ef3010591ee0d3883d834f38a29b13cf4e9a71a0291a1ac174852\": rpc error: code = NotFound desc = could not find container \"fb75e210972ef3010591ee0d3883d834f38a29b13cf4e9a71a0291a1ac174852\": container with ID starting with fb75e210972ef3010591ee0d3883d834f38a29b13cf4e9a71a0291a1ac174852 not found: ID does not exist" Dec 11 08:30:09 crc kubenswrapper[4860]: I1211 08:30:09.202401 4860 scope.go:117] "RemoveContainer" containerID="69e0e84cf252adb8a847241eca4fd03574e73a2fe440b02276821f9fab9eecd4" Dec 11 08:30:09 crc kubenswrapper[4860]: E1211 08:30:09.202799 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"69e0e84cf252adb8a847241eca4fd03574e73a2fe440b02276821f9fab9eecd4\": container with ID starting with 69e0e84cf252adb8a847241eca4fd03574e73a2fe440b02276821f9fab9eecd4 not found: ID does not exist" containerID="69e0e84cf252adb8a847241eca4fd03574e73a2fe440b02276821f9fab9eecd4" Dec 11 08:30:09 crc kubenswrapper[4860]: I1211 08:30:09.202827 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"69e0e84cf252adb8a847241eca4fd03574e73a2fe440b02276821f9fab9eecd4"} err="failed to get container status \"69e0e84cf252adb8a847241eca4fd03574e73a2fe440b02276821f9fab9eecd4\": rpc error: code = NotFound desc = could not find container \"69e0e84cf252adb8a847241eca4fd03574e73a2fe440b02276821f9fab9eecd4\": container with ID starting with 69e0e84cf252adb8a847241eca4fd03574e73a2fe440b02276821f9fab9eecd4 not found: ID does not exist" Dec 11 08:30:09 crc kubenswrapper[4860]: I1211 08:30:09.323589 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-55ccdb6bbd-grdt6" Dec 11 08:30:09 crc kubenswrapper[4860]: I1211 08:30:09.387955 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-776755b658-9fwr9" Dec 11 08:30:09 crc kubenswrapper[4860]: I1211 08:30:09.518623 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-55ccdb6bbd-grdt6" Dec 11 08:30:09 crc kubenswrapper[4860]: I1211 08:30:09.623341 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5bc875db-b1de-4250-b9e7-bdd3a54bb08f" path="/var/lib/kubelet/pods/5bc875db-b1de-4250-b9e7-bdd3a54bb08f/volumes" Dec 11 08:30:09 crc kubenswrapper[4860]: I1211 08:30:09.629218 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-6b4697fc68-vnd26" Dec 11 08:30:10 crc kubenswrapper[4860]: I1211 08:30:10.153755 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2e2c517e-81fe-4563-889e-af154d1f500a","Type":"ContainerStarted","Data":"c139b307041de43862507eac5d2f3deee5e84a25fa1e5cc7c663e05bf5f01ad8"} Dec 11 08:30:10 crc kubenswrapper[4860]: I1211 08:30:10.153815 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 11 08:30:10 crc kubenswrapper[4860]: I1211 08:30:10.179817 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.650651818 podStartE2EDuration="6.179797633s" podCreationTimestamp="2025-12-11 08:30:04 +0000 UTC" firstStartedPulling="2025-12-11 08:30:05.003051493 +0000 UTC m=+1137.731570548" lastFinishedPulling="2025-12-11 08:30:09.532197308 +0000 UTC m=+1142.260716363" observedRunningTime="2025-12-11 08:30:10.179277886 +0000 UTC m=+1142.907796941" watchObservedRunningTime="2025-12-11 08:30:10.179797633 +0000 UTC m=+1142.908316688" Dec 11 08:30:11 crc kubenswrapper[4860]: I1211 08:30:11.176500 4860 generic.go:334] "Generic (PLEG): container finished" podID="6075b482-4926-4cf5-b632-0a731d2686dc" containerID="cfac87528b0888c0e951b979a1fd4da21776995c318a81737d10b48d2d6886c6" exitCode=0 Dec 11 08:30:11 crc kubenswrapper[4860]: I1211 08:30:11.177118 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"6075b482-4926-4cf5-b632-0a731d2686dc","Type":"ContainerDied","Data":"cfac87528b0888c0e951b979a1fd4da21776995c318a81737d10b48d2d6886c6"} Dec 11 08:30:11 crc kubenswrapper[4860]: I1211 08:30:11.184320 4860 generic.go:334] "Generic (PLEG): container finished" podID="e076681d-ddec-44a2-9ce4-39b87e5542c6" containerID="fcaf6875741f0622368b408b61196853ecbfe68ddd9ec0f7cf84241df6b96926" exitCode=0 Dec 11 08:30:11 crc kubenswrapper[4860]: I1211 08:30:11.185543 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-ccf7c6598-jcbgb" event={"ID":"e076681d-ddec-44a2-9ce4-39b87e5542c6","Type":"ContainerDied","Data":"fcaf6875741f0622368b408b61196853ecbfe68ddd9ec0f7cf84241df6b96926"} Dec 11 08:30:11 crc kubenswrapper[4860]: I1211 08:30:11.185587 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-ccf7c6598-jcbgb" event={"ID":"e076681d-ddec-44a2-9ce4-39b87e5542c6","Type":"ContainerDied","Data":"f409338159e92c6ed18765cf7e90fd5ab1cabe9c8f2ab78e9b25db7c4312cd65"} Dec 11 08:30:11 crc kubenswrapper[4860]: I1211 08:30:11.185604 4860 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f409338159e92c6ed18765cf7e90fd5ab1cabe9c8f2ab78e9b25db7c4312cd65" Dec 11 08:30:11 crc kubenswrapper[4860]: I1211 08:30:11.194183 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-ccf7c6598-jcbgb" Dec 11 08:30:11 crc kubenswrapper[4860]: I1211 08:30:11.372907 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e076681d-ddec-44a2-9ce4-39b87e5542c6-httpd-config\") pod \"e076681d-ddec-44a2-9ce4-39b87e5542c6\" (UID: \"e076681d-ddec-44a2-9ce4-39b87e5542c6\") " Dec 11 08:30:11 crc kubenswrapper[4860]: I1211 08:30:11.373166 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e076681d-ddec-44a2-9ce4-39b87e5542c6-ovndb-tls-certs\") pod \"e076681d-ddec-44a2-9ce4-39b87e5542c6\" (UID: \"e076681d-ddec-44a2-9ce4-39b87e5542c6\") " Dec 11 08:30:11 crc kubenswrapper[4860]: I1211 08:30:11.373218 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e076681d-ddec-44a2-9ce4-39b87e5542c6-config\") pod \"e076681d-ddec-44a2-9ce4-39b87e5542c6\" (UID: \"e076681d-ddec-44a2-9ce4-39b87e5542c6\") " Dec 11 08:30:11 crc kubenswrapper[4860]: I1211 08:30:11.373292 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c7c9x\" (UniqueName: \"kubernetes.io/projected/e076681d-ddec-44a2-9ce4-39b87e5542c6-kube-api-access-c7c9x\") pod \"e076681d-ddec-44a2-9ce4-39b87e5542c6\" (UID: \"e076681d-ddec-44a2-9ce4-39b87e5542c6\") " Dec 11 08:30:11 crc kubenswrapper[4860]: I1211 08:30:11.373365 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e076681d-ddec-44a2-9ce4-39b87e5542c6-combined-ca-bundle\") pod \"e076681d-ddec-44a2-9ce4-39b87e5542c6\" (UID: \"e076681d-ddec-44a2-9ce4-39b87e5542c6\") " Dec 11 08:30:11 crc kubenswrapper[4860]: I1211 08:30:11.393022 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e076681d-ddec-44a2-9ce4-39b87e5542c6-kube-api-access-c7c9x" (OuterVolumeSpecName: "kube-api-access-c7c9x") pod "e076681d-ddec-44a2-9ce4-39b87e5542c6" (UID: "e076681d-ddec-44a2-9ce4-39b87e5542c6"). InnerVolumeSpecName "kube-api-access-c7c9x". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:30:11 crc kubenswrapper[4860]: I1211 08:30:11.403903 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e076681d-ddec-44a2-9ce4-39b87e5542c6-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "e076681d-ddec-44a2-9ce4-39b87e5542c6" (UID: "e076681d-ddec-44a2-9ce4-39b87e5542c6"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:30:11 crc kubenswrapper[4860]: I1211 08:30:11.446752 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e076681d-ddec-44a2-9ce4-39b87e5542c6-config" (OuterVolumeSpecName: "config") pod "e076681d-ddec-44a2-9ce4-39b87e5542c6" (UID: "e076681d-ddec-44a2-9ce4-39b87e5542c6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:30:11 crc kubenswrapper[4860]: I1211 08:30:11.471768 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e076681d-ddec-44a2-9ce4-39b87e5542c6-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "e076681d-ddec-44a2-9ce4-39b87e5542c6" (UID: "e076681d-ddec-44a2-9ce4-39b87e5542c6"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:30:11 crc kubenswrapper[4860]: I1211 08:30:11.475962 4860 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e076681d-ddec-44a2-9ce4-39b87e5542c6-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 11 08:30:11 crc kubenswrapper[4860]: I1211 08:30:11.475999 4860 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/e076681d-ddec-44a2-9ce4-39b87e5542c6-config\") on node \"crc\" DevicePath \"\"" Dec 11 08:30:11 crc kubenswrapper[4860]: I1211 08:30:11.476014 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c7c9x\" (UniqueName: \"kubernetes.io/projected/e076681d-ddec-44a2-9ce4-39b87e5542c6-kube-api-access-c7c9x\") on node \"crc\" DevicePath \"\"" Dec 11 08:30:11 crc kubenswrapper[4860]: I1211 08:30:11.476026 4860 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e076681d-ddec-44a2-9ce4-39b87e5542c6-httpd-config\") on node \"crc\" DevicePath \"\"" Dec 11 08:30:11 crc kubenswrapper[4860]: I1211 08:30:11.486444 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e076681d-ddec-44a2-9ce4-39b87e5542c6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e076681d-ddec-44a2-9ce4-39b87e5542c6" (UID: "e076681d-ddec-44a2-9ce4-39b87e5542c6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:30:11 crc kubenswrapper[4860]: I1211 08:30:11.585850 4860 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e076681d-ddec-44a2-9ce4-39b87e5542c6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 08:30:11 crc kubenswrapper[4860]: I1211 08:30:11.906042 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 11 08:30:12 crc kubenswrapper[4860]: I1211 08:30:12.035339 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-748d5bd978-4sz6v" Dec 11 08:30:12 crc kubenswrapper[4860]: I1211 08:30:12.053772 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 11 08:30:12 crc kubenswrapper[4860]: I1211 08:30:12.073629 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-776755b658-9fwr9" Dec 11 08:30:12 crc kubenswrapper[4860]: I1211 08:30:12.162230 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-6b4697fc68-vnd26" Dec 11 08:30:12 crc kubenswrapper[4860]: I1211 08:30:12.206275 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-ccf7c6598-jcbgb" Dec 11 08:30:12 crc kubenswrapper[4860]: I1211 08:30:12.251821 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-776755b658-9fwr9"] Dec 11 08:30:12 crc kubenswrapper[4860]: I1211 08:30:12.252176 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-776755b658-9fwr9" podUID="8f38e033-5243-4e65-a7e0-29f1282b063f" containerName="horizon-log" containerID="cri-o://0aab949e04f4b3a4703fdc0bff0fb82f6a451c3c24e3537820a25282e91ff2c8" gracePeriod=30 Dec 11 08:30:12 crc kubenswrapper[4860]: I1211 08:30:12.252801 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-776755b658-9fwr9" podUID="8f38e033-5243-4e65-a7e0-29f1282b063f" containerName="horizon" containerID="cri-o://948c71446dedd81878e5444ab6a4dbd7fbfe0ce699ae4e007ae7d21b188368ec" gracePeriod=30 Dec 11 08:30:12 crc kubenswrapper[4860]: I1211 08:30:12.270486 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-ccf7c6598-jcbgb"] Dec 11 08:30:12 crc kubenswrapper[4860]: I1211 08:30:12.289569 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-ccf7c6598-jcbgb"] Dec 11 08:30:13 crc kubenswrapper[4860]: I1211 08:30:13.219790 4860 generic.go:334] "Generic (PLEG): container finished" podID="6075b482-4926-4cf5-b632-0a731d2686dc" containerID="5c83bda1a7e1ac8d4f912f91a01f263592a538d5209e99c9a8bf509f9d11da72" exitCode=0 Dec 11 08:30:13 crc kubenswrapper[4860]: I1211 08:30:13.219995 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"6075b482-4926-4cf5-b632-0a731d2686dc","Type":"ContainerDied","Data":"5c83bda1a7e1ac8d4f912f91a01f263592a538d5209e99c9a8bf509f9d11da72"} Dec 11 08:30:13 crc kubenswrapper[4860]: I1211 08:30:13.220170 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"6075b482-4926-4cf5-b632-0a731d2686dc","Type":"ContainerDied","Data":"7032faeaef7a9f5f18ba1e486ffed6c02c2057f931f2ef0af05bf26658f60558"} Dec 11 08:30:13 crc kubenswrapper[4860]: I1211 08:30:13.220798 4860 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7032faeaef7a9f5f18ba1e486ffed6c02c2057f931f2ef0af05bf26658f60558" Dec 11 08:30:13 crc kubenswrapper[4860]: I1211 08:30:13.299919 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 11 08:30:13 crc kubenswrapper[4860]: I1211 08:30:13.457372 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6075b482-4926-4cf5-b632-0a731d2686dc-etc-machine-id\") pod \"6075b482-4926-4cf5-b632-0a731d2686dc\" (UID: \"6075b482-4926-4cf5-b632-0a731d2686dc\") " Dec 11 08:30:13 crc kubenswrapper[4860]: I1211 08:30:13.457744 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6075b482-4926-4cf5-b632-0a731d2686dc-combined-ca-bundle\") pod \"6075b482-4926-4cf5-b632-0a731d2686dc\" (UID: \"6075b482-4926-4cf5-b632-0a731d2686dc\") " Dec 11 08:30:13 crc kubenswrapper[4860]: I1211 08:30:13.457868 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6075b482-4926-4cf5-b632-0a731d2686dc-scripts\") pod \"6075b482-4926-4cf5-b632-0a731d2686dc\" (UID: \"6075b482-4926-4cf5-b632-0a731d2686dc\") " Dec 11 08:30:13 crc kubenswrapper[4860]: I1211 08:30:13.458047 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bsbt6\" (UniqueName: \"kubernetes.io/projected/6075b482-4926-4cf5-b632-0a731d2686dc-kube-api-access-bsbt6\") pod \"6075b482-4926-4cf5-b632-0a731d2686dc\" (UID: \"6075b482-4926-4cf5-b632-0a731d2686dc\") " Dec 11 08:30:13 crc kubenswrapper[4860]: I1211 08:30:13.458763 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6075b482-4926-4cf5-b632-0a731d2686dc-config-data\") pod \"6075b482-4926-4cf5-b632-0a731d2686dc\" (UID: \"6075b482-4926-4cf5-b632-0a731d2686dc\") " Dec 11 08:30:13 crc kubenswrapper[4860]: I1211 08:30:13.458933 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6075b482-4926-4cf5-b632-0a731d2686dc-config-data-custom\") pod \"6075b482-4926-4cf5-b632-0a731d2686dc\" (UID: \"6075b482-4926-4cf5-b632-0a731d2686dc\") " Dec 11 08:30:13 crc kubenswrapper[4860]: I1211 08:30:13.457511 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6075b482-4926-4cf5-b632-0a731d2686dc-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "6075b482-4926-4cf5-b632-0a731d2686dc" (UID: "6075b482-4926-4cf5-b632-0a731d2686dc"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 08:30:13 crc kubenswrapper[4860]: I1211 08:30:13.459817 4860 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6075b482-4926-4cf5-b632-0a731d2686dc-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 11 08:30:13 crc kubenswrapper[4860]: I1211 08:30:13.466882 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6075b482-4926-4cf5-b632-0a731d2686dc-kube-api-access-bsbt6" (OuterVolumeSpecName: "kube-api-access-bsbt6") pod "6075b482-4926-4cf5-b632-0a731d2686dc" (UID: "6075b482-4926-4cf5-b632-0a731d2686dc"). InnerVolumeSpecName "kube-api-access-bsbt6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:30:13 crc kubenswrapper[4860]: I1211 08:30:13.467271 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6075b482-4926-4cf5-b632-0a731d2686dc-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "6075b482-4926-4cf5-b632-0a731d2686dc" (UID: "6075b482-4926-4cf5-b632-0a731d2686dc"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:30:13 crc kubenswrapper[4860]: I1211 08:30:13.484277 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6075b482-4926-4cf5-b632-0a731d2686dc-scripts" (OuterVolumeSpecName: "scripts") pod "6075b482-4926-4cf5-b632-0a731d2686dc" (UID: "6075b482-4926-4cf5-b632-0a731d2686dc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:30:13 crc kubenswrapper[4860]: I1211 08:30:13.521104 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6075b482-4926-4cf5-b632-0a731d2686dc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6075b482-4926-4cf5-b632-0a731d2686dc" (UID: "6075b482-4926-4cf5-b632-0a731d2686dc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:30:13 crc kubenswrapper[4860]: I1211 08:30:13.564307 4860 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6075b482-4926-4cf5-b632-0a731d2686dc-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 11 08:30:13 crc kubenswrapper[4860]: I1211 08:30:13.564585 4860 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6075b482-4926-4cf5-b632-0a731d2686dc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 08:30:13 crc kubenswrapper[4860]: I1211 08:30:13.564671 4860 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6075b482-4926-4cf5-b632-0a731d2686dc-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 08:30:13 crc kubenswrapper[4860]: I1211 08:30:13.564730 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bsbt6\" (UniqueName: \"kubernetes.io/projected/6075b482-4926-4cf5-b632-0a731d2686dc-kube-api-access-bsbt6\") on node \"crc\" DevicePath \"\"" Dec 11 08:30:13 crc kubenswrapper[4860]: I1211 08:30:13.591268 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e076681d-ddec-44a2-9ce4-39b87e5542c6" path="/var/lib/kubelet/pods/e076681d-ddec-44a2-9ce4-39b87e5542c6/volumes" Dec 11 08:30:13 crc kubenswrapper[4860]: I1211 08:30:13.615858 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6075b482-4926-4cf5-b632-0a731d2686dc-config-data" (OuterVolumeSpecName: "config-data") pod "6075b482-4926-4cf5-b632-0a731d2686dc" (UID: "6075b482-4926-4cf5-b632-0a731d2686dc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:30:13 crc kubenswrapper[4860]: I1211 08:30:13.666366 4860 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6075b482-4926-4cf5-b632-0a731d2686dc-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 08:30:14 crc kubenswrapper[4860]: I1211 08:30:14.229809 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 11 08:30:14 crc kubenswrapper[4860]: I1211 08:30:14.269447 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 11 08:30:14 crc kubenswrapper[4860]: I1211 08:30:14.289977 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 11 08:30:14 crc kubenswrapper[4860]: I1211 08:30:14.301447 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 11 08:30:14 crc kubenswrapper[4860]: E1211 08:30:14.301996 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5bc875db-b1de-4250-b9e7-bdd3a54bb08f" containerName="init" Dec 11 08:30:14 crc kubenswrapper[4860]: I1211 08:30:14.302018 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="5bc875db-b1de-4250-b9e7-bdd3a54bb08f" containerName="init" Dec 11 08:30:14 crc kubenswrapper[4860]: E1211 08:30:14.302030 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e076681d-ddec-44a2-9ce4-39b87e5542c6" containerName="neutron-api" Dec 11 08:30:14 crc kubenswrapper[4860]: I1211 08:30:14.302039 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="e076681d-ddec-44a2-9ce4-39b87e5542c6" containerName="neutron-api" Dec 11 08:30:14 crc kubenswrapper[4860]: E1211 08:30:14.302084 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e076681d-ddec-44a2-9ce4-39b87e5542c6" containerName="neutron-httpd" Dec 11 08:30:14 crc kubenswrapper[4860]: I1211 08:30:14.302094 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="e076681d-ddec-44a2-9ce4-39b87e5542c6" containerName="neutron-httpd" Dec 11 08:30:14 crc kubenswrapper[4860]: E1211 08:30:14.302106 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5bc875db-b1de-4250-b9e7-bdd3a54bb08f" containerName="dnsmasq-dns" Dec 11 08:30:14 crc kubenswrapper[4860]: I1211 08:30:14.302114 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="5bc875db-b1de-4250-b9e7-bdd3a54bb08f" containerName="dnsmasq-dns" Dec 11 08:30:14 crc kubenswrapper[4860]: E1211 08:30:14.302149 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6075b482-4926-4cf5-b632-0a731d2686dc" containerName="probe" Dec 11 08:30:14 crc kubenswrapper[4860]: I1211 08:30:14.302157 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="6075b482-4926-4cf5-b632-0a731d2686dc" containerName="probe" Dec 11 08:30:14 crc kubenswrapper[4860]: E1211 08:30:14.302173 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6075b482-4926-4cf5-b632-0a731d2686dc" containerName="cinder-scheduler" Dec 11 08:30:14 crc kubenswrapper[4860]: I1211 08:30:14.302180 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="6075b482-4926-4cf5-b632-0a731d2686dc" containerName="cinder-scheduler" Dec 11 08:30:14 crc kubenswrapper[4860]: I1211 08:30:14.302407 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="e076681d-ddec-44a2-9ce4-39b87e5542c6" containerName="neutron-httpd" Dec 11 08:30:14 crc kubenswrapper[4860]: I1211 08:30:14.302422 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="5bc875db-b1de-4250-b9e7-bdd3a54bb08f" containerName="dnsmasq-dns" Dec 11 08:30:14 crc kubenswrapper[4860]: I1211 08:30:14.302441 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="e076681d-ddec-44a2-9ce4-39b87e5542c6" containerName="neutron-api" Dec 11 08:30:14 crc kubenswrapper[4860]: I1211 08:30:14.302467 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="6075b482-4926-4cf5-b632-0a731d2686dc" containerName="cinder-scheduler" Dec 11 08:30:14 crc kubenswrapper[4860]: I1211 08:30:14.302481 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="6075b482-4926-4cf5-b632-0a731d2686dc" containerName="probe" Dec 11 08:30:14 crc kubenswrapper[4860]: I1211 08:30:14.303696 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 11 08:30:14 crc kubenswrapper[4860]: I1211 08:30:14.307006 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 11 08:30:14 crc kubenswrapper[4860]: I1211 08:30:14.315917 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 11 08:30:14 crc kubenswrapper[4860]: I1211 08:30:14.385036 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5dc4bc3d-458e-4251-81f7-68b5b47a7275-config-data\") pod \"cinder-scheduler-0\" (UID: \"5dc4bc3d-458e-4251-81f7-68b5b47a7275\") " pod="openstack/cinder-scheduler-0" Dec 11 08:30:14 crc kubenswrapper[4860]: I1211 08:30:14.385114 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mqx9s\" (UniqueName: \"kubernetes.io/projected/5dc4bc3d-458e-4251-81f7-68b5b47a7275-kube-api-access-mqx9s\") pod \"cinder-scheduler-0\" (UID: \"5dc4bc3d-458e-4251-81f7-68b5b47a7275\") " pod="openstack/cinder-scheduler-0" Dec 11 08:30:14 crc kubenswrapper[4860]: I1211 08:30:14.385152 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5dc4bc3d-458e-4251-81f7-68b5b47a7275-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"5dc4bc3d-458e-4251-81f7-68b5b47a7275\") " pod="openstack/cinder-scheduler-0" Dec 11 08:30:14 crc kubenswrapper[4860]: I1211 08:30:14.385304 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5dc4bc3d-458e-4251-81f7-68b5b47a7275-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"5dc4bc3d-458e-4251-81f7-68b5b47a7275\") " pod="openstack/cinder-scheduler-0" Dec 11 08:30:14 crc kubenswrapper[4860]: I1211 08:30:14.385334 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5dc4bc3d-458e-4251-81f7-68b5b47a7275-scripts\") pod \"cinder-scheduler-0\" (UID: \"5dc4bc3d-458e-4251-81f7-68b5b47a7275\") " pod="openstack/cinder-scheduler-0" Dec 11 08:30:14 crc kubenswrapper[4860]: I1211 08:30:14.385448 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5dc4bc3d-458e-4251-81f7-68b5b47a7275-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"5dc4bc3d-458e-4251-81f7-68b5b47a7275\") " pod="openstack/cinder-scheduler-0" Dec 11 08:30:14 crc kubenswrapper[4860]: I1211 08:30:14.486172 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5dc4bc3d-458e-4251-81f7-68b5b47a7275-config-data\") pod \"cinder-scheduler-0\" (UID: \"5dc4bc3d-458e-4251-81f7-68b5b47a7275\") " pod="openstack/cinder-scheduler-0" Dec 11 08:30:14 crc kubenswrapper[4860]: I1211 08:30:14.486239 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mqx9s\" (UniqueName: \"kubernetes.io/projected/5dc4bc3d-458e-4251-81f7-68b5b47a7275-kube-api-access-mqx9s\") pod \"cinder-scheduler-0\" (UID: \"5dc4bc3d-458e-4251-81f7-68b5b47a7275\") " pod="openstack/cinder-scheduler-0" Dec 11 08:30:14 crc kubenswrapper[4860]: I1211 08:30:14.486262 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5dc4bc3d-458e-4251-81f7-68b5b47a7275-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"5dc4bc3d-458e-4251-81f7-68b5b47a7275\") " pod="openstack/cinder-scheduler-0" Dec 11 08:30:14 crc kubenswrapper[4860]: I1211 08:30:14.486344 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5dc4bc3d-458e-4251-81f7-68b5b47a7275-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"5dc4bc3d-458e-4251-81f7-68b5b47a7275\") " pod="openstack/cinder-scheduler-0" Dec 11 08:30:14 crc kubenswrapper[4860]: I1211 08:30:14.486371 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5dc4bc3d-458e-4251-81f7-68b5b47a7275-scripts\") pod \"cinder-scheduler-0\" (UID: \"5dc4bc3d-458e-4251-81f7-68b5b47a7275\") " pod="openstack/cinder-scheduler-0" Dec 11 08:30:14 crc kubenswrapper[4860]: I1211 08:30:14.486445 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5dc4bc3d-458e-4251-81f7-68b5b47a7275-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"5dc4bc3d-458e-4251-81f7-68b5b47a7275\") " pod="openstack/cinder-scheduler-0" Dec 11 08:30:14 crc kubenswrapper[4860]: I1211 08:30:14.486526 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5dc4bc3d-458e-4251-81f7-68b5b47a7275-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"5dc4bc3d-458e-4251-81f7-68b5b47a7275\") " pod="openstack/cinder-scheduler-0" Dec 11 08:30:14 crc kubenswrapper[4860]: I1211 08:30:14.493886 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5dc4bc3d-458e-4251-81f7-68b5b47a7275-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"5dc4bc3d-458e-4251-81f7-68b5b47a7275\") " pod="openstack/cinder-scheduler-0" Dec 11 08:30:14 crc kubenswrapper[4860]: I1211 08:30:14.495048 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5dc4bc3d-458e-4251-81f7-68b5b47a7275-scripts\") pod \"cinder-scheduler-0\" (UID: \"5dc4bc3d-458e-4251-81f7-68b5b47a7275\") " pod="openstack/cinder-scheduler-0" Dec 11 08:30:14 crc kubenswrapper[4860]: I1211 08:30:14.507871 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5dc4bc3d-458e-4251-81f7-68b5b47a7275-config-data\") pod \"cinder-scheduler-0\" (UID: \"5dc4bc3d-458e-4251-81f7-68b5b47a7275\") " pod="openstack/cinder-scheduler-0" Dec 11 08:30:14 crc kubenswrapper[4860]: I1211 08:30:14.515742 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mqx9s\" (UniqueName: \"kubernetes.io/projected/5dc4bc3d-458e-4251-81f7-68b5b47a7275-kube-api-access-mqx9s\") pod \"cinder-scheduler-0\" (UID: \"5dc4bc3d-458e-4251-81f7-68b5b47a7275\") " pod="openstack/cinder-scheduler-0" Dec 11 08:30:14 crc kubenswrapper[4860]: I1211 08:30:14.516268 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5dc4bc3d-458e-4251-81f7-68b5b47a7275-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"5dc4bc3d-458e-4251-81f7-68b5b47a7275\") " pod="openstack/cinder-scheduler-0" Dec 11 08:30:14 crc kubenswrapper[4860]: I1211 08:30:14.625407 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 11 08:30:15 crc kubenswrapper[4860]: I1211 08:30:15.183773 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 11 08:30:15 crc kubenswrapper[4860]: I1211 08:30:15.244264 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"5dc4bc3d-458e-4251-81f7-68b5b47a7275","Type":"ContainerStarted","Data":"969e749d14f1513051e19b2c4b4fe33305e86709aaf65653e3e130bd83341687"} Dec 11 08:30:15 crc kubenswrapper[4860]: I1211 08:30:15.321683 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 11 08:30:15 crc kubenswrapper[4860]: I1211 08:30:15.340664 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 11 08:30:15 crc kubenswrapper[4860]: I1211 08:30:15.340858 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 11 08:30:15 crc kubenswrapper[4860]: I1211 08:30:15.345262 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-z7vc4" Dec 11 08:30:15 crc kubenswrapper[4860]: I1211 08:30:15.345488 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Dec 11 08:30:15 crc kubenswrapper[4860]: I1211 08:30:15.345597 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Dec 11 08:30:15 crc kubenswrapper[4860]: I1211 08:30:15.505151 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/6f6e1905-c5e6-495b-9076-fa5edd9a01ac-openstack-config-secret\") pod \"openstackclient\" (UID: \"6f6e1905-c5e6-495b-9076-fa5edd9a01ac\") " pod="openstack/openstackclient" Dec 11 08:30:15 crc kubenswrapper[4860]: I1211 08:30:15.505377 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/6f6e1905-c5e6-495b-9076-fa5edd9a01ac-openstack-config\") pod \"openstackclient\" (UID: \"6f6e1905-c5e6-495b-9076-fa5edd9a01ac\") " pod="openstack/openstackclient" Dec 11 08:30:15 crc kubenswrapper[4860]: I1211 08:30:15.505437 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xtdmz\" (UniqueName: \"kubernetes.io/projected/6f6e1905-c5e6-495b-9076-fa5edd9a01ac-kube-api-access-xtdmz\") pod \"openstackclient\" (UID: \"6f6e1905-c5e6-495b-9076-fa5edd9a01ac\") " pod="openstack/openstackclient" Dec 11 08:30:15 crc kubenswrapper[4860]: I1211 08:30:15.505542 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f6e1905-c5e6-495b-9076-fa5edd9a01ac-combined-ca-bundle\") pod \"openstackclient\" (UID: \"6f6e1905-c5e6-495b-9076-fa5edd9a01ac\") " pod="openstack/openstackclient" Dec 11 08:30:15 crc kubenswrapper[4860]: I1211 08:30:15.612428 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/6f6e1905-c5e6-495b-9076-fa5edd9a01ac-openstack-config\") pod \"openstackclient\" (UID: \"6f6e1905-c5e6-495b-9076-fa5edd9a01ac\") " pod="openstack/openstackclient" Dec 11 08:30:15 crc kubenswrapper[4860]: I1211 08:30:15.612504 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xtdmz\" (UniqueName: \"kubernetes.io/projected/6f6e1905-c5e6-495b-9076-fa5edd9a01ac-kube-api-access-xtdmz\") pod \"openstackclient\" (UID: \"6f6e1905-c5e6-495b-9076-fa5edd9a01ac\") " pod="openstack/openstackclient" Dec 11 08:30:15 crc kubenswrapper[4860]: I1211 08:30:15.612950 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f6e1905-c5e6-495b-9076-fa5edd9a01ac-combined-ca-bundle\") pod \"openstackclient\" (UID: \"6f6e1905-c5e6-495b-9076-fa5edd9a01ac\") " pod="openstack/openstackclient" Dec 11 08:30:15 crc kubenswrapper[4860]: I1211 08:30:15.613213 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/6f6e1905-c5e6-495b-9076-fa5edd9a01ac-openstack-config-secret\") pod \"openstackclient\" (UID: \"6f6e1905-c5e6-495b-9076-fa5edd9a01ac\") " pod="openstack/openstackclient" Dec 11 08:30:15 crc kubenswrapper[4860]: I1211 08:30:15.615785 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/6f6e1905-c5e6-495b-9076-fa5edd9a01ac-openstack-config\") pod \"openstackclient\" (UID: \"6f6e1905-c5e6-495b-9076-fa5edd9a01ac\") " pod="openstack/openstackclient" Dec 11 08:30:15 crc kubenswrapper[4860]: I1211 08:30:15.643746 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/6f6e1905-c5e6-495b-9076-fa5edd9a01ac-openstack-config-secret\") pod \"openstackclient\" (UID: \"6f6e1905-c5e6-495b-9076-fa5edd9a01ac\") " pod="openstack/openstackclient" Dec 11 08:30:15 crc kubenswrapper[4860]: I1211 08:30:15.644358 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f6e1905-c5e6-495b-9076-fa5edd9a01ac-combined-ca-bundle\") pod \"openstackclient\" (UID: \"6f6e1905-c5e6-495b-9076-fa5edd9a01ac\") " pod="openstack/openstackclient" Dec 11 08:30:15 crc kubenswrapper[4860]: I1211 08:30:15.650001 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xtdmz\" (UniqueName: \"kubernetes.io/projected/6f6e1905-c5e6-495b-9076-fa5edd9a01ac-kube-api-access-xtdmz\") pod \"openstackclient\" (UID: \"6f6e1905-c5e6-495b-9076-fa5edd9a01ac\") " pod="openstack/openstackclient" Dec 11 08:30:15 crc kubenswrapper[4860]: I1211 08:30:15.664891 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6075b482-4926-4cf5-b632-0a731d2686dc" path="/var/lib/kubelet/pods/6075b482-4926-4cf5-b632-0a731d2686dc/volumes" Dec 11 08:30:15 crc kubenswrapper[4860]: I1211 08:30:15.671392 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 11 08:30:15 crc kubenswrapper[4860]: E1211 08:30:15.775382 4860 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8f38e033_5243_4e65_a7e0_29f1282b063f.slice/crio-conmon-948c71446dedd81878e5444ab6a4dbd7fbfe0ce699ae4e007ae7d21b188368ec.scope\": RecentStats: unable to find data in memory cache]" Dec 11 08:30:16 crc kubenswrapper[4860]: I1211 08:30:16.094402 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Dec 11 08:30:16 crc kubenswrapper[4860]: I1211 08:30:16.256303 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"5dc4bc3d-458e-4251-81f7-68b5b47a7275","Type":"ContainerStarted","Data":"ad68357b9b06703710b5befce0d84272c90aebf2fb78993edb73c05d3bebb238"} Dec 11 08:30:16 crc kubenswrapper[4860]: I1211 08:30:16.259392 4860 generic.go:334] "Generic (PLEG): container finished" podID="8f38e033-5243-4e65-a7e0-29f1282b063f" containerID="948c71446dedd81878e5444ab6a4dbd7fbfe0ce699ae4e007ae7d21b188368ec" exitCode=0 Dec 11 08:30:16 crc kubenswrapper[4860]: I1211 08:30:16.259434 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-776755b658-9fwr9" event={"ID":"8f38e033-5243-4e65-a7e0-29f1282b063f","Type":"ContainerDied","Data":"948c71446dedd81878e5444ab6a4dbd7fbfe0ce699ae4e007ae7d21b188368ec"} Dec 11 08:30:16 crc kubenswrapper[4860]: I1211 08:30:16.312151 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 11 08:30:16 crc kubenswrapper[4860]: I1211 08:30:16.699551 4860 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-776755b658-9fwr9" podUID="8f38e033-5243-4e65-a7e0-29f1282b063f" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.142:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.142:8443: connect: connection refused" Dec 11 08:30:17 crc kubenswrapper[4860]: I1211 08:30:17.275071 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"5dc4bc3d-458e-4251-81f7-68b5b47a7275","Type":"ContainerStarted","Data":"67294908d5d32c3cd661f66b7e481cec6c7573650cdb9d6004ade0eab25d27de"} Dec 11 08:30:17 crc kubenswrapper[4860]: I1211 08:30:17.280311 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"6f6e1905-c5e6-495b-9076-fa5edd9a01ac","Type":"ContainerStarted","Data":"cffc9d8dfab1f57c0ba97aa62760b15d431f20a4853d36d25dbb3ca70cfb6121"} Dec 11 08:30:17 crc kubenswrapper[4860]: I1211 08:30:17.306752 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.3067228379999998 podStartE2EDuration="3.306722838s" podCreationTimestamp="2025-12-11 08:30:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:30:17.297441286 +0000 UTC m=+1150.025960341" watchObservedRunningTime="2025-12-11 08:30:17.306722838 +0000 UTC m=+1150.035241913" Dec 11 08:30:19 crc kubenswrapper[4860]: I1211 08:30:19.626306 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 11 08:30:20 crc kubenswrapper[4860]: I1211 08:30:20.837056 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-865bbb47ff-66xpd"] Dec 11 08:30:20 crc kubenswrapper[4860]: I1211 08:30:20.839953 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-865bbb47ff-66xpd" Dec 11 08:30:20 crc kubenswrapper[4860]: I1211 08:30:20.843156 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Dec 11 08:30:20 crc kubenswrapper[4860]: I1211 08:30:20.844190 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Dec 11 08:30:20 crc kubenswrapper[4860]: I1211 08:30:20.851737 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 11 08:30:20 crc kubenswrapper[4860]: I1211 08:30:20.862030 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-865bbb47ff-66xpd"] Dec 11 08:30:20 crc kubenswrapper[4860]: I1211 08:30:20.929377 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca8cd60f-8a69-44f4-9512-387dd82af334-config-data\") pod \"swift-proxy-865bbb47ff-66xpd\" (UID: \"ca8cd60f-8a69-44f4-9512-387dd82af334\") " pod="openstack/swift-proxy-865bbb47ff-66xpd" Dec 11 08:30:20 crc kubenswrapper[4860]: I1211 08:30:20.929419 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x7h2m\" (UniqueName: \"kubernetes.io/projected/ca8cd60f-8a69-44f4-9512-387dd82af334-kube-api-access-x7h2m\") pod \"swift-proxy-865bbb47ff-66xpd\" (UID: \"ca8cd60f-8a69-44f4-9512-387dd82af334\") " pod="openstack/swift-proxy-865bbb47ff-66xpd" Dec 11 08:30:20 crc kubenswrapper[4860]: I1211 08:30:20.929570 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ca8cd60f-8a69-44f4-9512-387dd82af334-run-httpd\") pod \"swift-proxy-865bbb47ff-66xpd\" (UID: \"ca8cd60f-8a69-44f4-9512-387dd82af334\") " pod="openstack/swift-proxy-865bbb47ff-66xpd" Dec 11 08:30:20 crc kubenswrapper[4860]: I1211 08:30:20.929715 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ca8cd60f-8a69-44f4-9512-387dd82af334-log-httpd\") pod \"swift-proxy-865bbb47ff-66xpd\" (UID: \"ca8cd60f-8a69-44f4-9512-387dd82af334\") " pod="openstack/swift-proxy-865bbb47ff-66xpd" Dec 11 08:30:20 crc kubenswrapper[4860]: I1211 08:30:20.929859 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca8cd60f-8a69-44f4-9512-387dd82af334-internal-tls-certs\") pod \"swift-proxy-865bbb47ff-66xpd\" (UID: \"ca8cd60f-8a69-44f4-9512-387dd82af334\") " pod="openstack/swift-proxy-865bbb47ff-66xpd" Dec 11 08:30:20 crc kubenswrapper[4860]: I1211 08:30:20.929933 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca8cd60f-8a69-44f4-9512-387dd82af334-public-tls-certs\") pod \"swift-proxy-865bbb47ff-66xpd\" (UID: \"ca8cd60f-8a69-44f4-9512-387dd82af334\") " pod="openstack/swift-proxy-865bbb47ff-66xpd" Dec 11 08:30:20 crc kubenswrapper[4860]: I1211 08:30:20.930147 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/ca8cd60f-8a69-44f4-9512-387dd82af334-etc-swift\") pod \"swift-proxy-865bbb47ff-66xpd\" (UID: \"ca8cd60f-8a69-44f4-9512-387dd82af334\") " pod="openstack/swift-proxy-865bbb47ff-66xpd" Dec 11 08:30:20 crc kubenswrapper[4860]: I1211 08:30:20.930261 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca8cd60f-8a69-44f4-9512-387dd82af334-combined-ca-bundle\") pod \"swift-proxy-865bbb47ff-66xpd\" (UID: \"ca8cd60f-8a69-44f4-9512-387dd82af334\") " pod="openstack/swift-proxy-865bbb47ff-66xpd" Dec 11 08:30:21 crc kubenswrapper[4860]: I1211 08:30:21.030937 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/ca8cd60f-8a69-44f4-9512-387dd82af334-etc-swift\") pod \"swift-proxy-865bbb47ff-66xpd\" (UID: \"ca8cd60f-8a69-44f4-9512-387dd82af334\") " pod="openstack/swift-proxy-865bbb47ff-66xpd" Dec 11 08:30:21 crc kubenswrapper[4860]: I1211 08:30:21.031010 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca8cd60f-8a69-44f4-9512-387dd82af334-combined-ca-bundle\") pod \"swift-proxy-865bbb47ff-66xpd\" (UID: \"ca8cd60f-8a69-44f4-9512-387dd82af334\") " pod="openstack/swift-proxy-865bbb47ff-66xpd" Dec 11 08:30:21 crc kubenswrapper[4860]: I1211 08:30:21.031055 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca8cd60f-8a69-44f4-9512-387dd82af334-config-data\") pod \"swift-proxy-865bbb47ff-66xpd\" (UID: \"ca8cd60f-8a69-44f4-9512-387dd82af334\") " pod="openstack/swift-proxy-865bbb47ff-66xpd" Dec 11 08:30:21 crc kubenswrapper[4860]: I1211 08:30:21.031100 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x7h2m\" (UniqueName: \"kubernetes.io/projected/ca8cd60f-8a69-44f4-9512-387dd82af334-kube-api-access-x7h2m\") pod \"swift-proxy-865bbb47ff-66xpd\" (UID: \"ca8cd60f-8a69-44f4-9512-387dd82af334\") " pod="openstack/swift-proxy-865bbb47ff-66xpd" Dec 11 08:30:21 crc kubenswrapper[4860]: I1211 08:30:21.031953 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ca8cd60f-8a69-44f4-9512-387dd82af334-run-httpd\") pod \"swift-proxy-865bbb47ff-66xpd\" (UID: \"ca8cd60f-8a69-44f4-9512-387dd82af334\") " pod="openstack/swift-proxy-865bbb47ff-66xpd" Dec 11 08:30:21 crc kubenswrapper[4860]: I1211 08:30:21.032055 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ca8cd60f-8a69-44f4-9512-387dd82af334-run-httpd\") pod \"swift-proxy-865bbb47ff-66xpd\" (UID: \"ca8cd60f-8a69-44f4-9512-387dd82af334\") " pod="openstack/swift-proxy-865bbb47ff-66xpd" Dec 11 08:30:21 crc kubenswrapper[4860]: I1211 08:30:21.032105 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ca8cd60f-8a69-44f4-9512-387dd82af334-log-httpd\") pod \"swift-proxy-865bbb47ff-66xpd\" (UID: \"ca8cd60f-8a69-44f4-9512-387dd82af334\") " pod="openstack/swift-proxy-865bbb47ff-66xpd" Dec 11 08:30:21 crc kubenswrapper[4860]: I1211 08:30:21.032158 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca8cd60f-8a69-44f4-9512-387dd82af334-internal-tls-certs\") pod \"swift-proxy-865bbb47ff-66xpd\" (UID: \"ca8cd60f-8a69-44f4-9512-387dd82af334\") " pod="openstack/swift-proxy-865bbb47ff-66xpd" Dec 11 08:30:21 crc kubenswrapper[4860]: I1211 08:30:21.032201 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca8cd60f-8a69-44f4-9512-387dd82af334-public-tls-certs\") pod \"swift-proxy-865bbb47ff-66xpd\" (UID: \"ca8cd60f-8a69-44f4-9512-387dd82af334\") " pod="openstack/swift-proxy-865bbb47ff-66xpd" Dec 11 08:30:21 crc kubenswrapper[4860]: I1211 08:30:21.032567 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ca8cd60f-8a69-44f4-9512-387dd82af334-log-httpd\") pod \"swift-proxy-865bbb47ff-66xpd\" (UID: \"ca8cd60f-8a69-44f4-9512-387dd82af334\") " pod="openstack/swift-proxy-865bbb47ff-66xpd" Dec 11 08:30:21 crc kubenswrapper[4860]: I1211 08:30:21.036957 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca8cd60f-8a69-44f4-9512-387dd82af334-config-data\") pod \"swift-proxy-865bbb47ff-66xpd\" (UID: \"ca8cd60f-8a69-44f4-9512-387dd82af334\") " pod="openstack/swift-proxy-865bbb47ff-66xpd" Dec 11 08:30:21 crc kubenswrapper[4860]: I1211 08:30:21.037024 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca8cd60f-8a69-44f4-9512-387dd82af334-internal-tls-certs\") pod \"swift-proxy-865bbb47ff-66xpd\" (UID: \"ca8cd60f-8a69-44f4-9512-387dd82af334\") " pod="openstack/swift-proxy-865bbb47ff-66xpd" Dec 11 08:30:21 crc kubenswrapper[4860]: I1211 08:30:21.037685 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca8cd60f-8a69-44f4-9512-387dd82af334-combined-ca-bundle\") pod \"swift-proxy-865bbb47ff-66xpd\" (UID: \"ca8cd60f-8a69-44f4-9512-387dd82af334\") " pod="openstack/swift-proxy-865bbb47ff-66xpd" Dec 11 08:30:21 crc kubenswrapper[4860]: I1211 08:30:21.038385 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/ca8cd60f-8a69-44f4-9512-387dd82af334-etc-swift\") pod \"swift-proxy-865bbb47ff-66xpd\" (UID: \"ca8cd60f-8a69-44f4-9512-387dd82af334\") " pod="openstack/swift-proxy-865bbb47ff-66xpd" Dec 11 08:30:21 crc kubenswrapper[4860]: I1211 08:30:21.046769 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca8cd60f-8a69-44f4-9512-387dd82af334-public-tls-certs\") pod \"swift-proxy-865bbb47ff-66xpd\" (UID: \"ca8cd60f-8a69-44f4-9512-387dd82af334\") " pod="openstack/swift-proxy-865bbb47ff-66xpd" Dec 11 08:30:21 crc kubenswrapper[4860]: I1211 08:30:21.053492 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x7h2m\" (UniqueName: \"kubernetes.io/projected/ca8cd60f-8a69-44f4-9512-387dd82af334-kube-api-access-x7h2m\") pod \"swift-proxy-865bbb47ff-66xpd\" (UID: \"ca8cd60f-8a69-44f4-9512-387dd82af334\") " pod="openstack/swift-proxy-865bbb47ff-66xpd" Dec 11 08:30:21 crc kubenswrapper[4860]: I1211 08:30:21.164141 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-865bbb47ff-66xpd" Dec 11 08:30:22 crc kubenswrapper[4860]: I1211 08:30:22.019192 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 11 08:30:22 crc kubenswrapper[4860]: I1211 08:30:22.019998 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2e2c517e-81fe-4563-889e-af154d1f500a" containerName="sg-core" containerID="cri-o://0ebcc0b0cc22669fec7bffa7bda1858a957e5d7ac94dedf0cefa8a030cca3fe6" gracePeriod=30 Dec 11 08:30:22 crc kubenswrapper[4860]: I1211 08:30:22.019972 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2e2c517e-81fe-4563-889e-af154d1f500a" containerName="ceilometer-central-agent" containerID="cri-o://938851d2fa08ce2e25fbee8b87015ca7ff04119efecbd755fcbb891a2b000533" gracePeriod=30 Dec 11 08:30:22 crc kubenswrapper[4860]: I1211 08:30:22.020010 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2e2c517e-81fe-4563-889e-af154d1f500a" containerName="ceilometer-notification-agent" containerID="cri-o://2851fc988d20e4bf3cae08111c3baea890cba97eaace2938f550e3e6cb24f15b" gracePeriod=30 Dec 11 08:30:22 crc kubenswrapper[4860]: I1211 08:30:22.020105 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2e2c517e-81fe-4563-889e-af154d1f500a" containerName="proxy-httpd" containerID="cri-o://c139b307041de43862507eac5d2f3deee5e84a25fa1e5cc7c663e05bf5f01ad8" gracePeriod=30 Dec 11 08:30:22 crc kubenswrapper[4860]: I1211 08:30:22.025960 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 11 08:30:22 crc kubenswrapper[4860]: I1211 08:30:22.363012 4860 generic.go:334] "Generic (PLEG): container finished" podID="2e2c517e-81fe-4563-889e-af154d1f500a" containerID="c139b307041de43862507eac5d2f3deee5e84a25fa1e5cc7c663e05bf5f01ad8" exitCode=0 Dec 11 08:30:22 crc kubenswrapper[4860]: I1211 08:30:22.363054 4860 generic.go:334] "Generic (PLEG): container finished" podID="2e2c517e-81fe-4563-889e-af154d1f500a" containerID="0ebcc0b0cc22669fec7bffa7bda1858a957e5d7ac94dedf0cefa8a030cca3fe6" exitCode=2 Dec 11 08:30:22 crc kubenswrapper[4860]: I1211 08:30:22.363068 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2e2c517e-81fe-4563-889e-af154d1f500a","Type":"ContainerDied","Data":"c139b307041de43862507eac5d2f3deee5e84a25fa1e5cc7c663e05bf5f01ad8"} Dec 11 08:30:22 crc kubenswrapper[4860]: I1211 08:30:22.363483 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2e2c517e-81fe-4563-889e-af154d1f500a","Type":"ContainerDied","Data":"0ebcc0b0cc22669fec7bffa7bda1858a957e5d7ac94dedf0cefa8a030cca3fe6"} Dec 11 08:30:23 crc kubenswrapper[4860]: I1211 08:30:23.380387 4860 generic.go:334] "Generic (PLEG): container finished" podID="2e2c517e-81fe-4563-889e-af154d1f500a" containerID="938851d2fa08ce2e25fbee8b87015ca7ff04119efecbd755fcbb891a2b000533" exitCode=0 Dec 11 08:30:23 crc kubenswrapper[4860]: I1211 08:30:23.380440 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2e2c517e-81fe-4563-889e-af154d1f500a","Type":"ContainerDied","Data":"938851d2fa08ce2e25fbee8b87015ca7ff04119efecbd755fcbb891a2b000533"} Dec 11 08:30:24 crc kubenswrapper[4860]: I1211 08:30:24.975348 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 11 08:30:25 crc kubenswrapper[4860]: I1211 08:30:25.943949 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 11 08:30:26 crc kubenswrapper[4860]: I1211 08:30:26.411514 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="bbb1e117-b966-48e4-bc09-08d66a1ce9eb" containerName="kube-state-metrics" containerID="cri-o://23e919a40caf8668edf5943b9e36eecf6963a5c9dddb7d9d93a14c7cffe56e62" gracePeriod=30 Dec 11 08:30:26 crc kubenswrapper[4860]: I1211 08:30:26.696182 4860 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-776755b658-9fwr9" podUID="8f38e033-5243-4e65-a7e0-29f1282b063f" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.142:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.142:8443: connect: connection refused" Dec 11 08:30:27 crc kubenswrapper[4860]: I1211 08:30:27.426835 4860 generic.go:334] "Generic (PLEG): container finished" podID="bbb1e117-b966-48e4-bc09-08d66a1ce9eb" containerID="23e919a40caf8668edf5943b9e36eecf6963a5c9dddb7d9d93a14c7cffe56e62" exitCode=2 Dec 11 08:30:27 crc kubenswrapper[4860]: I1211 08:30:27.427107 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"bbb1e117-b966-48e4-bc09-08d66a1ce9eb","Type":"ContainerDied","Data":"23e919a40caf8668edf5943b9e36eecf6963a5c9dddb7d9d93a14c7cffe56e62"} Dec 11 08:30:27 crc kubenswrapper[4860]: I1211 08:30:27.435337 4860 generic.go:334] "Generic (PLEG): container finished" podID="2e2c517e-81fe-4563-889e-af154d1f500a" containerID="2851fc988d20e4bf3cae08111c3baea890cba97eaace2938f550e3e6cb24f15b" exitCode=0 Dec 11 08:30:27 crc kubenswrapper[4860]: I1211 08:30:27.435398 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2e2c517e-81fe-4563-889e-af154d1f500a","Type":"ContainerDied","Data":"2851fc988d20e4bf3cae08111c3baea890cba97eaace2938f550e3e6cb24f15b"} Dec 11 08:30:27 crc kubenswrapper[4860]: I1211 08:30:27.696567 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 11 08:30:27 crc kubenswrapper[4860]: I1211 08:30:27.794278 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p469v\" (UniqueName: \"kubernetes.io/projected/bbb1e117-b966-48e4-bc09-08d66a1ce9eb-kube-api-access-p469v\") pod \"bbb1e117-b966-48e4-bc09-08d66a1ce9eb\" (UID: \"bbb1e117-b966-48e4-bc09-08d66a1ce9eb\") " Dec 11 08:30:27 crc kubenswrapper[4860]: I1211 08:30:27.802880 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 11 08:30:27 crc kubenswrapper[4860]: I1211 08:30:27.802967 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bbb1e117-b966-48e4-bc09-08d66a1ce9eb-kube-api-access-p469v" (OuterVolumeSpecName: "kube-api-access-p469v") pod "bbb1e117-b966-48e4-bc09-08d66a1ce9eb" (UID: "bbb1e117-b966-48e4-bc09-08d66a1ce9eb"). InnerVolumeSpecName "kube-api-access-p469v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:30:27 crc kubenswrapper[4860]: I1211 08:30:27.897836 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e2c517e-81fe-4563-889e-af154d1f500a-config-data\") pod \"2e2c517e-81fe-4563-889e-af154d1f500a\" (UID: \"2e2c517e-81fe-4563-889e-af154d1f500a\") " Dec 11 08:30:27 crc kubenswrapper[4860]: I1211 08:30:27.897935 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e2c517e-81fe-4563-889e-af154d1f500a-scripts\") pod \"2e2c517e-81fe-4563-889e-af154d1f500a\" (UID: \"2e2c517e-81fe-4563-889e-af154d1f500a\") " Dec 11 08:30:27 crc kubenswrapper[4860]: I1211 08:30:27.898014 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e2c517e-81fe-4563-889e-af154d1f500a-log-httpd\") pod \"2e2c517e-81fe-4563-889e-af154d1f500a\" (UID: \"2e2c517e-81fe-4563-889e-af154d1f500a\") " Dec 11 08:30:27 crc kubenswrapper[4860]: I1211 08:30:27.898154 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e2c517e-81fe-4563-889e-af154d1f500a-combined-ca-bundle\") pod \"2e2c517e-81fe-4563-889e-af154d1f500a\" (UID: \"2e2c517e-81fe-4563-889e-af154d1f500a\") " Dec 11 08:30:27 crc kubenswrapper[4860]: I1211 08:30:27.898253 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5q92p\" (UniqueName: \"kubernetes.io/projected/2e2c517e-81fe-4563-889e-af154d1f500a-kube-api-access-5q92p\") pod \"2e2c517e-81fe-4563-889e-af154d1f500a\" (UID: \"2e2c517e-81fe-4563-889e-af154d1f500a\") " Dec 11 08:30:27 crc kubenswrapper[4860]: I1211 08:30:27.898290 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2e2c517e-81fe-4563-889e-af154d1f500a-sg-core-conf-yaml\") pod \"2e2c517e-81fe-4563-889e-af154d1f500a\" (UID: \"2e2c517e-81fe-4563-889e-af154d1f500a\") " Dec 11 08:30:27 crc kubenswrapper[4860]: I1211 08:30:27.898337 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e2c517e-81fe-4563-889e-af154d1f500a-run-httpd\") pod \"2e2c517e-81fe-4563-889e-af154d1f500a\" (UID: \"2e2c517e-81fe-4563-889e-af154d1f500a\") " Dec 11 08:30:27 crc kubenswrapper[4860]: I1211 08:30:27.899206 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p469v\" (UniqueName: \"kubernetes.io/projected/bbb1e117-b966-48e4-bc09-08d66a1ce9eb-kube-api-access-p469v\") on node \"crc\" DevicePath \"\"" Dec 11 08:30:27 crc kubenswrapper[4860]: I1211 08:30:27.899230 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2e2c517e-81fe-4563-889e-af154d1f500a-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "2e2c517e-81fe-4563-889e-af154d1f500a" (UID: "2e2c517e-81fe-4563-889e-af154d1f500a"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:30:27 crc kubenswrapper[4860]: I1211 08:30:27.900751 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2e2c517e-81fe-4563-889e-af154d1f500a-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "2e2c517e-81fe-4563-889e-af154d1f500a" (UID: "2e2c517e-81fe-4563-889e-af154d1f500a"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:30:27 crc kubenswrapper[4860]: I1211 08:30:27.905031 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e2c517e-81fe-4563-889e-af154d1f500a-scripts" (OuterVolumeSpecName: "scripts") pod "2e2c517e-81fe-4563-889e-af154d1f500a" (UID: "2e2c517e-81fe-4563-889e-af154d1f500a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:30:27 crc kubenswrapper[4860]: I1211 08:30:27.908968 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e2c517e-81fe-4563-889e-af154d1f500a-kube-api-access-5q92p" (OuterVolumeSpecName: "kube-api-access-5q92p") pod "2e2c517e-81fe-4563-889e-af154d1f500a" (UID: "2e2c517e-81fe-4563-889e-af154d1f500a"). InnerVolumeSpecName "kube-api-access-5q92p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:30:27 crc kubenswrapper[4860]: I1211 08:30:27.929332 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e2c517e-81fe-4563-889e-af154d1f500a-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "2e2c517e-81fe-4563-889e-af154d1f500a" (UID: "2e2c517e-81fe-4563-889e-af154d1f500a"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:30:27 crc kubenswrapper[4860]: I1211 08:30:27.991242 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e2c517e-81fe-4563-889e-af154d1f500a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2e2c517e-81fe-4563-889e-af154d1f500a" (UID: "2e2c517e-81fe-4563-889e-af154d1f500a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:30:28 crc kubenswrapper[4860]: I1211 08:30:28.000978 4860 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2e2c517e-81fe-4563-889e-af154d1f500a-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 11 08:30:28 crc kubenswrapper[4860]: I1211 08:30:28.001016 4860 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e2c517e-81fe-4563-889e-af154d1f500a-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 11 08:30:28 crc kubenswrapper[4860]: I1211 08:30:28.001028 4860 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e2c517e-81fe-4563-889e-af154d1f500a-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 08:30:28 crc kubenswrapper[4860]: I1211 08:30:28.001041 4860 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e2c517e-81fe-4563-889e-af154d1f500a-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 11 08:30:28 crc kubenswrapper[4860]: I1211 08:30:28.001054 4860 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e2c517e-81fe-4563-889e-af154d1f500a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 08:30:28 crc kubenswrapper[4860]: I1211 08:30:28.001066 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5q92p\" (UniqueName: \"kubernetes.io/projected/2e2c517e-81fe-4563-889e-af154d1f500a-kube-api-access-5q92p\") on node \"crc\" DevicePath \"\"" Dec 11 08:30:28 crc kubenswrapper[4860]: I1211 08:30:28.035060 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e2c517e-81fe-4563-889e-af154d1f500a-config-data" (OuterVolumeSpecName: "config-data") pod "2e2c517e-81fe-4563-889e-af154d1f500a" (UID: "2e2c517e-81fe-4563-889e-af154d1f500a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:30:28 crc kubenswrapper[4860]: W1211 08:30:28.047895 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podca8cd60f_8a69_44f4_9512_387dd82af334.slice/crio-e571310d6c9ec1bd2b22a35af6225a458c7be725b9523a65b36de7564b1ef54a WatchSource:0}: Error finding container e571310d6c9ec1bd2b22a35af6225a458c7be725b9523a65b36de7564b1ef54a: Status 404 returned error can't find the container with id e571310d6c9ec1bd2b22a35af6225a458c7be725b9523a65b36de7564b1ef54a Dec 11 08:30:28 crc kubenswrapper[4860]: I1211 08:30:28.049891 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-865bbb47ff-66xpd"] Dec 11 08:30:28 crc kubenswrapper[4860]: I1211 08:30:28.103445 4860 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e2c517e-81fe-4563-889e-af154d1f500a-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 08:30:28 crc kubenswrapper[4860]: I1211 08:30:28.455788 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 11 08:30:28 crc kubenswrapper[4860]: I1211 08:30:28.455797 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"bbb1e117-b966-48e4-bc09-08d66a1ce9eb","Type":"ContainerDied","Data":"5fafe2235e6146e5bdcc9cb3c753ad5723d2ba870a789157311460891393f8d7"} Dec 11 08:30:28 crc kubenswrapper[4860]: I1211 08:30:28.456194 4860 scope.go:117] "RemoveContainer" containerID="23e919a40caf8668edf5943b9e36eecf6963a5c9dddb7d9d93a14c7cffe56e62" Dec 11 08:30:28 crc kubenswrapper[4860]: I1211 08:30:28.463265 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2e2c517e-81fe-4563-889e-af154d1f500a","Type":"ContainerDied","Data":"d716be20ff6c5eb0ac20517b3c44bd25c32944509c239d09b7b182d5dbc5e195"} Dec 11 08:30:28 crc kubenswrapper[4860]: I1211 08:30:28.463334 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 11 08:30:28 crc kubenswrapper[4860]: I1211 08:30:28.466195 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-865bbb47ff-66xpd" event={"ID":"ca8cd60f-8a69-44f4-9512-387dd82af334","Type":"ContainerStarted","Data":"7c08ad706aa9d4ae87c26872a64f8e6dc5896f06acb603abb5cd8683a0c13d45"} Dec 11 08:30:28 crc kubenswrapper[4860]: I1211 08:30:28.466225 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-865bbb47ff-66xpd" event={"ID":"ca8cd60f-8a69-44f4-9512-387dd82af334","Type":"ContainerStarted","Data":"e571310d6c9ec1bd2b22a35af6225a458c7be725b9523a65b36de7564b1ef54a"} Dec 11 08:30:28 crc kubenswrapper[4860]: I1211 08:30:28.468137 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"6f6e1905-c5e6-495b-9076-fa5edd9a01ac","Type":"ContainerStarted","Data":"092ee56074295700380af2f0490976d829e16e1e0f32881f98dac20180c29cfa"} Dec 11 08:30:28 crc kubenswrapper[4860]: I1211 08:30:28.490908 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.440604256 podStartE2EDuration="13.490882427s" podCreationTimestamp="2025-12-11 08:30:15 +0000 UTC" firstStartedPulling="2025-12-11 08:30:16.32366715 +0000 UTC m=+1149.052186205" lastFinishedPulling="2025-12-11 08:30:27.373945321 +0000 UTC m=+1160.102464376" observedRunningTime="2025-12-11 08:30:28.487967315 +0000 UTC m=+1161.216486370" watchObservedRunningTime="2025-12-11 08:30:28.490882427 +0000 UTC m=+1161.219401482" Dec 11 08:30:28 crc kubenswrapper[4860]: I1211 08:30:28.515872 4860 scope.go:117] "RemoveContainer" containerID="c139b307041de43862507eac5d2f3deee5e84a25fa1e5cc7c663e05bf5f01ad8" Dec 11 08:30:28 crc kubenswrapper[4860]: I1211 08:30:28.552147 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 11 08:30:28 crc kubenswrapper[4860]: I1211 08:30:28.560856 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 11 08:30:28 crc kubenswrapper[4860]: I1211 08:30:28.562865 4860 scope.go:117] "RemoveContainer" containerID="0ebcc0b0cc22669fec7bffa7bda1858a957e5d7ac94dedf0cefa8a030cca3fe6" Dec 11 08:30:28 crc kubenswrapper[4860]: I1211 08:30:28.571690 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 11 08:30:28 crc kubenswrapper[4860]: I1211 08:30:28.580295 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 11 08:30:28 crc kubenswrapper[4860]: I1211 08:30:28.588527 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 11 08:30:28 crc kubenswrapper[4860]: E1211 08:30:28.589068 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e2c517e-81fe-4563-889e-af154d1f500a" containerName="ceilometer-notification-agent" Dec 11 08:30:28 crc kubenswrapper[4860]: I1211 08:30:28.589089 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e2c517e-81fe-4563-889e-af154d1f500a" containerName="ceilometer-notification-agent" Dec 11 08:30:28 crc kubenswrapper[4860]: E1211 08:30:28.589109 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bbb1e117-b966-48e4-bc09-08d66a1ce9eb" containerName="kube-state-metrics" Dec 11 08:30:28 crc kubenswrapper[4860]: I1211 08:30:28.589119 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="bbb1e117-b966-48e4-bc09-08d66a1ce9eb" containerName="kube-state-metrics" Dec 11 08:30:28 crc kubenswrapper[4860]: E1211 08:30:28.589134 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e2c517e-81fe-4563-889e-af154d1f500a" containerName="ceilometer-central-agent" Dec 11 08:30:28 crc kubenswrapper[4860]: I1211 08:30:28.589141 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e2c517e-81fe-4563-889e-af154d1f500a" containerName="ceilometer-central-agent" Dec 11 08:30:28 crc kubenswrapper[4860]: E1211 08:30:28.589157 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e2c517e-81fe-4563-889e-af154d1f500a" containerName="sg-core" Dec 11 08:30:28 crc kubenswrapper[4860]: I1211 08:30:28.589163 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e2c517e-81fe-4563-889e-af154d1f500a" containerName="sg-core" Dec 11 08:30:28 crc kubenswrapper[4860]: E1211 08:30:28.589172 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e2c517e-81fe-4563-889e-af154d1f500a" containerName="proxy-httpd" Dec 11 08:30:28 crc kubenswrapper[4860]: I1211 08:30:28.589178 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e2c517e-81fe-4563-889e-af154d1f500a" containerName="proxy-httpd" Dec 11 08:30:28 crc kubenswrapper[4860]: I1211 08:30:28.589353 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e2c517e-81fe-4563-889e-af154d1f500a" containerName="proxy-httpd" Dec 11 08:30:28 crc kubenswrapper[4860]: I1211 08:30:28.589374 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e2c517e-81fe-4563-889e-af154d1f500a" containerName="sg-core" Dec 11 08:30:28 crc kubenswrapper[4860]: I1211 08:30:28.589383 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e2c517e-81fe-4563-889e-af154d1f500a" containerName="ceilometer-central-agent" Dec 11 08:30:28 crc kubenswrapper[4860]: I1211 08:30:28.589392 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="bbb1e117-b966-48e4-bc09-08d66a1ce9eb" containerName="kube-state-metrics" Dec 11 08:30:28 crc kubenswrapper[4860]: I1211 08:30:28.589406 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e2c517e-81fe-4563-889e-af154d1f500a" containerName="ceilometer-notification-agent" Dec 11 08:30:28 crc kubenswrapper[4860]: I1211 08:30:28.590127 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 11 08:30:28 crc kubenswrapper[4860]: I1211 08:30:28.592605 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Dec 11 08:30:28 crc kubenswrapper[4860]: I1211 08:30:28.592724 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-8wb2b" Dec 11 08:30:28 crc kubenswrapper[4860]: I1211 08:30:28.594710 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Dec 11 08:30:28 crc kubenswrapper[4860]: I1211 08:30:28.601155 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 11 08:30:28 crc kubenswrapper[4860]: I1211 08:30:28.607168 4860 scope.go:117] "RemoveContainer" containerID="2851fc988d20e4bf3cae08111c3baea890cba97eaace2938f550e3e6cb24f15b" Dec 11 08:30:28 crc kubenswrapper[4860]: I1211 08:30:28.610165 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 11 08:30:28 crc kubenswrapper[4860]: I1211 08:30:28.625349 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eceff086-19de-4aed-8ac3-5242d9b949ab-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"eceff086-19de-4aed-8ac3-5242d9b949ab\") " pod="openstack/kube-state-metrics-0" Dec 11 08:30:28 crc kubenswrapper[4860]: I1211 08:30:28.628574 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/eceff086-19de-4aed-8ac3-5242d9b949ab-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"eceff086-19de-4aed-8ac3-5242d9b949ab\") " pod="openstack/kube-state-metrics-0" Dec 11 08:30:28 crc kubenswrapper[4860]: I1211 08:30:28.628896 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m7vbd\" (UniqueName: \"kubernetes.io/projected/eceff086-19de-4aed-8ac3-5242d9b949ab-kube-api-access-m7vbd\") pod \"kube-state-metrics-0\" (UID: \"eceff086-19de-4aed-8ac3-5242d9b949ab\") " pod="openstack/kube-state-metrics-0" Dec 11 08:30:28 crc kubenswrapper[4860]: I1211 08:30:28.629062 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/eceff086-19de-4aed-8ac3-5242d9b949ab-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"eceff086-19de-4aed-8ac3-5242d9b949ab\") " pod="openstack/kube-state-metrics-0" Dec 11 08:30:28 crc kubenswrapper[4860]: I1211 08:30:28.636685 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 11 08:30:28 crc kubenswrapper[4860]: I1211 08:30:28.642169 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 11 08:30:28 crc kubenswrapper[4860]: I1211 08:30:28.642612 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 11 08:30:28 crc kubenswrapper[4860]: I1211 08:30:28.642920 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 11 08:30:28 crc kubenswrapper[4860]: I1211 08:30:28.654597 4860 scope.go:117] "RemoveContainer" containerID="938851d2fa08ce2e25fbee8b87015ca7ff04119efecbd755fcbb891a2b000533" Dec 11 08:30:28 crc kubenswrapper[4860]: I1211 08:30:28.670929 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 11 08:30:28 crc kubenswrapper[4860]: I1211 08:30:28.731733 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fb9e2dd0-591f-4a0e-88aa-5f21285fbc77-scripts\") pod \"ceilometer-0\" (UID: \"fb9e2dd0-591f-4a0e-88aa-5f21285fbc77\") " pod="openstack/ceilometer-0" Dec 11 08:30:28 crc kubenswrapper[4860]: I1211 08:30:28.731828 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/eceff086-19de-4aed-8ac3-5242d9b949ab-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"eceff086-19de-4aed-8ac3-5242d9b949ab\") " pod="openstack/kube-state-metrics-0" Dec 11 08:30:28 crc kubenswrapper[4860]: I1211 08:30:28.731899 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb9e2dd0-591f-4a0e-88aa-5f21285fbc77-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"fb9e2dd0-591f-4a0e-88aa-5f21285fbc77\") " pod="openstack/ceilometer-0" Dec 11 08:30:28 crc kubenswrapper[4860]: I1211 08:30:28.731948 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2c7rl\" (UniqueName: \"kubernetes.io/projected/fb9e2dd0-591f-4a0e-88aa-5f21285fbc77-kube-api-access-2c7rl\") pod \"ceilometer-0\" (UID: \"fb9e2dd0-591f-4a0e-88aa-5f21285fbc77\") " pod="openstack/ceilometer-0" Dec 11 08:30:28 crc kubenswrapper[4860]: I1211 08:30:28.731971 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m7vbd\" (UniqueName: \"kubernetes.io/projected/eceff086-19de-4aed-8ac3-5242d9b949ab-kube-api-access-m7vbd\") pod \"kube-state-metrics-0\" (UID: \"eceff086-19de-4aed-8ac3-5242d9b949ab\") " pod="openstack/kube-state-metrics-0" Dec 11 08:30:28 crc kubenswrapper[4860]: I1211 08:30:28.731996 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb9e2dd0-591f-4a0e-88aa-5f21285fbc77-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"fb9e2dd0-591f-4a0e-88aa-5f21285fbc77\") " pod="openstack/ceilometer-0" Dec 11 08:30:28 crc kubenswrapper[4860]: I1211 08:30:28.732062 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fb9e2dd0-591f-4a0e-88aa-5f21285fbc77-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"fb9e2dd0-591f-4a0e-88aa-5f21285fbc77\") " pod="openstack/ceilometer-0" Dec 11 08:30:28 crc kubenswrapper[4860]: I1211 08:30:28.732103 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/eceff086-19de-4aed-8ac3-5242d9b949ab-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"eceff086-19de-4aed-8ac3-5242d9b949ab\") " pod="openstack/kube-state-metrics-0" Dec 11 08:30:28 crc kubenswrapper[4860]: I1211 08:30:28.732158 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eceff086-19de-4aed-8ac3-5242d9b949ab-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"eceff086-19de-4aed-8ac3-5242d9b949ab\") " pod="openstack/kube-state-metrics-0" Dec 11 08:30:28 crc kubenswrapper[4860]: I1211 08:30:28.732178 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fb9e2dd0-591f-4a0e-88aa-5f21285fbc77-log-httpd\") pod \"ceilometer-0\" (UID: \"fb9e2dd0-591f-4a0e-88aa-5f21285fbc77\") " pod="openstack/ceilometer-0" Dec 11 08:30:28 crc kubenswrapper[4860]: I1211 08:30:28.732219 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fb9e2dd0-591f-4a0e-88aa-5f21285fbc77-run-httpd\") pod \"ceilometer-0\" (UID: \"fb9e2dd0-591f-4a0e-88aa-5f21285fbc77\") " pod="openstack/ceilometer-0" Dec 11 08:30:28 crc kubenswrapper[4860]: I1211 08:30:28.732241 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb9e2dd0-591f-4a0e-88aa-5f21285fbc77-config-data\") pod \"ceilometer-0\" (UID: \"fb9e2dd0-591f-4a0e-88aa-5f21285fbc77\") " pod="openstack/ceilometer-0" Dec 11 08:30:28 crc kubenswrapper[4860]: I1211 08:30:28.737797 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/eceff086-19de-4aed-8ac3-5242d9b949ab-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"eceff086-19de-4aed-8ac3-5242d9b949ab\") " pod="openstack/kube-state-metrics-0" Dec 11 08:30:28 crc kubenswrapper[4860]: I1211 08:30:28.738504 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/eceff086-19de-4aed-8ac3-5242d9b949ab-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"eceff086-19de-4aed-8ac3-5242d9b949ab\") " pod="openstack/kube-state-metrics-0" Dec 11 08:30:28 crc kubenswrapper[4860]: I1211 08:30:28.749144 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m7vbd\" (UniqueName: \"kubernetes.io/projected/eceff086-19de-4aed-8ac3-5242d9b949ab-kube-api-access-m7vbd\") pod \"kube-state-metrics-0\" (UID: \"eceff086-19de-4aed-8ac3-5242d9b949ab\") " pod="openstack/kube-state-metrics-0" Dec 11 08:30:28 crc kubenswrapper[4860]: I1211 08:30:28.756980 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eceff086-19de-4aed-8ac3-5242d9b949ab-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"eceff086-19de-4aed-8ac3-5242d9b949ab\") " pod="openstack/kube-state-metrics-0" Dec 11 08:30:28 crc kubenswrapper[4860]: I1211 08:30:28.833667 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fb9e2dd0-591f-4a0e-88aa-5f21285fbc77-log-httpd\") pod \"ceilometer-0\" (UID: \"fb9e2dd0-591f-4a0e-88aa-5f21285fbc77\") " pod="openstack/ceilometer-0" Dec 11 08:30:28 crc kubenswrapper[4860]: I1211 08:30:28.833737 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fb9e2dd0-591f-4a0e-88aa-5f21285fbc77-run-httpd\") pod \"ceilometer-0\" (UID: \"fb9e2dd0-591f-4a0e-88aa-5f21285fbc77\") " pod="openstack/ceilometer-0" Dec 11 08:30:28 crc kubenswrapper[4860]: I1211 08:30:28.833763 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb9e2dd0-591f-4a0e-88aa-5f21285fbc77-config-data\") pod \"ceilometer-0\" (UID: \"fb9e2dd0-591f-4a0e-88aa-5f21285fbc77\") " pod="openstack/ceilometer-0" Dec 11 08:30:28 crc kubenswrapper[4860]: I1211 08:30:28.833821 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fb9e2dd0-591f-4a0e-88aa-5f21285fbc77-scripts\") pod \"ceilometer-0\" (UID: \"fb9e2dd0-591f-4a0e-88aa-5f21285fbc77\") " pod="openstack/ceilometer-0" Dec 11 08:30:28 crc kubenswrapper[4860]: I1211 08:30:28.833864 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb9e2dd0-591f-4a0e-88aa-5f21285fbc77-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"fb9e2dd0-591f-4a0e-88aa-5f21285fbc77\") " pod="openstack/ceilometer-0" Dec 11 08:30:28 crc kubenswrapper[4860]: I1211 08:30:28.833893 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2c7rl\" (UniqueName: \"kubernetes.io/projected/fb9e2dd0-591f-4a0e-88aa-5f21285fbc77-kube-api-access-2c7rl\") pod \"ceilometer-0\" (UID: \"fb9e2dd0-591f-4a0e-88aa-5f21285fbc77\") " pod="openstack/ceilometer-0" Dec 11 08:30:28 crc kubenswrapper[4860]: I1211 08:30:28.833920 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb9e2dd0-591f-4a0e-88aa-5f21285fbc77-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"fb9e2dd0-591f-4a0e-88aa-5f21285fbc77\") " pod="openstack/ceilometer-0" Dec 11 08:30:28 crc kubenswrapper[4860]: I1211 08:30:28.833943 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fb9e2dd0-591f-4a0e-88aa-5f21285fbc77-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"fb9e2dd0-591f-4a0e-88aa-5f21285fbc77\") " pod="openstack/ceilometer-0" Dec 11 08:30:28 crc kubenswrapper[4860]: I1211 08:30:28.835011 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fb9e2dd0-591f-4a0e-88aa-5f21285fbc77-run-httpd\") pod \"ceilometer-0\" (UID: \"fb9e2dd0-591f-4a0e-88aa-5f21285fbc77\") " pod="openstack/ceilometer-0" Dec 11 08:30:28 crc kubenswrapper[4860]: I1211 08:30:28.835077 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fb9e2dd0-591f-4a0e-88aa-5f21285fbc77-log-httpd\") pod \"ceilometer-0\" (UID: \"fb9e2dd0-591f-4a0e-88aa-5f21285fbc77\") " pod="openstack/ceilometer-0" Dec 11 08:30:28 crc kubenswrapper[4860]: I1211 08:30:28.838787 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb9e2dd0-591f-4a0e-88aa-5f21285fbc77-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"fb9e2dd0-591f-4a0e-88aa-5f21285fbc77\") " pod="openstack/ceilometer-0" Dec 11 08:30:28 crc kubenswrapper[4860]: I1211 08:30:28.839171 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fb9e2dd0-591f-4a0e-88aa-5f21285fbc77-scripts\") pod \"ceilometer-0\" (UID: \"fb9e2dd0-591f-4a0e-88aa-5f21285fbc77\") " pod="openstack/ceilometer-0" Dec 11 08:30:28 crc kubenswrapper[4860]: I1211 08:30:28.839175 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fb9e2dd0-591f-4a0e-88aa-5f21285fbc77-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"fb9e2dd0-591f-4a0e-88aa-5f21285fbc77\") " pod="openstack/ceilometer-0" Dec 11 08:30:28 crc kubenswrapper[4860]: I1211 08:30:28.851629 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb9e2dd0-591f-4a0e-88aa-5f21285fbc77-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"fb9e2dd0-591f-4a0e-88aa-5f21285fbc77\") " pod="openstack/ceilometer-0" Dec 11 08:30:28 crc kubenswrapper[4860]: I1211 08:30:28.851903 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb9e2dd0-591f-4a0e-88aa-5f21285fbc77-config-data\") pod \"ceilometer-0\" (UID: \"fb9e2dd0-591f-4a0e-88aa-5f21285fbc77\") " pod="openstack/ceilometer-0" Dec 11 08:30:28 crc kubenswrapper[4860]: I1211 08:30:28.858268 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2c7rl\" (UniqueName: \"kubernetes.io/projected/fb9e2dd0-591f-4a0e-88aa-5f21285fbc77-kube-api-access-2c7rl\") pod \"ceilometer-0\" (UID: \"fb9e2dd0-591f-4a0e-88aa-5f21285fbc77\") " pod="openstack/ceilometer-0" Dec 11 08:30:28 crc kubenswrapper[4860]: I1211 08:30:28.932519 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 11 08:30:28 crc kubenswrapper[4860]: I1211 08:30:28.981329 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 11 08:30:29 crc kubenswrapper[4860]: I1211 08:30:29.447632 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 11 08:30:29 crc kubenswrapper[4860]: W1211 08:30:29.461237 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfb9e2dd0_591f_4a0e_88aa_5f21285fbc77.slice/crio-a496c61f077dcb30e98e70f519be52f2cb0fe7bff1cedf2fa450405d885f360b WatchSource:0}: Error finding container a496c61f077dcb30e98e70f519be52f2cb0fe7bff1cedf2fa450405d885f360b: Status 404 returned error can't find the container with id a496c61f077dcb30e98e70f519be52f2cb0fe7bff1cedf2fa450405d885f360b Dec 11 08:30:29 crc kubenswrapper[4860]: I1211 08:30:29.485857 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 11 08:30:29 crc kubenswrapper[4860]: I1211 08:30:29.498201 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-865bbb47ff-66xpd" event={"ID":"ca8cd60f-8a69-44f4-9512-387dd82af334","Type":"ContainerStarted","Data":"2cd5a06f079f020f54ba3e310b83480fb25ac9b333730d2e1cb483f9439fff7c"} Dec 11 08:30:29 crc kubenswrapper[4860]: I1211 08:30:29.498665 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-865bbb47ff-66xpd" Dec 11 08:30:29 crc kubenswrapper[4860]: I1211 08:30:29.498768 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-865bbb47ff-66xpd" Dec 11 08:30:29 crc kubenswrapper[4860]: I1211 08:30:29.503724 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fb9e2dd0-591f-4a0e-88aa-5f21285fbc77","Type":"ContainerStarted","Data":"a496c61f077dcb30e98e70f519be52f2cb0fe7bff1cedf2fa450405d885f360b"} Dec 11 08:30:29 crc kubenswrapper[4860]: I1211 08:30:29.528415 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-865bbb47ff-66xpd" podStartSLOduration=9.528392197 podStartE2EDuration="9.528392197s" podCreationTimestamp="2025-12-11 08:30:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:30:29.515946756 +0000 UTC m=+1162.244465811" watchObservedRunningTime="2025-12-11 08:30:29.528392197 +0000 UTC m=+1162.256911252" Dec 11 08:30:29 crc kubenswrapper[4860]: I1211 08:30:29.592464 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2e2c517e-81fe-4563-889e-af154d1f500a" path="/var/lib/kubelet/pods/2e2c517e-81fe-4563-889e-af154d1f500a/volumes" Dec 11 08:30:29 crc kubenswrapper[4860]: I1211 08:30:29.593187 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bbb1e117-b966-48e4-bc09-08d66a1ce9eb" path="/var/lib/kubelet/pods/bbb1e117-b966-48e4-bc09-08d66a1ce9eb/volumes" Dec 11 08:30:30 crc kubenswrapper[4860]: I1211 08:30:30.516371 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"eceff086-19de-4aed-8ac3-5242d9b949ab","Type":"ContainerStarted","Data":"cce827c651106f54b4828cea85c785fe4feb5eb498491aea955d302678b31eb0"} Dec 11 08:30:31 crc kubenswrapper[4860]: I1211 08:30:31.984866 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 11 08:30:32 crc kubenswrapper[4860]: I1211 08:30:32.555690 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fb9e2dd0-591f-4a0e-88aa-5f21285fbc77","Type":"ContainerStarted","Data":"2333689d78c9ee84e0452445e3f70919ddfeeccadb977eb6bd41e9ac35f23fd0"} Dec 11 08:30:32 crc kubenswrapper[4860]: I1211 08:30:32.557919 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"eceff086-19de-4aed-8ac3-5242d9b949ab","Type":"ContainerStarted","Data":"ce347b44e932e5d50391e091d719a212e8ee9686790a3ee89bbf8a07267463fd"} Dec 11 08:30:32 crc kubenswrapper[4860]: I1211 08:30:32.558081 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 11 08:30:32 crc kubenswrapper[4860]: I1211 08:30:32.592047 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=4.087051687 podStartE2EDuration="4.592027028s" podCreationTimestamp="2025-12-11 08:30:28 +0000 UTC" firstStartedPulling="2025-12-11 08:30:29.484742305 +0000 UTC m=+1162.213261360" lastFinishedPulling="2025-12-11 08:30:29.989717646 +0000 UTC m=+1162.718236701" observedRunningTime="2025-12-11 08:30:32.579759653 +0000 UTC m=+1165.308278718" watchObservedRunningTime="2025-12-11 08:30:32.592027028 +0000 UTC m=+1165.320546083" Dec 11 08:30:33 crc kubenswrapper[4860]: I1211 08:30:33.220810 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-hc9mt"] Dec 11 08:30:33 crc kubenswrapper[4860]: I1211 08:30:33.222672 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-hc9mt" Dec 11 08:30:33 crc kubenswrapper[4860]: I1211 08:30:33.237405 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-hc9mt"] Dec 11 08:30:33 crc kubenswrapper[4860]: I1211 08:30:33.305475 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-7bmlr"] Dec 11 08:30:33 crc kubenswrapper[4860]: I1211 08:30:33.315594 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-7bmlr" Dec 11 08:30:33 crc kubenswrapper[4860]: I1211 08:30:33.322103 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b4598c5c-a584-487e-83ed-0126b5fd253e-operator-scripts\") pod \"nova-api-db-create-hc9mt\" (UID: \"b4598c5c-a584-487e-83ed-0126b5fd253e\") " pod="openstack/nova-api-db-create-hc9mt" Dec 11 08:30:33 crc kubenswrapper[4860]: I1211 08:30:33.322290 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-7bmlr"] Dec 11 08:30:33 crc kubenswrapper[4860]: I1211 08:30:33.322304 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k7w52\" (UniqueName: \"kubernetes.io/projected/b4598c5c-a584-487e-83ed-0126b5fd253e-kube-api-access-k7w52\") pod \"nova-api-db-create-hc9mt\" (UID: \"b4598c5c-a584-487e-83ed-0126b5fd253e\") " pod="openstack/nova-api-db-create-hc9mt" Dec 11 08:30:33 crc kubenswrapper[4860]: I1211 08:30:33.406603 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-vzvdr"] Dec 11 08:30:33 crc kubenswrapper[4860]: I1211 08:30:33.408165 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-vzvdr" Dec 11 08:30:33 crc kubenswrapper[4860]: I1211 08:30:33.416009 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-7093-account-create-update-4m7w4"] Dec 11 08:30:33 crc kubenswrapper[4860]: I1211 08:30:33.417794 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-7093-account-create-update-4m7w4" Dec 11 08:30:33 crc kubenswrapper[4860]: I1211 08:30:33.421346 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Dec 11 08:30:33 crc kubenswrapper[4860]: I1211 08:30:33.425009 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b4598c5c-a584-487e-83ed-0126b5fd253e-operator-scripts\") pod \"nova-api-db-create-hc9mt\" (UID: \"b4598c5c-a584-487e-83ed-0126b5fd253e\") " pod="openstack/nova-api-db-create-hc9mt" Dec 11 08:30:33 crc kubenswrapper[4860]: I1211 08:30:33.425157 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k7w52\" (UniqueName: \"kubernetes.io/projected/b4598c5c-a584-487e-83ed-0126b5fd253e-kube-api-access-k7w52\") pod \"nova-api-db-create-hc9mt\" (UID: \"b4598c5c-a584-487e-83ed-0126b5fd253e\") " pod="openstack/nova-api-db-create-hc9mt" Dec 11 08:30:33 crc kubenswrapper[4860]: I1211 08:30:33.425229 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e01fc44a-9c71-49bb-bc43-f4a20d44b41a-operator-scripts\") pod \"nova-cell0-db-create-7bmlr\" (UID: \"e01fc44a-9c71-49bb-bc43-f4a20d44b41a\") " pod="openstack/nova-cell0-db-create-7bmlr" Dec 11 08:30:33 crc kubenswrapper[4860]: I1211 08:30:33.425304 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q8vhc\" (UniqueName: \"kubernetes.io/projected/e01fc44a-9c71-49bb-bc43-f4a20d44b41a-kube-api-access-q8vhc\") pod \"nova-cell0-db-create-7bmlr\" (UID: \"e01fc44a-9c71-49bb-bc43-f4a20d44b41a\") " pod="openstack/nova-cell0-db-create-7bmlr" Dec 11 08:30:33 crc kubenswrapper[4860]: I1211 08:30:33.425895 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b4598c5c-a584-487e-83ed-0126b5fd253e-operator-scripts\") pod \"nova-api-db-create-hc9mt\" (UID: \"b4598c5c-a584-487e-83ed-0126b5fd253e\") " pod="openstack/nova-api-db-create-hc9mt" Dec 11 08:30:33 crc kubenswrapper[4860]: I1211 08:30:33.429293 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-vzvdr"] Dec 11 08:30:33 crc kubenswrapper[4860]: I1211 08:30:33.446319 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-7093-account-create-update-4m7w4"] Dec 11 08:30:33 crc kubenswrapper[4860]: I1211 08:30:33.491026 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k7w52\" (UniqueName: \"kubernetes.io/projected/b4598c5c-a584-487e-83ed-0126b5fd253e-kube-api-access-k7w52\") pod \"nova-api-db-create-hc9mt\" (UID: \"b4598c5c-a584-487e-83ed-0126b5fd253e\") " pod="openstack/nova-api-db-create-hc9mt" Dec 11 08:30:33 crc kubenswrapper[4860]: I1211 08:30:33.528464 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/31d35fc5-e254-4e3f-86bf-9a944f8a44af-operator-scripts\") pod \"nova-cell1-db-create-vzvdr\" (UID: \"31d35fc5-e254-4e3f-86bf-9a944f8a44af\") " pod="openstack/nova-cell1-db-create-vzvdr" Dec 11 08:30:33 crc kubenswrapper[4860]: I1211 08:30:33.528774 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-28z7f\" (UniqueName: \"kubernetes.io/projected/293660ed-0401-45cc-811e-a612da96284f-kube-api-access-28z7f\") pod \"nova-api-7093-account-create-update-4m7w4\" (UID: \"293660ed-0401-45cc-811e-a612da96284f\") " pod="openstack/nova-api-7093-account-create-update-4m7w4" Dec 11 08:30:33 crc kubenswrapper[4860]: I1211 08:30:33.528830 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e01fc44a-9c71-49bb-bc43-f4a20d44b41a-operator-scripts\") pod \"nova-cell0-db-create-7bmlr\" (UID: \"e01fc44a-9c71-49bb-bc43-f4a20d44b41a\") " pod="openstack/nova-cell0-db-create-7bmlr" Dec 11 08:30:33 crc kubenswrapper[4860]: I1211 08:30:33.528849 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/293660ed-0401-45cc-811e-a612da96284f-operator-scripts\") pod \"nova-api-7093-account-create-update-4m7w4\" (UID: \"293660ed-0401-45cc-811e-a612da96284f\") " pod="openstack/nova-api-7093-account-create-update-4m7w4" Dec 11 08:30:33 crc kubenswrapper[4860]: I1211 08:30:33.528897 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q8vhc\" (UniqueName: \"kubernetes.io/projected/e01fc44a-9c71-49bb-bc43-f4a20d44b41a-kube-api-access-q8vhc\") pod \"nova-cell0-db-create-7bmlr\" (UID: \"e01fc44a-9c71-49bb-bc43-f4a20d44b41a\") " pod="openstack/nova-cell0-db-create-7bmlr" Dec 11 08:30:33 crc kubenswrapper[4860]: I1211 08:30:33.528956 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fhszw\" (UniqueName: \"kubernetes.io/projected/31d35fc5-e254-4e3f-86bf-9a944f8a44af-kube-api-access-fhszw\") pod \"nova-cell1-db-create-vzvdr\" (UID: \"31d35fc5-e254-4e3f-86bf-9a944f8a44af\") " pod="openstack/nova-cell1-db-create-vzvdr" Dec 11 08:30:33 crc kubenswrapper[4860]: I1211 08:30:33.529724 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e01fc44a-9c71-49bb-bc43-f4a20d44b41a-operator-scripts\") pod \"nova-cell0-db-create-7bmlr\" (UID: \"e01fc44a-9c71-49bb-bc43-f4a20d44b41a\") " pod="openstack/nova-cell0-db-create-7bmlr" Dec 11 08:30:33 crc kubenswrapper[4860]: I1211 08:30:33.547067 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q8vhc\" (UniqueName: \"kubernetes.io/projected/e01fc44a-9c71-49bb-bc43-f4a20d44b41a-kube-api-access-q8vhc\") pod \"nova-cell0-db-create-7bmlr\" (UID: \"e01fc44a-9c71-49bb-bc43-f4a20d44b41a\") " pod="openstack/nova-cell0-db-create-7bmlr" Dec 11 08:30:33 crc kubenswrapper[4860]: I1211 08:30:33.558141 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-hc9mt" Dec 11 08:30:33 crc kubenswrapper[4860]: I1211 08:30:33.617391 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fb9e2dd0-591f-4a0e-88aa-5f21285fbc77","Type":"ContainerStarted","Data":"967f589ab90b87704934b82341d0412adecad46c59feec2439df3412a407ac73"} Dec 11 08:30:33 crc kubenswrapper[4860]: I1211 08:30:33.626632 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-b7a3-account-create-update-qgft5"] Dec 11 08:30:33 crc kubenswrapper[4860]: I1211 08:30:33.628153 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-b7a3-account-create-update-qgft5" Dec 11 08:30:33 crc kubenswrapper[4860]: I1211 08:30:33.630311 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fhszw\" (UniqueName: \"kubernetes.io/projected/31d35fc5-e254-4e3f-86bf-9a944f8a44af-kube-api-access-fhszw\") pod \"nova-cell1-db-create-vzvdr\" (UID: \"31d35fc5-e254-4e3f-86bf-9a944f8a44af\") " pod="openstack/nova-cell1-db-create-vzvdr" Dec 11 08:30:33 crc kubenswrapper[4860]: I1211 08:30:33.630411 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/31d35fc5-e254-4e3f-86bf-9a944f8a44af-operator-scripts\") pod \"nova-cell1-db-create-vzvdr\" (UID: \"31d35fc5-e254-4e3f-86bf-9a944f8a44af\") " pod="openstack/nova-cell1-db-create-vzvdr" Dec 11 08:30:33 crc kubenswrapper[4860]: I1211 08:30:33.630443 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-28z7f\" (UniqueName: \"kubernetes.io/projected/293660ed-0401-45cc-811e-a612da96284f-kube-api-access-28z7f\") pod \"nova-api-7093-account-create-update-4m7w4\" (UID: \"293660ed-0401-45cc-811e-a612da96284f\") " pod="openstack/nova-api-7093-account-create-update-4m7w4" Dec 11 08:30:33 crc kubenswrapper[4860]: I1211 08:30:33.630495 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/293660ed-0401-45cc-811e-a612da96284f-operator-scripts\") pod \"nova-api-7093-account-create-update-4m7w4\" (UID: \"293660ed-0401-45cc-811e-a612da96284f\") " pod="openstack/nova-api-7093-account-create-update-4m7w4" Dec 11 08:30:33 crc kubenswrapper[4860]: I1211 08:30:33.631896 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/293660ed-0401-45cc-811e-a612da96284f-operator-scripts\") pod \"nova-api-7093-account-create-update-4m7w4\" (UID: \"293660ed-0401-45cc-811e-a612da96284f\") " pod="openstack/nova-api-7093-account-create-update-4m7w4" Dec 11 08:30:33 crc kubenswrapper[4860]: I1211 08:30:33.632033 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/31d35fc5-e254-4e3f-86bf-9a944f8a44af-operator-scripts\") pod \"nova-cell1-db-create-vzvdr\" (UID: \"31d35fc5-e254-4e3f-86bf-9a944f8a44af\") " pod="openstack/nova-cell1-db-create-vzvdr" Dec 11 08:30:33 crc kubenswrapper[4860]: I1211 08:30:33.635039 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Dec 11 08:30:33 crc kubenswrapper[4860]: I1211 08:30:33.638685 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-b7a3-account-create-update-qgft5"] Dec 11 08:30:33 crc kubenswrapper[4860]: I1211 08:30:33.661987 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fhszw\" (UniqueName: \"kubernetes.io/projected/31d35fc5-e254-4e3f-86bf-9a944f8a44af-kube-api-access-fhszw\") pod \"nova-cell1-db-create-vzvdr\" (UID: \"31d35fc5-e254-4e3f-86bf-9a944f8a44af\") " pod="openstack/nova-cell1-db-create-vzvdr" Dec 11 08:30:33 crc kubenswrapper[4860]: I1211 08:30:33.674973 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-7bmlr" Dec 11 08:30:33 crc kubenswrapper[4860]: I1211 08:30:33.675016 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-28z7f\" (UniqueName: \"kubernetes.io/projected/293660ed-0401-45cc-811e-a612da96284f-kube-api-access-28z7f\") pod \"nova-api-7093-account-create-update-4m7w4\" (UID: \"293660ed-0401-45cc-811e-a612da96284f\") " pod="openstack/nova-api-7093-account-create-update-4m7w4" Dec 11 08:30:33 crc kubenswrapper[4860]: I1211 08:30:33.732153 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fdae480c-a60d-4508-add1-e1e89bd3e8e3-operator-scripts\") pod \"nova-cell0-b7a3-account-create-update-qgft5\" (UID: \"fdae480c-a60d-4508-add1-e1e89bd3e8e3\") " pod="openstack/nova-cell0-b7a3-account-create-update-qgft5" Dec 11 08:30:33 crc kubenswrapper[4860]: I1211 08:30:33.732510 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dfb55\" (UniqueName: \"kubernetes.io/projected/fdae480c-a60d-4508-add1-e1e89bd3e8e3-kube-api-access-dfb55\") pod \"nova-cell0-b7a3-account-create-update-qgft5\" (UID: \"fdae480c-a60d-4508-add1-e1e89bd3e8e3\") " pod="openstack/nova-cell0-b7a3-account-create-update-qgft5" Dec 11 08:30:33 crc kubenswrapper[4860]: I1211 08:30:33.734259 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-vzvdr" Dec 11 08:30:33 crc kubenswrapper[4860]: I1211 08:30:33.748436 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-7093-account-create-update-4m7w4" Dec 11 08:30:33 crc kubenswrapper[4860]: I1211 08:30:33.818736 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-caef-account-create-update-jwjlc"] Dec 11 08:30:33 crc kubenswrapper[4860]: I1211 08:30:33.820281 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-caef-account-create-update-jwjlc" Dec 11 08:30:33 crc kubenswrapper[4860]: I1211 08:30:33.827879 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Dec 11 08:30:33 crc kubenswrapper[4860]: I1211 08:30:33.834929 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-caef-account-create-update-jwjlc"] Dec 11 08:30:33 crc kubenswrapper[4860]: I1211 08:30:33.834991 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fdae480c-a60d-4508-add1-e1e89bd3e8e3-operator-scripts\") pod \"nova-cell0-b7a3-account-create-update-qgft5\" (UID: \"fdae480c-a60d-4508-add1-e1e89bd3e8e3\") " pod="openstack/nova-cell0-b7a3-account-create-update-qgft5" Dec 11 08:30:33 crc kubenswrapper[4860]: I1211 08:30:33.835061 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dfb55\" (UniqueName: \"kubernetes.io/projected/fdae480c-a60d-4508-add1-e1e89bd3e8e3-kube-api-access-dfb55\") pod \"nova-cell0-b7a3-account-create-update-qgft5\" (UID: \"fdae480c-a60d-4508-add1-e1e89bd3e8e3\") " pod="openstack/nova-cell0-b7a3-account-create-update-qgft5" Dec 11 08:30:33 crc kubenswrapper[4860]: I1211 08:30:33.837068 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fdae480c-a60d-4508-add1-e1e89bd3e8e3-operator-scripts\") pod \"nova-cell0-b7a3-account-create-update-qgft5\" (UID: \"fdae480c-a60d-4508-add1-e1e89bd3e8e3\") " pod="openstack/nova-cell0-b7a3-account-create-update-qgft5" Dec 11 08:30:33 crc kubenswrapper[4860]: I1211 08:30:33.867709 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dfb55\" (UniqueName: \"kubernetes.io/projected/fdae480c-a60d-4508-add1-e1e89bd3e8e3-kube-api-access-dfb55\") pod \"nova-cell0-b7a3-account-create-update-qgft5\" (UID: \"fdae480c-a60d-4508-add1-e1e89bd3e8e3\") " pod="openstack/nova-cell0-b7a3-account-create-update-qgft5" Dec 11 08:30:33 crc kubenswrapper[4860]: I1211 08:30:33.938913 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6a7614eb-5a8b-44ca-afe2-22a8b1151065-operator-scripts\") pod \"nova-cell1-caef-account-create-update-jwjlc\" (UID: \"6a7614eb-5a8b-44ca-afe2-22a8b1151065\") " pod="openstack/nova-cell1-caef-account-create-update-jwjlc" Dec 11 08:30:33 crc kubenswrapper[4860]: I1211 08:30:33.939010 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zzwwl\" (UniqueName: \"kubernetes.io/projected/6a7614eb-5a8b-44ca-afe2-22a8b1151065-kube-api-access-zzwwl\") pod \"nova-cell1-caef-account-create-update-jwjlc\" (UID: \"6a7614eb-5a8b-44ca-afe2-22a8b1151065\") " pod="openstack/nova-cell1-caef-account-create-update-jwjlc" Dec 11 08:30:34 crc kubenswrapper[4860]: I1211 08:30:34.041277 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zzwwl\" (UniqueName: \"kubernetes.io/projected/6a7614eb-5a8b-44ca-afe2-22a8b1151065-kube-api-access-zzwwl\") pod \"nova-cell1-caef-account-create-update-jwjlc\" (UID: \"6a7614eb-5a8b-44ca-afe2-22a8b1151065\") " pod="openstack/nova-cell1-caef-account-create-update-jwjlc" Dec 11 08:30:34 crc kubenswrapper[4860]: I1211 08:30:34.041705 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6a7614eb-5a8b-44ca-afe2-22a8b1151065-operator-scripts\") pod \"nova-cell1-caef-account-create-update-jwjlc\" (UID: \"6a7614eb-5a8b-44ca-afe2-22a8b1151065\") " pod="openstack/nova-cell1-caef-account-create-update-jwjlc" Dec 11 08:30:34 crc kubenswrapper[4860]: I1211 08:30:34.055169 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6a7614eb-5a8b-44ca-afe2-22a8b1151065-operator-scripts\") pod \"nova-cell1-caef-account-create-update-jwjlc\" (UID: \"6a7614eb-5a8b-44ca-afe2-22a8b1151065\") " pod="openstack/nova-cell1-caef-account-create-update-jwjlc" Dec 11 08:30:34 crc kubenswrapper[4860]: I1211 08:30:34.069181 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zzwwl\" (UniqueName: \"kubernetes.io/projected/6a7614eb-5a8b-44ca-afe2-22a8b1151065-kube-api-access-zzwwl\") pod \"nova-cell1-caef-account-create-update-jwjlc\" (UID: \"6a7614eb-5a8b-44ca-afe2-22a8b1151065\") " pod="openstack/nova-cell1-caef-account-create-update-jwjlc" Dec 11 08:30:34 crc kubenswrapper[4860]: I1211 08:30:34.107573 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-b7a3-account-create-update-qgft5" Dec 11 08:30:34 crc kubenswrapper[4860]: I1211 08:30:34.145805 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-hc9mt"] Dec 11 08:30:34 crc kubenswrapper[4860]: I1211 08:30:34.167402 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-caef-account-create-update-jwjlc" Dec 11 08:30:34 crc kubenswrapper[4860]: I1211 08:30:34.300486 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-7bmlr"] Dec 11 08:30:34 crc kubenswrapper[4860]: W1211 08:30:34.321309 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode01fc44a_9c71_49bb_bc43_f4a20d44b41a.slice/crio-92ccb2b36b176354594ded05902a8b9a880bb0084ce5527271d59cede2c118a3 WatchSource:0}: Error finding container 92ccb2b36b176354594ded05902a8b9a880bb0084ce5527271d59cede2c118a3: Status 404 returned error can't find the container with id 92ccb2b36b176354594ded05902a8b9a880bb0084ce5527271d59cede2c118a3 Dec 11 08:30:34 crc kubenswrapper[4860]: I1211 08:30:34.494421 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-vzvdr"] Dec 11 08:30:34 crc kubenswrapper[4860]: I1211 08:30:34.511554 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-7093-account-create-update-4m7w4"] Dec 11 08:30:34 crc kubenswrapper[4860]: I1211 08:30:34.628244 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-hc9mt" event={"ID":"b4598c5c-a584-487e-83ed-0126b5fd253e","Type":"ContainerStarted","Data":"d5b4720e97798804622ed48c582631bd6b457e29bca173b7f14ad4cc5860fe14"} Dec 11 08:30:34 crc kubenswrapper[4860]: I1211 08:30:34.628306 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-hc9mt" event={"ID":"b4598c5c-a584-487e-83ed-0126b5fd253e","Type":"ContainerStarted","Data":"06b65ff37789f6104d5d2e21c482d9c68dfca6de50b945f6ee92a9fbf6935d06"} Dec 11 08:30:34 crc kubenswrapper[4860]: I1211 08:30:34.638450 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-vzvdr" event={"ID":"31d35fc5-e254-4e3f-86bf-9a944f8a44af","Type":"ContainerStarted","Data":"98a9eb3b7b11acc8fb27e81e4bff1554dec6e80cd787804b9a961eabf6a0596d"} Dec 11 08:30:34 crc kubenswrapper[4860]: I1211 08:30:34.642405 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-7bmlr" event={"ID":"e01fc44a-9c71-49bb-bc43-f4a20d44b41a","Type":"ContainerStarted","Data":"a8dfc96d64f0aed235dcb8437d52cc0a75402373d7c90da2ec678b3aecf52e4f"} Dec 11 08:30:34 crc kubenswrapper[4860]: I1211 08:30:34.642472 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-7bmlr" event={"ID":"e01fc44a-9c71-49bb-bc43-f4a20d44b41a","Type":"ContainerStarted","Data":"92ccb2b36b176354594ded05902a8b9a880bb0084ce5527271d59cede2c118a3"} Dec 11 08:30:34 crc kubenswrapper[4860]: I1211 08:30:34.643788 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-7093-account-create-update-4m7w4" event={"ID":"293660ed-0401-45cc-811e-a612da96284f","Type":"ContainerStarted","Data":"9dc66b8425abc81f3979c19435b4698a3069e8b8bc6a06a1a8daeb2f9ae06dd5"} Dec 11 08:30:34 crc kubenswrapper[4860]: I1211 08:30:34.650013 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fb9e2dd0-591f-4a0e-88aa-5f21285fbc77","Type":"ContainerStarted","Data":"bc42015dc1ea63f1842e6450d911e1e49b5ea23e5bf2c6aaeeab303b302b119a"} Dec 11 08:30:34 crc kubenswrapper[4860]: I1211 08:30:34.659899 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-db-create-hc9mt" podStartSLOduration=1.659844905 podStartE2EDuration="1.659844905s" podCreationTimestamp="2025-12-11 08:30:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:30:34.649371854 +0000 UTC m=+1167.377890919" watchObservedRunningTime="2025-12-11 08:30:34.659844905 +0000 UTC m=+1167.388363960" Dec 11 08:30:34 crc kubenswrapper[4860]: I1211 08:30:34.679780 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-db-create-7bmlr" podStartSLOduration=1.679756434 podStartE2EDuration="1.679756434s" podCreationTimestamp="2025-12-11 08:30:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:30:34.679195446 +0000 UTC m=+1167.407714511" watchObservedRunningTime="2025-12-11 08:30:34.679756434 +0000 UTC m=+1167.408275499" Dec 11 08:30:34 crc kubenswrapper[4860]: I1211 08:30:34.749268 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-b7a3-account-create-update-qgft5"] Dec 11 08:30:34 crc kubenswrapper[4860]: I1211 08:30:34.847901 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-caef-account-create-update-jwjlc"] Dec 11 08:30:34 crc kubenswrapper[4860]: W1211 08:30:34.853601 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6a7614eb_5a8b_44ca_afe2_22a8b1151065.slice/crio-c6aace91493bb1377e7c78627d5e4eb67f43c7a5ea3d1eab7c8efb7329ee2090 WatchSource:0}: Error finding container c6aace91493bb1377e7c78627d5e4eb67f43c7a5ea3d1eab7c8efb7329ee2090: Status 404 returned error can't find the container with id c6aace91493bb1377e7c78627d5e4eb67f43c7a5ea3d1eab7c8efb7329ee2090 Dec 11 08:30:35 crc kubenswrapper[4860]: I1211 08:30:35.665558 4860 generic.go:334] "Generic (PLEG): container finished" podID="b4598c5c-a584-487e-83ed-0126b5fd253e" containerID="d5b4720e97798804622ed48c582631bd6b457e29bca173b7f14ad4cc5860fe14" exitCode=0 Dec 11 08:30:35 crc kubenswrapper[4860]: I1211 08:30:35.665728 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-hc9mt" event={"ID":"b4598c5c-a584-487e-83ed-0126b5fd253e","Type":"ContainerDied","Data":"d5b4720e97798804622ed48c582631bd6b457e29bca173b7f14ad4cc5860fe14"} Dec 11 08:30:35 crc kubenswrapper[4860]: I1211 08:30:35.668918 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-b7a3-account-create-update-qgft5" event={"ID":"fdae480c-a60d-4508-add1-e1e89bd3e8e3","Type":"ContainerStarted","Data":"b217915b36edf21ab47701779f8cba774a5f626850151fb5e7bd9eeb8370d4d6"} Dec 11 08:30:35 crc kubenswrapper[4860]: I1211 08:30:35.668950 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-b7a3-account-create-update-qgft5" event={"ID":"fdae480c-a60d-4508-add1-e1e89bd3e8e3","Type":"ContainerStarted","Data":"46f7bbc83c743a6fd3b0aacd93c3f748f1bf03298be9d84d8d2cec7609810842"} Dec 11 08:30:35 crc kubenswrapper[4860]: I1211 08:30:35.676994 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-vzvdr" event={"ID":"31d35fc5-e254-4e3f-86bf-9a944f8a44af","Type":"ContainerStarted","Data":"ec87f01347028dc991f26c633f3ca640d51582e9ba4678b414d94b21b0235ef4"} Dec 11 08:30:35 crc kubenswrapper[4860]: I1211 08:30:35.681225 4860 generic.go:334] "Generic (PLEG): container finished" podID="e01fc44a-9c71-49bb-bc43-f4a20d44b41a" containerID="a8dfc96d64f0aed235dcb8437d52cc0a75402373d7c90da2ec678b3aecf52e4f" exitCode=0 Dec 11 08:30:35 crc kubenswrapper[4860]: I1211 08:30:35.681308 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-7bmlr" event={"ID":"e01fc44a-9c71-49bb-bc43-f4a20d44b41a","Type":"ContainerDied","Data":"a8dfc96d64f0aed235dcb8437d52cc0a75402373d7c90da2ec678b3aecf52e4f"} Dec 11 08:30:35 crc kubenswrapper[4860]: I1211 08:30:35.687113 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-7093-account-create-update-4m7w4" event={"ID":"293660ed-0401-45cc-811e-a612da96284f","Type":"ContainerStarted","Data":"4de218dffc3e8e24efb4fe4428903323ac601fe2154ece0ae358ea367ca41eb9"} Dec 11 08:30:35 crc kubenswrapper[4860]: I1211 08:30:35.689592 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-caef-account-create-update-jwjlc" event={"ID":"6a7614eb-5a8b-44ca-afe2-22a8b1151065","Type":"ContainerStarted","Data":"be956e7e27260afa9cb66d3f8d92a9d6ba13ffd3eb5429dc7166b50c41401383"} Dec 11 08:30:35 crc kubenswrapper[4860]: I1211 08:30:35.689694 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-caef-account-create-update-jwjlc" event={"ID":"6a7614eb-5a8b-44ca-afe2-22a8b1151065","Type":"ContainerStarted","Data":"c6aace91493bb1377e7c78627d5e4eb67f43c7a5ea3d1eab7c8efb7329ee2090"} Dec 11 08:30:35 crc kubenswrapper[4860]: I1211 08:30:35.692351 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fb9e2dd0-591f-4a0e-88aa-5f21285fbc77","Type":"ContainerStarted","Data":"0654f036c46ff9fe2261a835fd570095454a3493304ad137b4dc5ee75c6a22cb"} Dec 11 08:30:35 crc kubenswrapper[4860]: I1211 08:30:35.692491 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="fb9e2dd0-591f-4a0e-88aa-5f21285fbc77" containerName="ceilometer-central-agent" containerID="cri-o://2333689d78c9ee84e0452445e3f70919ddfeeccadb977eb6bd41e9ac35f23fd0" gracePeriod=30 Dec 11 08:30:35 crc kubenswrapper[4860]: I1211 08:30:35.692554 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 11 08:30:35 crc kubenswrapper[4860]: I1211 08:30:35.692596 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="fb9e2dd0-591f-4a0e-88aa-5f21285fbc77" containerName="sg-core" containerID="cri-o://bc42015dc1ea63f1842e6450d911e1e49b5ea23e5bf2c6aaeeab303b302b119a" gracePeriod=30 Dec 11 08:30:35 crc kubenswrapper[4860]: I1211 08:30:35.692609 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="fb9e2dd0-591f-4a0e-88aa-5f21285fbc77" containerName="ceilometer-notification-agent" containerID="cri-o://967f589ab90b87704934b82341d0412adecad46c59feec2439df3412a407ac73" gracePeriod=30 Dec 11 08:30:35 crc kubenswrapper[4860]: I1211 08:30:35.692697 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="fb9e2dd0-591f-4a0e-88aa-5f21285fbc77" containerName="proxy-httpd" containerID="cri-o://0654f036c46ff9fe2261a835fd570095454a3493304ad137b4dc5ee75c6a22cb" gracePeriod=30 Dec 11 08:30:35 crc kubenswrapper[4860]: I1211 08:30:35.710291 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-db-create-vzvdr" podStartSLOduration=2.710271485 podStartE2EDuration="2.710271485s" podCreationTimestamp="2025-12-11 08:30:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:30:35.698607668 +0000 UTC m=+1168.427126753" watchObservedRunningTime="2025-12-11 08:30:35.710271485 +0000 UTC m=+1168.438790540" Dec 11 08:30:35 crc kubenswrapper[4860]: I1211 08:30:35.744370 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-b7a3-account-create-update-qgft5" podStartSLOduration=2.744337382 podStartE2EDuration="2.744337382s" podCreationTimestamp="2025-12-11 08:30:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:30:35.718809875 +0000 UTC m=+1168.447328940" watchObservedRunningTime="2025-12-11 08:30:35.744337382 +0000 UTC m=+1168.472856437" Dec 11 08:30:35 crc kubenswrapper[4860]: I1211 08:30:35.754296 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.89225178 podStartE2EDuration="7.754268656s" podCreationTimestamp="2025-12-11 08:30:28 +0000 UTC" firstStartedPulling="2025-12-11 08:30:29.470809667 +0000 UTC m=+1162.199328722" lastFinishedPulling="2025-12-11 08:30:35.332826543 +0000 UTC m=+1168.061345598" observedRunningTime="2025-12-11 08:30:35.750269829 +0000 UTC m=+1168.478788884" watchObservedRunningTime="2025-12-11 08:30:35.754268656 +0000 UTC m=+1168.482787711" Dec 11 08:30:35 crc kubenswrapper[4860]: I1211 08:30:35.782962 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-caef-account-create-update-jwjlc" podStartSLOduration=2.782933881 podStartE2EDuration="2.782933881s" podCreationTimestamp="2025-12-11 08:30:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:30:35.769181616 +0000 UTC m=+1168.497700681" watchObservedRunningTime="2025-12-11 08:30:35.782933881 +0000 UTC m=+1168.511452956" Dec 11 08:30:35 crc kubenswrapper[4860]: I1211 08:30:35.806332 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-7093-account-create-update-4m7w4" podStartSLOduration=2.806309309 podStartE2EDuration="2.806309309s" podCreationTimestamp="2025-12-11 08:30:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:30:35.803915344 +0000 UTC m=+1168.532434409" watchObservedRunningTime="2025-12-11 08:30:35.806309309 +0000 UTC m=+1168.534828364" Dec 11 08:30:36 crc kubenswrapper[4860]: I1211 08:30:36.171407 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-865bbb47ff-66xpd" Dec 11 08:30:36 crc kubenswrapper[4860]: I1211 08:30:36.178654 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-865bbb47ff-66xpd" Dec 11 08:30:36 crc kubenswrapper[4860]: I1211 08:30:36.698518 4860 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-776755b658-9fwr9" podUID="8f38e033-5243-4e65-a7e0-29f1282b063f" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.142:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.142:8443: connect: connection refused" Dec 11 08:30:36 crc kubenswrapper[4860]: I1211 08:30:36.698679 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-776755b658-9fwr9" Dec 11 08:30:36 crc kubenswrapper[4860]: I1211 08:30:36.702873 4860 generic.go:334] "Generic (PLEG): container finished" podID="fdae480c-a60d-4508-add1-e1e89bd3e8e3" containerID="b217915b36edf21ab47701779f8cba774a5f626850151fb5e7bd9eeb8370d4d6" exitCode=0 Dec 11 08:30:36 crc kubenswrapper[4860]: I1211 08:30:36.702952 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-b7a3-account-create-update-qgft5" event={"ID":"fdae480c-a60d-4508-add1-e1e89bd3e8e3","Type":"ContainerDied","Data":"b217915b36edf21ab47701779f8cba774a5f626850151fb5e7bd9eeb8370d4d6"} Dec 11 08:30:36 crc kubenswrapper[4860]: I1211 08:30:36.705126 4860 generic.go:334] "Generic (PLEG): container finished" podID="31d35fc5-e254-4e3f-86bf-9a944f8a44af" containerID="ec87f01347028dc991f26c633f3ca640d51582e9ba4678b414d94b21b0235ef4" exitCode=0 Dec 11 08:30:36 crc kubenswrapper[4860]: I1211 08:30:36.705193 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-vzvdr" event={"ID":"31d35fc5-e254-4e3f-86bf-9a944f8a44af","Type":"ContainerDied","Data":"ec87f01347028dc991f26c633f3ca640d51582e9ba4678b414d94b21b0235ef4"} Dec 11 08:30:36 crc kubenswrapper[4860]: I1211 08:30:36.707232 4860 generic.go:334] "Generic (PLEG): container finished" podID="293660ed-0401-45cc-811e-a612da96284f" containerID="4de218dffc3e8e24efb4fe4428903323ac601fe2154ece0ae358ea367ca41eb9" exitCode=0 Dec 11 08:30:36 crc kubenswrapper[4860]: I1211 08:30:36.707275 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-7093-account-create-update-4m7w4" event={"ID":"293660ed-0401-45cc-811e-a612da96284f","Type":"ContainerDied","Data":"4de218dffc3e8e24efb4fe4428903323ac601fe2154ece0ae358ea367ca41eb9"} Dec 11 08:30:36 crc kubenswrapper[4860]: I1211 08:30:36.709011 4860 generic.go:334] "Generic (PLEG): container finished" podID="6a7614eb-5a8b-44ca-afe2-22a8b1151065" containerID="be956e7e27260afa9cb66d3f8d92a9d6ba13ffd3eb5429dc7166b50c41401383" exitCode=0 Dec 11 08:30:36 crc kubenswrapper[4860]: I1211 08:30:36.709068 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-caef-account-create-update-jwjlc" event={"ID":"6a7614eb-5a8b-44ca-afe2-22a8b1151065","Type":"ContainerDied","Data":"be956e7e27260afa9cb66d3f8d92a9d6ba13ffd3eb5429dc7166b50c41401383"} Dec 11 08:30:36 crc kubenswrapper[4860]: I1211 08:30:36.713003 4860 generic.go:334] "Generic (PLEG): container finished" podID="fb9e2dd0-591f-4a0e-88aa-5f21285fbc77" containerID="bc42015dc1ea63f1842e6450d911e1e49b5ea23e5bf2c6aaeeab303b302b119a" exitCode=2 Dec 11 08:30:36 crc kubenswrapper[4860]: I1211 08:30:36.713041 4860 generic.go:334] "Generic (PLEG): container finished" podID="fb9e2dd0-591f-4a0e-88aa-5f21285fbc77" containerID="967f589ab90b87704934b82341d0412adecad46c59feec2439df3412a407ac73" exitCode=0 Dec 11 08:30:36 crc kubenswrapper[4860]: I1211 08:30:36.713036 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fb9e2dd0-591f-4a0e-88aa-5f21285fbc77","Type":"ContainerDied","Data":"bc42015dc1ea63f1842e6450d911e1e49b5ea23e5bf2c6aaeeab303b302b119a"} Dec 11 08:30:36 crc kubenswrapper[4860]: I1211 08:30:36.713154 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fb9e2dd0-591f-4a0e-88aa-5f21285fbc77","Type":"ContainerDied","Data":"967f589ab90b87704934b82341d0412adecad46c59feec2439df3412a407ac73"} Dec 11 08:30:37 crc kubenswrapper[4860]: I1211 08:30:37.163972 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-7bmlr" Dec 11 08:30:37 crc kubenswrapper[4860]: I1211 08:30:37.170538 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-hc9mt" Dec 11 08:30:37 crc kubenswrapper[4860]: I1211 08:30:37.273776 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b4598c5c-a584-487e-83ed-0126b5fd253e-operator-scripts\") pod \"b4598c5c-a584-487e-83ed-0126b5fd253e\" (UID: \"b4598c5c-a584-487e-83ed-0126b5fd253e\") " Dec 11 08:30:37 crc kubenswrapper[4860]: I1211 08:30:37.274176 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e01fc44a-9c71-49bb-bc43-f4a20d44b41a-operator-scripts\") pod \"e01fc44a-9c71-49bb-bc43-f4a20d44b41a\" (UID: \"e01fc44a-9c71-49bb-bc43-f4a20d44b41a\") " Dec 11 08:30:37 crc kubenswrapper[4860]: I1211 08:30:37.274336 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q8vhc\" (UniqueName: \"kubernetes.io/projected/e01fc44a-9c71-49bb-bc43-f4a20d44b41a-kube-api-access-q8vhc\") pod \"e01fc44a-9c71-49bb-bc43-f4a20d44b41a\" (UID: \"e01fc44a-9c71-49bb-bc43-f4a20d44b41a\") " Dec 11 08:30:37 crc kubenswrapper[4860]: I1211 08:30:37.274558 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k7w52\" (UniqueName: \"kubernetes.io/projected/b4598c5c-a584-487e-83ed-0126b5fd253e-kube-api-access-k7w52\") pod \"b4598c5c-a584-487e-83ed-0126b5fd253e\" (UID: \"b4598c5c-a584-487e-83ed-0126b5fd253e\") " Dec 11 08:30:37 crc kubenswrapper[4860]: I1211 08:30:37.274668 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e01fc44a-9c71-49bb-bc43-f4a20d44b41a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e01fc44a-9c71-49bb-bc43-f4a20d44b41a" (UID: "e01fc44a-9c71-49bb-bc43-f4a20d44b41a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:30:37 crc kubenswrapper[4860]: I1211 08:30:37.274667 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b4598c5c-a584-487e-83ed-0126b5fd253e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b4598c5c-a584-487e-83ed-0126b5fd253e" (UID: "b4598c5c-a584-487e-83ed-0126b5fd253e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:30:37 crc kubenswrapper[4860]: I1211 08:30:37.275312 4860 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b4598c5c-a584-487e-83ed-0126b5fd253e-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 08:30:37 crc kubenswrapper[4860]: I1211 08:30:37.275727 4860 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e01fc44a-9c71-49bb-bc43-f4a20d44b41a-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 08:30:37 crc kubenswrapper[4860]: I1211 08:30:37.280910 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e01fc44a-9c71-49bb-bc43-f4a20d44b41a-kube-api-access-q8vhc" (OuterVolumeSpecName: "kube-api-access-q8vhc") pod "e01fc44a-9c71-49bb-bc43-f4a20d44b41a" (UID: "e01fc44a-9c71-49bb-bc43-f4a20d44b41a"). InnerVolumeSpecName "kube-api-access-q8vhc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:30:37 crc kubenswrapper[4860]: I1211 08:30:37.283627 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b4598c5c-a584-487e-83ed-0126b5fd253e-kube-api-access-k7w52" (OuterVolumeSpecName: "kube-api-access-k7w52") pod "b4598c5c-a584-487e-83ed-0126b5fd253e" (UID: "b4598c5c-a584-487e-83ed-0126b5fd253e"). InnerVolumeSpecName "kube-api-access-k7w52". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:30:37 crc kubenswrapper[4860]: I1211 08:30:37.378015 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q8vhc\" (UniqueName: \"kubernetes.io/projected/e01fc44a-9c71-49bb-bc43-f4a20d44b41a-kube-api-access-q8vhc\") on node \"crc\" DevicePath \"\"" Dec 11 08:30:37 crc kubenswrapper[4860]: I1211 08:30:37.378055 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k7w52\" (UniqueName: \"kubernetes.io/projected/b4598c5c-a584-487e-83ed-0126b5fd253e-kube-api-access-k7w52\") on node \"crc\" DevicePath \"\"" Dec 11 08:30:37 crc kubenswrapper[4860]: I1211 08:30:37.723201 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-7bmlr" Dec 11 08:30:37 crc kubenswrapper[4860]: I1211 08:30:37.723359 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-7bmlr" event={"ID":"e01fc44a-9c71-49bb-bc43-f4a20d44b41a","Type":"ContainerDied","Data":"92ccb2b36b176354594ded05902a8b9a880bb0084ce5527271d59cede2c118a3"} Dec 11 08:30:37 crc kubenswrapper[4860]: I1211 08:30:37.723405 4860 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="92ccb2b36b176354594ded05902a8b9a880bb0084ce5527271d59cede2c118a3" Dec 11 08:30:37 crc kubenswrapper[4860]: I1211 08:30:37.725540 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-hc9mt" Dec 11 08:30:37 crc kubenswrapper[4860]: I1211 08:30:37.725531 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-hc9mt" event={"ID":"b4598c5c-a584-487e-83ed-0126b5fd253e","Type":"ContainerDied","Data":"06b65ff37789f6104d5d2e21c482d9c68dfca6de50b945f6ee92a9fbf6935d06"} Dec 11 08:30:37 crc kubenswrapper[4860]: I1211 08:30:37.725756 4860 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="06b65ff37789f6104d5d2e21c482d9c68dfca6de50b945f6ee92a9fbf6935d06" Dec 11 08:30:38 crc kubenswrapper[4860]: I1211 08:30:38.056000 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 11 08:30:38 crc kubenswrapper[4860]: I1211 08:30:38.056627 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="a3d65995-e403-4e92-b7aa-b39b8084d1bc" containerName="glance-log" containerID="cri-o://e0910e8fa92a5f01a5e9fcf10245b4f567cd1f91da1a21ce559bfaf3ce103f95" gracePeriod=30 Dec 11 08:30:38 crc kubenswrapper[4860]: I1211 08:30:38.057201 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="a3d65995-e403-4e92-b7aa-b39b8084d1bc" containerName="glance-httpd" containerID="cri-o://b2623116284268f161ab7f24e48bff42f3f1d8de1c286d66200076c14429dc28" gracePeriod=30 Dec 11 08:30:38 crc kubenswrapper[4860]: I1211 08:30:38.285136 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-b7a3-account-create-update-qgft5" Dec 11 08:30:38 crc kubenswrapper[4860]: I1211 08:30:38.419032 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-caef-account-create-update-jwjlc" Dec 11 08:30:38 crc kubenswrapper[4860]: I1211 08:30:38.422922 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fdae480c-a60d-4508-add1-e1e89bd3e8e3-operator-scripts\") pod \"fdae480c-a60d-4508-add1-e1e89bd3e8e3\" (UID: \"fdae480c-a60d-4508-add1-e1e89bd3e8e3\") " Dec 11 08:30:38 crc kubenswrapper[4860]: I1211 08:30:38.424723 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dfb55\" (UniqueName: \"kubernetes.io/projected/fdae480c-a60d-4508-add1-e1e89bd3e8e3-kube-api-access-dfb55\") pod \"fdae480c-a60d-4508-add1-e1e89bd3e8e3\" (UID: \"fdae480c-a60d-4508-add1-e1e89bd3e8e3\") " Dec 11 08:30:38 crc kubenswrapper[4860]: I1211 08:30:38.424678 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fdae480c-a60d-4508-add1-e1e89bd3e8e3-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "fdae480c-a60d-4508-add1-e1e89bd3e8e3" (UID: "fdae480c-a60d-4508-add1-e1e89bd3e8e3"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:30:38 crc kubenswrapper[4860]: I1211 08:30:38.427098 4860 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fdae480c-a60d-4508-add1-e1e89bd3e8e3-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 08:30:38 crc kubenswrapper[4860]: I1211 08:30:38.428000 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-7093-account-create-update-4m7w4" Dec 11 08:30:38 crc kubenswrapper[4860]: I1211 08:30:38.429133 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fdae480c-a60d-4508-add1-e1e89bd3e8e3-kube-api-access-dfb55" (OuterVolumeSpecName: "kube-api-access-dfb55") pod "fdae480c-a60d-4508-add1-e1e89bd3e8e3" (UID: "fdae480c-a60d-4508-add1-e1e89bd3e8e3"). InnerVolumeSpecName "kube-api-access-dfb55". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:30:38 crc kubenswrapper[4860]: I1211 08:30:38.448312 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-vzvdr" Dec 11 08:30:38 crc kubenswrapper[4860]: I1211 08:30:38.527873 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6a7614eb-5a8b-44ca-afe2-22a8b1151065-operator-scripts\") pod \"6a7614eb-5a8b-44ca-afe2-22a8b1151065\" (UID: \"6a7614eb-5a8b-44ca-afe2-22a8b1151065\") " Dec 11 08:30:38 crc kubenswrapper[4860]: I1211 08:30:38.528762 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6a7614eb-5a8b-44ca-afe2-22a8b1151065-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "6a7614eb-5a8b-44ca-afe2-22a8b1151065" (UID: "6a7614eb-5a8b-44ca-afe2-22a8b1151065"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:30:38 crc kubenswrapper[4860]: I1211 08:30:38.528919 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zzwwl\" (UniqueName: \"kubernetes.io/projected/6a7614eb-5a8b-44ca-afe2-22a8b1151065-kube-api-access-zzwwl\") pod \"6a7614eb-5a8b-44ca-afe2-22a8b1151065\" (UID: \"6a7614eb-5a8b-44ca-afe2-22a8b1151065\") " Dec 11 08:30:38 crc kubenswrapper[4860]: I1211 08:30:38.529491 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/293660ed-0401-45cc-811e-a612da96284f-operator-scripts\") pod \"293660ed-0401-45cc-811e-a612da96284f\" (UID: \"293660ed-0401-45cc-811e-a612da96284f\") " Dec 11 08:30:38 crc kubenswrapper[4860]: I1211 08:30:38.529523 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-28z7f\" (UniqueName: \"kubernetes.io/projected/293660ed-0401-45cc-811e-a612da96284f-kube-api-access-28z7f\") pod \"293660ed-0401-45cc-811e-a612da96284f\" (UID: \"293660ed-0401-45cc-811e-a612da96284f\") " Dec 11 08:30:38 crc kubenswrapper[4860]: I1211 08:30:38.530000 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/293660ed-0401-45cc-811e-a612da96284f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "293660ed-0401-45cc-811e-a612da96284f" (UID: "293660ed-0401-45cc-811e-a612da96284f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:30:38 crc kubenswrapper[4860]: I1211 08:30:38.530696 4860 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/293660ed-0401-45cc-811e-a612da96284f-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 08:30:38 crc kubenswrapper[4860]: I1211 08:30:38.530718 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dfb55\" (UniqueName: \"kubernetes.io/projected/fdae480c-a60d-4508-add1-e1e89bd3e8e3-kube-api-access-dfb55\") on node \"crc\" DevicePath \"\"" Dec 11 08:30:38 crc kubenswrapper[4860]: I1211 08:30:38.530731 4860 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6a7614eb-5a8b-44ca-afe2-22a8b1151065-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 08:30:38 crc kubenswrapper[4860]: I1211 08:30:38.533941 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a7614eb-5a8b-44ca-afe2-22a8b1151065-kube-api-access-zzwwl" (OuterVolumeSpecName: "kube-api-access-zzwwl") pod "6a7614eb-5a8b-44ca-afe2-22a8b1151065" (UID: "6a7614eb-5a8b-44ca-afe2-22a8b1151065"). InnerVolumeSpecName "kube-api-access-zzwwl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:30:38 crc kubenswrapper[4860]: I1211 08:30:38.535264 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/293660ed-0401-45cc-811e-a612da96284f-kube-api-access-28z7f" (OuterVolumeSpecName: "kube-api-access-28z7f") pod "293660ed-0401-45cc-811e-a612da96284f" (UID: "293660ed-0401-45cc-811e-a612da96284f"). InnerVolumeSpecName "kube-api-access-28z7f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:30:38 crc kubenswrapper[4860]: I1211 08:30:38.631599 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fhszw\" (UniqueName: \"kubernetes.io/projected/31d35fc5-e254-4e3f-86bf-9a944f8a44af-kube-api-access-fhszw\") pod \"31d35fc5-e254-4e3f-86bf-9a944f8a44af\" (UID: \"31d35fc5-e254-4e3f-86bf-9a944f8a44af\") " Dec 11 08:30:38 crc kubenswrapper[4860]: I1211 08:30:38.631881 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/31d35fc5-e254-4e3f-86bf-9a944f8a44af-operator-scripts\") pod \"31d35fc5-e254-4e3f-86bf-9a944f8a44af\" (UID: \"31d35fc5-e254-4e3f-86bf-9a944f8a44af\") " Dec 11 08:30:38 crc kubenswrapper[4860]: I1211 08:30:38.632334 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d35fc5-e254-4e3f-86bf-9a944f8a44af-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "31d35fc5-e254-4e3f-86bf-9a944f8a44af" (UID: "31d35fc5-e254-4e3f-86bf-9a944f8a44af"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:30:38 crc kubenswrapper[4860]: I1211 08:30:38.632385 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zzwwl\" (UniqueName: \"kubernetes.io/projected/6a7614eb-5a8b-44ca-afe2-22a8b1151065-kube-api-access-zzwwl\") on node \"crc\" DevicePath \"\"" Dec 11 08:30:38 crc kubenswrapper[4860]: I1211 08:30:38.632401 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-28z7f\" (UniqueName: \"kubernetes.io/projected/293660ed-0401-45cc-811e-a612da96284f-kube-api-access-28z7f\") on node \"crc\" DevicePath \"\"" Dec 11 08:30:38 crc kubenswrapper[4860]: I1211 08:30:38.636146 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d35fc5-e254-4e3f-86bf-9a944f8a44af-kube-api-access-fhszw" (OuterVolumeSpecName: "kube-api-access-fhszw") pod "31d35fc5-e254-4e3f-86bf-9a944f8a44af" (UID: "31d35fc5-e254-4e3f-86bf-9a944f8a44af"). InnerVolumeSpecName "kube-api-access-fhszw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:30:38 crc kubenswrapper[4860]: I1211 08:30:38.734836 4860 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/31d35fc5-e254-4e3f-86bf-9a944f8a44af-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 08:30:38 crc kubenswrapper[4860]: I1211 08:30:38.734868 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fhszw\" (UniqueName: \"kubernetes.io/projected/31d35fc5-e254-4e3f-86bf-9a944f8a44af-kube-api-access-fhszw\") on node \"crc\" DevicePath \"\"" Dec 11 08:30:38 crc kubenswrapper[4860]: I1211 08:30:38.738877 4860 generic.go:334] "Generic (PLEG): container finished" podID="a3d65995-e403-4e92-b7aa-b39b8084d1bc" containerID="e0910e8fa92a5f01a5e9fcf10245b4f567cd1f91da1a21ce559bfaf3ce103f95" exitCode=143 Dec 11 08:30:38 crc kubenswrapper[4860]: I1211 08:30:38.738975 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"a3d65995-e403-4e92-b7aa-b39b8084d1bc","Type":"ContainerDied","Data":"e0910e8fa92a5f01a5e9fcf10245b4f567cd1f91da1a21ce559bfaf3ce103f95"} Dec 11 08:30:38 crc kubenswrapper[4860]: I1211 08:30:38.740685 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-vzvdr" event={"ID":"31d35fc5-e254-4e3f-86bf-9a944f8a44af","Type":"ContainerDied","Data":"98a9eb3b7b11acc8fb27e81e4bff1554dec6e80cd787804b9a961eabf6a0596d"} Dec 11 08:30:38 crc kubenswrapper[4860]: I1211 08:30:38.740715 4860 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="98a9eb3b7b11acc8fb27e81e4bff1554dec6e80cd787804b9a961eabf6a0596d" Dec 11 08:30:38 crc kubenswrapper[4860]: I1211 08:30:38.740778 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-vzvdr" Dec 11 08:30:38 crc kubenswrapper[4860]: I1211 08:30:38.754386 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-7093-account-create-update-4m7w4" event={"ID":"293660ed-0401-45cc-811e-a612da96284f","Type":"ContainerDied","Data":"9dc66b8425abc81f3979c19435b4698a3069e8b8bc6a06a1a8daeb2f9ae06dd5"} Dec 11 08:30:38 crc kubenswrapper[4860]: I1211 08:30:38.754440 4860 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9dc66b8425abc81f3979c19435b4698a3069e8b8bc6a06a1a8daeb2f9ae06dd5" Dec 11 08:30:38 crc kubenswrapper[4860]: I1211 08:30:38.754509 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-7093-account-create-update-4m7w4" Dec 11 08:30:38 crc kubenswrapper[4860]: I1211 08:30:38.758959 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-caef-account-create-update-jwjlc" Dec 11 08:30:38 crc kubenswrapper[4860]: I1211 08:30:38.759740 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-caef-account-create-update-jwjlc" event={"ID":"6a7614eb-5a8b-44ca-afe2-22a8b1151065","Type":"ContainerDied","Data":"c6aace91493bb1377e7c78627d5e4eb67f43c7a5ea3d1eab7c8efb7329ee2090"} Dec 11 08:30:38 crc kubenswrapper[4860]: I1211 08:30:38.759788 4860 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c6aace91493bb1377e7c78627d5e4eb67f43c7a5ea3d1eab7c8efb7329ee2090" Dec 11 08:30:38 crc kubenswrapper[4860]: I1211 08:30:38.762183 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-b7a3-account-create-update-qgft5" event={"ID":"fdae480c-a60d-4508-add1-e1e89bd3e8e3","Type":"ContainerDied","Data":"46f7bbc83c743a6fd3b0aacd93c3f748f1bf03298be9d84d8d2cec7609810842"} Dec 11 08:30:38 crc kubenswrapper[4860]: I1211 08:30:38.762219 4860 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="46f7bbc83c743a6fd3b0aacd93c3f748f1bf03298be9d84d8d2cec7609810842" Dec 11 08:30:38 crc kubenswrapper[4860]: I1211 08:30:38.762985 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-b7a3-account-create-update-qgft5" Dec 11 08:30:38 crc kubenswrapper[4860]: I1211 08:30:38.943827 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 11 08:30:39 crc kubenswrapper[4860]: I1211 08:30:39.757774 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 11 08:30:39 crc kubenswrapper[4860]: I1211 08:30:39.758393 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="83082bff-37e1-4ac1-9074-ae9d69ed66b0" containerName="glance-log" containerID="cri-o://80eab5ca76a36476cefed22689137813096f4b92f167de2ab1d721595040852f" gracePeriod=30 Dec 11 08:30:39 crc kubenswrapper[4860]: I1211 08:30:39.758500 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="83082bff-37e1-4ac1-9074-ae9d69ed66b0" containerName="glance-httpd" containerID="cri-o://239173e5da0a533df7671655d48736e1bc38b5098483280871fd5ef02669df4c" gracePeriod=30 Dec 11 08:30:40 crc kubenswrapper[4860]: I1211 08:30:40.785271 4860 generic.go:334] "Generic (PLEG): container finished" podID="83082bff-37e1-4ac1-9074-ae9d69ed66b0" containerID="80eab5ca76a36476cefed22689137813096f4b92f167de2ab1d721595040852f" exitCode=143 Dec 11 08:30:40 crc kubenswrapper[4860]: I1211 08:30:40.785357 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"83082bff-37e1-4ac1-9074-ae9d69ed66b0","Type":"ContainerDied","Data":"80eab5ca76a36476cefed22689137813096f4b92f167de2ab1d721595040852f"} Dec 11 08:30:41 crc kubenswrapper[4860]: I1211 08:30:41.221997 4860 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-external-api-0" podUID="a3d65995-e403-4e92-b7aa-b39b8084d1bc" containerName="glance-httpd" probeResult="failure" output="Get \"https://10.217.0.152:9292/healthcheck\": read tcp 10.217.0.2:48406->10.217.0.152:9292: read: connection reset by peer" Dec 11 08:30:41 crc kubenswrapper[4860]: I1211 08:30:41.222402 4860 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-external-api-0" podUID="a3d65995-e403-4e92-b7aa-b39b8084d1bc" containerName="glance-log" probeResult="failure" output="Get \"https://10.217.0.152:9292/healthcheck\": read tcp 10.217.0.2:48422->10.217.0.152:9292: read: connection reset by peer" Dec 11 08:30:41 crc kubenswrapper[4860]: I1211 08:30:41.801110 4860 generic.go:334] "Generic (PLEG): container finished" podID="fb9e2dd0-591f-4a0e-88aa-5f21285fbc77" containerID="2333689d78c9ee84e0452445e3f70919ddfeeccadb977eb6bd41e9ac35f23fd0" exitCode=0 Dec 11 08:30:41 crc kubenswrapper[4860]: I1211 08:30:41.801179 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fb9e2dd0-591f-4a0e-88aa-5f21285fbc77","Type":"ContainerDied","Data":"2333689d78c9ee84e0452445e3f70919ddfeeccadb977eb6bd41e9ac35f23fd0"} Dec 11 08:30:41 crc kubenswrapper[4860]: I1211 08:30:41.805384 4860 generic.go:334] "Generic (PLEG): container finished" podID="a3d65995-e403-4e92-b7aa-b39b8084d1bc" containerID="b2623116284268f161ab7f24e48bff42f3f1d8de1c286d66200076c14429dc28" exitCode=0 Dec 11 08:30:41 crc kubenswrapper[4860]: I1211 08:30:41.805456 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"a3d65995-e403-4e92-b7aa-b39b8084d1bc","Type":"ContainerDied","Data":"b2623116284268f161ab7f24e48bff42f3f1d8de1c286d66200076c14429dc28"} Dec 11 08:30:41 crc kubenswrapper[4860]: I1211 08:30:41.805496 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"a3d65995-e403-4e92-b7aa-b39b8084d1bc","Type":"ContainerDied","Data":"099601ebce1a3974f9e908e2166615be7d1eb2fba71bc7455e8b36301a7886da"} Dec 11 08:30:41 crc kubenswrapper[4860]: I1211 08:30:41.805511 4860 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="099601ebce1a3974f9e908e2166615be7d1eb2fba71bc7455e8b36301a7886da" Dec 11 08:30:41 crc kubenswrapper[4860]: I1211 08:30:41.852054 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 11 08:30:42 crc kubenswrapper[4860]: I1211 08:30:42.010736 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a3d65995-e403-4e92-b7aa-b39b8084d1bc-public-tls-certs\") pod \"a3d65995-e403-4e92-b7aa-b39b8084d1bc\" (UID: \"a3d65995-e403-4e92-b7aa-b39b8084d1bc\") " Dec 11 08:30:42 crc kubenswrapper[4860]: I1211 08:30:42.010795 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a3d65995-e403-4e92-b7aa-b39b8084d1bc-logs\") pod \"a3d65995-e403-4e92-b7aa-b39b8084d1bc\" (UID: \"a3d65995-e403-4e92-b7aa-b39b8084d1bc\") " Dec 11 08:30:42 crc kubenswrapper[4860]: I1211 08:30:42.010957 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a3d65995-e403-4e92-b7aa-b39b8084d1bc-scripts\") pod \"a3d65995-e403-4e92-b7aa-b39b8084d1bc\" (UID: \"a3d65995-e403-4e92-b7aa-b39b8084d1bc\") " Dec 11 08:30:42 crc kubenswrapper[4860]: I1211 08:30:42.011043 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dngvk\" (UniqueName: \"kubernetes.io/projected/a3d65995-e403-4e92-b7aa-b39b8084d1bc-kube-api-access-dngvk\") pod \"a3d65995-e403-4e92-b7aa-b39b8084d1bc\" (UID: \"a3d65995-e403-4e92-b7aa-b39b8084d1bc\") " Dec 11 08:30:42 crc kubenswrapper[4860]: I1211 08:30:42.011101 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3d65995-e403-4e92-b7aa-b39b8084d1bc-config-data\") pod \"a3d65995-e403-4e92-b7aa-b39b8084d1bc\" (UID: \"a3d65995-e403-4e92-b7aa-b39b8084d1bc\") " Dec 11 08:30:42 crc kubenswrapper[4860]: I1211 08:30:42.011198 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3d65995-e403-4e92-b7aa-b39b8084d1bc-combined-ca-bundle\") pod \"a3d65995-e403-4e92-b7aa-b39b8084d1bc\" (UID: \"a3d65995-e403-4e92-b7aa-b39b8084d1bc\") " Dec 11 08:30:42 crc kubenswrapper[4860]: I1211 08:30:42.011242 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"a3d65995-e403-4e92-b7aa-b39b8084d1bc\" (UID: \"a3d65995-e403-4e92-b7aa-b39b8084d1bc\") " Dec 11 08:30:42 crc kubenswrapper[4860]: I1211 08:30:42.011298 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a3d65995-e403-4e92-b7aa-b39b8084d1bc-httpd-run\") pod \"a3d65995-e403-4e92-b7aa-b39b8084d1bc\" (UID: \"a3d65995-e403-4e92-b7aa-b39b8084d1bc\") " Dec 11 08:30:42 crc kubenswrapper[4860]: I1211 08:30:42.011331 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a3d65995-e403-4e92-b7aa-b39b8084d1bc-logs" (OuterVolumeSpecName: "logs") pod "a3d65995-e403-4e92-b7aa-b39b8084d1bc" (UID: "a3d65995-e403-4e92-b7aa-b39b8084d1bc"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:30:42 crc kubenswrapper[4860]: I1211 08:30:42.011975 4860 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a3d65995-e403-4e92-b7aa-b39b8084d1bc-logs\") on node \"crc\" DevicePath \"\"" Dec 11 08:30:42 crc kubenswrapper[4860]: I1211 08:30:42.012386 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a3d65995-e403-4e92-b7aa-b39b8084d1bc-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "a3d65995-e403-4e92-b7aa-b39b8084d1bc" (UID: "a3d65995-e403-4e92-b7aa-b39b8084d1bc"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:30:42 crc kubenswrapper[4860]: I1211 08:30:42.017314 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a3d65995-e403-4e92-b7aa-b39b8084d1bc-kube-api-access-dngvk" (OuterVolumeSpecName: "kube-api-access-dngvk") pod "a3d65995-e403-4e92-b7aa-b39b8084d1bc" (UID: "a3d65995-e403-4e92-b7aa-b39b8084d1bc"). InnerVolumeSpecName "kube-api-access-dngvk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:30:42 crc kubenswrapper[4860]: I1211 08:30:42.031841 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "glance") pod "a3d65995-e403-4e92-b7aa-b39b8084d1bc" (UID: "a3d65995-e403-4e92-b7aa-b39b8084d1bc"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 11 08:30:42 crc kubenswrapper[4860]: I1211 08:30:42.032784 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3d65995-e403-4e92-b7aa-b39b8084d1bc-scripts" (OuterVolumeSpecName: "scripts") pod "a3d65995-e403-4e92-b7aa-b39b8084d1bc" (UID: "a3d65995-e403-4e92-b7aa-b39b8084d1bc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:30:42 crc kubenswrapper[4860]: I1211 08:30:42.045343 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3d65995-e403-4e92-b7aa-b39b8084d1bc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a3d65995-e403-4e92-b7aa-b39b8084d1bc" (UID: "a3d65995-e403-4e92-b7aa-b39b8084d1bc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:30:42 crc kubenswrapper[4860]: I1211 08:30:42.082886 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3d65995-e403-4e92-b7aa-b39b8084d1bc-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "a3d65995-e403-4e92-b7aa-b39b8084d1bc" (UID: "a3d65995-e403-4e92-b7aa-b39b8084d1bc"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:30:42 crc kubenswrapper[4860]: I1211 08:30:42.091241 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3d65995-e403-4e92-b7aa-b39b8084d1bc-config-data" (OuterVolumeSpecName: "config-data") pod "a3d65995-e403-4e92-b7aa-b39b8084d1bc" (UID: "a3d65995-e403-4e92-b7aa-b39b8084d1bc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:30:42 crc kubenswrapper[4860]: I1211 08:30:42.113615 4860 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a3d65995-e403-4e92-b7aa-b39b8084d1bc-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 11 08:30:42 crc kubenswrapper[4860]: I1211 08:30:42.113677 4860 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a3d65995-e403-4e92-b7aa-b39b8084d1bc-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 08:30:42 crc kubenswrapper[4860]: I1211 08:30:42.113688 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dngvk\" (UniqueName: \"kubernetes.io/projected/a3d65995-e403-4e92-b7aa-b39b8084d1bc-kube-api-access-dngvk\") on node \"crc\" DevicePath \"\"" Dec 11 08:30:42 crc kubenswrapper[4860]: I1211 08:30:42.113702 4860 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3d65995-e403-4e92-b7aa-b39b8084d1bc-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 08:30:42 crc kubenswrapper[4860]: I1211 08:30:42.113710 4860 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3d65995-e403-4e92-b7aa-b39b8084d1bc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 08:30:42 crc kubenswrapper[4860]: I1211 08:30:42.113750 4860 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Dec 11 08:30:42 crc kubenswrapper[4860]: I1211 08:30:42.113760 4860 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a3d65995-e403-4e92-b7aa-b39b8084d1bc-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 11 08:30:42 crc kubenswrapper[4860]: I1211 08:30:42.145356 4860 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Dec 11 08:30:42 crc kubenswrapper[4860]: I1211 08:30:42.215552 4860 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Dec 11 08:30:42 crc kubenswrapper[4860]: I1211 08:30:42.817971 4860 generic.go:334] "Generic (PLEG): container finished" podID="8f38e033-5243-4e65-a7e0-29f1282b063f" containerID="0aab949e04f4b3a4703fdc0bff0fb82f6a451c3c24e3537820a25282e91ff2c8" exitCode=137 Dec 11 08:30:42 crc kubenswrapper[4860]: I1211 08:30:42.818073 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 11 08:30:42 crc kubenswrapper[4860]: I1211 08:30:42.818080 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-776755b658-9fwr9" event={"ID":"8f38e033-5243-4e65-a7e0-29f1282b063f","Type":"ContainerDied","Data":"0aab949e04f4b3a4703fdc0bff0fb82f6a451c3c24e3537820a25282e91ff2c8"} Dec 11 08:30:42 crc kubenswrapper[4860]: I1211 08:30:42.818150 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-776755b658-9fwr9" event={"ID":"8f38e033-5243-4e65-a7e0-29f1282b063f","Type":"ContainerDied","Data":"e978a4eb3ce30930d0707ffadd7da47a1d76fbcdb63711451ee10b5c7580b359"} Dec 11 08:30:42 crc kubenswrapper[4860]: I1211 08:30:42.818164 4860 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e978a4eb3ce30930d0707ffadd7da47a1d76fbcdb63711451ee10b5c7580b359" Dec 11 08:30:42 crc kubenswrapper[4860]: I1211 08:30:42.885432 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-776755b658-9fwr9" Dec 11 08:30:42 crc kubenswrapper[4860]: I1211 08:30:42.902500 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 11 08:30:42 crc kubenswrapper[4860]: I1211 08:30:42.914829 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 11 08:30:42 crc kubenswrapper[4860]: I1211 08:30:42.932923 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 11 08:30:42 crc kubenswrapper[4860]: E1211 08:30:42.933313 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e01fc44a-9c71-49bb-bc43-f4a20d44b41a" containerName="mariadb-database-create" Dec 11 08:30:42 crc kubenswrapper[4860]: I1211 08:30:42.933334 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="e01fc44a-9c71-49bb-bc43-f4a20d44b41a" containerName="mariadb-database-create" Dec 11 08:30:42 crc kubenswrapper[4860]: E1211 08:30:42.933363 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f38e033-5243-4e65-a7e0-29f1282b063f" containerName="horizon" Dec 11 08:30:42 crc kubenswrapper[4860]: I1211 08:30:42.933371 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f38e033-5243-4e65-a7e0-29f1282b063f" containerName="horizon" Dec 11 08:30:42 crc kubenswrapper[4860]: E1211 08:30:42.933390 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f38e033-5243-4e65-a7e0-29f1282b063f" containerName="horizon-log" Dec 11 08:30:42 crc kubenswrapper[4860]: I1211 08:30:42.933396 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f38e033-5243-4e65-a7e0-29f1282b063f" containerName="horizon-log" Dec 11 08:30:42 crc kubenswrapper[4860]: E1211 08:30:42.933406 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31d35fc5-e254-4e3f-86bf-9a944f8a44af" containerName="mariadb-database-create" Dec 11 08:30:42 crc kubenswrapper[4860]: I1211 08:30:42.933413 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="31d35fc5-e254-4e3f-86bf-9a944f8a44af" containerName="mariadb-database-create" Dec 11 08:30:42 crc kubenswrapper[4860]: E1211 08:30:42.933420 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fdae480c-a60d-4508-add1-e1e89bd3e8e3" containerName="mariadb-account-create-update" Dec 11 08:30:42 crc kubenswrapper[4860]: I1211 08:30:42.933427 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="fdae480c-a60d-4508-add1-e1e89bd3e8e3" containerName="mariadb-account-create-update" Dec 11 08:30:42 crc kubenswrapper[4860]: E1211 08:30:42.933439 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="293660ed-0401-45cc-811e-a612da96284f" containerName="mariadb-account-create-update" Dec 11 08:30:42 crc kubenswrapper[4860]: I1211 08:30:42.933446 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="293660ed-0401-45cc-811e-a612da96284f" containerName="mariadb-account-create-update" Dec 11 08:30:42 crc kubenswrapper[4860]: E1211 08:30:42.933467 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3d65995-e403-4e92-b7aa-b39b8084d1bc" containerName="glance-log" Dec 11 08:30:42 crc kubenswrapper[4860]: I1211 08:30:42.933473 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3d65995-e403-4e92-b7aa-b39b8084d1bc" containerName="glance-log" Dec 11 08:30:42 crc kubenswrapper[4860]: E1211 08:30:42.933484 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4598c5c-a584-487e-83ed-0126b5fd253e" containerName="mariadb-database-create" Dec 11 08:30:42 crc kubenswrapper[4860]: I1211 08:30:42.933490 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4598c5c-a584-487e-83ed-0126b5fd253e" containerName="mariadb-database-create" Dec 11 08:30:42 crc kubenswrapper[4860]: E1211 08:30:42.933505 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3d65995-e403-4e92-b7aa-b39b8084d1bc" containerName="glance-httpd" Dec 11 08:30:42 crc kubenswrapper[4860]: I1211 08:30:42.933511 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3d65995-e403-4e92-b7aa-b39b8084d1bc" containerName="glance-httpd" Dec 11 08:30:42 crc kubenswrapper[4860]: E1211 08:30:42.933519 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a7614eb-5a8b-44ca-afe2-22a8b1151065" containerName="mariadb-account-create-update" Dec 11 08:30:42 crc kubenswrapper[4860]: I1211 08:30:42.933524 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a7614eb-5a8b-44ca-afe2-22a8b1151065" containerName="mariadb-account-create-update" Dec 11 08:30:42 crc kubenswrapper[4860]: I1211 08:30:42.933742 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="b4598c5c-a584-487e-83ed-0126b5fd253e" containerName="mariadb-database-create" Dec 11 08:30:42 crc kubenswrapper[4860]: I1211 08:30:42.933763 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f38e033-5243-4e65-a7e0-29f1282b063f" containerName="horizon" Dec 11 08:30:42 crc kubenswrapper[4860]: I1211 08:30:42.933775 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="e01fc44a-9c71-49bb-bc43-f4a20d44b41a" containerName="mariadb-database-create" Dec 11 08:30:42 crc kubenswrapper[4860]: I1211 08:30:42.933784 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3d65995-e403-4e92-b7aa-b39b8084d1bc" containerName="glance-log" Dec 11 08:30:42 crc kubenswrapper[4860]: I1211 08:30:42.933793 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="31d35fc5-e254-4e3f-86bf-9a944f8a44af" containerName="mariadb-database-create" Dec 11 08:30:42 crc kubenswrapper[4860]: I1211 08:30:42.933803 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f38e033-5243-4e65-a7e0-29f1282b063f" containerName="horizon-log" Dec 11 08:30:42 crc kubenswrapper[4860]: I1211 08:30:42.933813 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="fdae480c-a60d-4508-add1-e1e89bd3e8e3" containerName="mariadb-account-create-update" Dec 11 08:30:42 crc kubenswrapper[4860]: I1211 08:30:42.933824 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="293660ed-0401-45cc-811e-a612da96284f" containerName="mariadb-account-create-update" Dec 11 08:30:42 crc kubenswrapper[4860]: I1211 08:30:42.933836 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3d65995-e403-4e92-b7aa-b39b8084d1bc" containerName="glance-httpd" Dec 11 08:30:42 crc kubenswrapper[4860]: I1211 08:30:42.933844 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a7614eb-5a8b-44ca-afe2-22a8b1151065" containerName="mariadb-account-create-update" Dec 11 08:30:42 crc kubenswrapper[4860]: I1211 08:30:42.934843 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 11 08:30:42 crc kubenswrapper[4860]: I1211 08:30:42.940282 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 11 08:30:42 crc kubenswrapper[4860]: I1211 08:30:42.940596 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 11 08:30:42 crc kubenswrapper[4860]: I1211 08:30:42.956318 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 11 08:30:43 crc kubenswrapper[4860]: I1211 08:30:43.029501 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/8f38e033-5243-4e65-a7e0-29f1282b063f-horizon-tls-certs\") pod \"8f38e033-5243-4e65-a7e0-29f1282b063f\" (UID: \"8f38e033-5243-4e65-a7e0-29f1282b063f\") " Dec 11 08:30:43 crc kubenswrapper[4860]: I1211 08:30:43.029613 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8f38e033-5243-4e65-a7e0-29f1282b063f-config-data\") pod \"8f38e033-5243-4e65-a7e0-29f1282b063f\" (UID: \"8f38e033-5243-4e65-a7e0-29f1282b063f\") " Dec 11 08:30:43 crc kubenswrapper[4860]: I1211 08:30:43.031864 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8f38e033-5243-4e65-a7e0-29f1282b063f-logs\") pod \"8f38e033-5243-4e65-a7e0-29f1282b063f\" (UID: \"8f38e033-5243-4e65-a7e0-29f1282b063f\") " Dec 11 08:30:43 crc kubenswrapper[4860]: I1211 08:30:43.031966 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/8f38e033-5243-4e65-a7e0-29f1282b063f-horizon-secret-key\") pod \"8f38e033-5243-4e65-a7e0-29f1282b063f\" (UID: \"8f38e033-5243-4e65-a7e0-29f1282b063f\") " Dec 11 08:30:43 crc kubenswrapper[4860]: I1211 08:30:43.032000 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f38e033-5243-4e65-a7e0-29f1282b063f-combined-ca-bundle\") pod \"8f38e033-5243-4e65-a7e0-29f1282b063f\" (UID: \"8f38e033-5243-4e65-a7e0-29f1282b063f\") " Dec 11 08:30:43 crc kubenswrapper[4860]: I1211 08:30:43.032086 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8f38e033-5243-4e65-a7e0-29f1282b063f-scripts\") pod \"8f38e033-5243-4e65-a7e0-29f1282b063f\" (UID: \"8f38e033-5243-4e65-a7e0-29f1282b063f\") " Dec 11 08:30:43 crc kubenswrapper[4860]: I1211 08:30:43.032130 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nx4wx\" (UniqueName: \"kubernetes.io/projected/8f38e033-5243-4e65-a7e0-29f1282b063f-kube-api-access-nx4wx\") pod \"8f38e033-5243-4e65-a7e0-29f1282b063f\" (UID: \"8f38e033-5243-4e65-a7e0-29f1282b063f\") " Dec 11 08:30:43 crc kubenswrapper[4860]: I1211 08:30:43.032503 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f51c05dd-1881-4644-9a40-7e6e2174c591-scripts\") pod \"glance-default-external-api-0\" (UID: \"f51c05dd-1881-4644-9a40-7e6e2174c591\") " pod="openstack/glance-default-external-api-0" Dec 11 08:30:43 crc kubenswrapper[4860]: I1211 08:30:43.032560 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f51c05dd-1881-4644-9a40-7e6e2174c591-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"f51c05dd-1881-4644-9a40-7e6e2174c591\") " pod="openstack/glance-default-external-api-0" Dec 11 08:30:43 crc kubenswrapper[4860]: I1211 08:30:43.032600 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"f51c05dd-1881-4644-9a40-7e6e2174c591\") " pod="openstack/glance-default-external-api-0" Dec 11 08:30:43 crc kubenswrapper[4860]: I1211 08:30:43.032774 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9vsvj\" (UniqueName: \"kubernetes.io/projected/f51c05dd-1881-4644-9a40-7e6e2174c591-kube-api-access-9vsvj\") pod \"glance-default-external-api-0\" (UID: \"f51c05dd-1881-4644-9a40-7e6e2174c591\") " pod="openstack/glance-default-external-api-0" Dec 11 08:30:43 crc kubenswrapper[4860]: I1211 08:30:43.032925 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f51c05dd-1881-4644-9a40-7e6e2174c591-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"f51c05dd-1881-4644-9a40-7e6e2174c591\") " pod="openstack/glance-default-external-api-0" Dec 11 08:30:43 crc kubenswrapper[4860]: I1211 08:30:43.033074 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f51c05dd-1881-4644-9a40-7e6e2174c591-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"f51c05dd-1881-4644-9a40-7e6e2174c591\") " pod="openstack/glance-default-external-api-0" Dec 11 08:30:43 crc kubenswrapper[4860]: I1211 08:30:43.033122 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f51c05dd-1881-4644-9a40-7e6e2174c591-config-data\") pod \"glance-default-external-api-0\" (UID: \"f51c05dd-1881-4644-9a40-7e6e2174c591\") " pod="openstack/glance-default-external-api-0" Dec 11 08:30:43 crc kubenswrapper[4860]: I1211 08:30:43.033219 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f51c05dd-1881-4644-9a40-7e6e2174c591-logs\") pod \"glance-default-external-api-0\" (UID: \"f51c05dd-1881-4644-9a40-7e6e2174c591\") " pod="openstack/glance-default-external-api-0" Dec 11 08:30:43 crc kubenswrapper[4860]: I1211 08:30:43.033967 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f38e033-5243-4e65-a7e0-29f1282b063f-logs" (OuterVolumeSpecName: "logs") pod "8f38e033-5243-4e65-a7e0-29f1282b063f" (UID: "8f38e033-5243-4e65-a7e0-29f1282b063f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:30:43 crc kubenswrapper[4860]: I1211 08:30:43.061818 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f38e033-5243-4e65-a7e0-29f1282b063f-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "8f38e033-5243-4e65-a7e0-29f1282b063f" (UID: "8f38e033-5243-4e65-a7e0-29f1282b063f"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:30:43 crc kubenswrapper[4860]: I1211 08:30:43.094868 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f38e033-5243-4e65-a7e0-29f1282b063f-kube-api-access-nx4wx" (OuterVolumeSpecName: "kube-api-access-nx4wx") pod "8f38e033-5243-4e65-a7e0-29f1282b063f" (UID: "8f38e033-5243-4e65-a7e0-29f1282b063f"). InnerVolumeSpecName "kube-api-access-nx4wx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:30:43 crc kubenswrapper[4860]: I1211 08:30:43.134653 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f38e033-5243-4e65-a7e0-29f1282b063f-scripts" (OuterVolumeSpecName: "scripts") pod "8f38e033-5243-4e65-a7e0-29f1282b063f" (UID: "8f38e033-5243-4e65-a7e0-29f1282b063f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:30:43 crc kubenswrapper[4860]: I1211 08:30:43.136876 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f51c05dd-1881-4644-9a40-7e6e2174c591-logs\") pod \"glance-default-external-api-0\" (UID: \"f51c05dd-1881-4644-9a40-7e6e2174c591\") " pod="openstack/glance-default-external-api-0" Dec 11 08:30:43 crc kubenswrapper[4860]: I1211 08:30:43.136959 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f51c05dd-1881-4644-9a40-7e6e2174c591-scripts\") pod \"glance-default-external-api-0\" (UID: \"f51c05dd-1881-4644-9a40-7e6e2174c591\") " pod="openstack/glance-default-external-api-0" Dec 11 08:30:43 crc kubenswrapper[4860]: I1211 08:30:43.136991 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f51c05dd-1881-4644-9a40-7e6e2174c591-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"f51c05dd-1881-4644-9a40-7e6e2174c591\") " pod="openstack/glance-default-external-api-0" Dec 11 08:30:43 crc kubenswrapper[4860]: I1211 08:30:43.137014 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"f51c05dd-1881-4644-9a40-7e6e2174c591\") " pod="openstack/glance-default-external-api-0" Dec 11 08:30:43 crc kubenswrapper[4860]: I1211 08:30:43.137066 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9vsvj\" (UniqueName: \"kubernetes.io/projected/f51c05dd-1881-4644-9a40-7e6e2174c591-kube-api-access-9vsvj\") pod \"glance-default-external-api-0\" (UID: \"f51c05dd-1881-4644-9a40-7e6e2174c591\") " pod="openstack/glance-default-external-api-0" Dec 11 08:30:43 crc kubenswrapper[4860]: I1211 08:30:43.137120 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f51c05dd-1881-4644-9a40-7e6e2174c591-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"f51c05dd-1881-4644-9a40-7e6e2174c591\") " pod="openstack/glance-default-external-api-0" Dec 11 08:30:43 crc kubenswrapper[4860]: I1211 08:30:43.137179 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f51c05dd-1881-4644-9a40-7e6e2174c591-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"f51c05dd-1881-4644-9a40-7e6e2174c591\") " pod="openstack/glance-default-external-api-0" Dec 11 08:30:43 crc kubenswrapper[4860]: I1211 08:30:43.137197 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f51c05dd-1881-4644-9a40-7e6e2174c591-config-data\") pod \"glance-default-external-api-0\" (UID: \"f51c05dd-1881-4644-9a40-7e6e2174c591\") " pod="openstack/glance-default-external-api-0" Dec 11 08:30:43 crc kubenswrapper[4860]: I1211 08:30:43.137257 4860 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/8f38e033-5243-4e65-a7e0-29f1282b063f-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 11 08:30:43 crc kubenswrapper[4860]: I1211 08:30:43.137268 4860 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8f38e033-5243-4e65-a7e0-29f1282b063f-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 08:30:43 crc kubenswrapper[4860]: I1211 08:30:43.137277 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nx4wx\" (UniqueName: \"kubernetes.io/projected/8f38e033-5243-4e65-a7e0-29f1282b063f-kube-api-access-nx4wx\") on node \"crc\" DevicePath \"\"" Dec 11 08:30:43 crc kubenswrapper[4860]: I1211 08:30:43.137291 4860 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8f38e033-5243-4e65-a7e0-29f1282b063f-logs\") on node \"crc\" DevicePath \"\"" Dec 11 08:30:43 crc kubenswrapper[4860]: I1211 08:30:43.139025 4860 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"f51c05dd-1881-4644-9a40-7e6e2174c591\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-external-api-0" Dec 11 08:30:43 crc kubenswrapper[4860]: I1211 08:30:43.147415 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f38e033-5243-4e65-a7e0-29f1282b063f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8f38e033-5243-4e65-a7e0-29f1282b063f" (UID: "8f38e033-5243-4e65-a7e0-29f1282b063f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:30:43 crc kubenswrapper[4860]: I1211 08:30:43.147555 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f51c05dd-1881-4644-9a40-7e6e2174c591-scripts\") pod \"glance-default-external-api-0\" (UID: \"f51c05dd-1881-4644-9a40-7e6e2174c591\") " pod="openstack/glance-default-external-api-0" Dec 11 08:30:43 crc kubenswrapper[4860]: I1211 08:30:43.147881 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f51c05dd-1881-4644-9a40-7e6e2174c591-logs\") pod \"glance-default-external-api-0\" (UID: \"f51c05dd-1881-4644-9a40-7e6e2174c591\") " pod="openstack/glance-default-external-api-0" Dec 11 08:30:43 crc kubenswrapper[4860]: I1211 08:30:43.147939 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f51c05dd-1881-4644-9a40-7e6e2174c591-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"f51c05dd-1881-4644-9a40-7e6e2174c591\") " pod="openstack/glance-default-external-api-0" Dec 11 08:30:43 crc kubenswrapper[4860]: I1211 08:30:43.148535 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f38e033-5243-4e65-a7e0-29f1282b063f-config-data" (OuterVolumeSpecName: "config-data") pod "8f38e033-5243-4e65-a7e0-29f1282b063f" (UID: "8f38e033-5243-4e65-a7e0-29f1282b063f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:30:43 crc kubenswrapper[4860]: I1211 08:30:43.169586 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f51c05dd-1881-4644-9a40-7e6e2174c591-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"f51c05dd-1881-4644-9a40-7e6e2174c591\") " pod="openstack/glance-default-external-api-0" Dec 11 08:30:43 crc kubenswrapper[4860]: I1211 08:30:43.186457 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9vsvj\" (UniqueName: \"kubernetes.io/projected/f51c05dd-1881-4644-9a40-7e6e2174c591-kube-api-access-9vsvj\") pod \"glance-default-external-api-0\" (UID: \"f51c05dd-1881-4644-9a40-7e6e2174c591\") " pod="openstack/glance-default-external-api-0" Dec 11 08:30:43 crc kubenswrapper[4860]: I1211 08:30:43.188762 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f51c05dd-1881-4644-9a40-7e6e2174c591-config-data\") pod \"glance-default-external-api-0\" (UID: \"f51c05dd-1881-4644-9a40-7e6e2174c591\") " pod="openstack/glance-default-external-api-0" Dec 11 08:30:43 crc kubenswrapper[4860]: I1211 08:30:43.196551 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f51c05dd-1881-4644-9a40-7e6e2174c591-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"f51c05dd-1881-4644-9a40-7e6e2174c591\") " pod="openstack/glance-default-external-api-0" Dec 11 08:30:43 crc kubenswrapper[4860]: I1211 08:30:43.240314 4860 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8f38e033-5243-4e65-a7e0-29f1282b063f-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 08:30:43 crc kubenswrapper[4860]: I1211 08:30:43.240661 4860 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f38e033-5243-4e65-a7e0-29f1282b063f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 08:30:43 crc kubenswrapper[4860]: I1211 08:30:43.261780 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f38e033-5243-4e65-a7e0-29f1282b063f-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "8f38e033-5243-4e65-a7e0-29f1282b063f" (UID: "8f38e033-5243-4e65-a7e0-29f1282b063f"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:30:43 crc kubenswrapper[4860]: I1211 08:30:43.295212 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"f51c05dd-1881-4644-9a40-7e6e2174c591\") " pod="openstack/glance-default-external-api-0" Dec 11 08:30:43 crc kubenswrapper[4860]: I1211 08:30:43.342156 4860 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/8f38e033-5243-4e65-a7e0-29f1282b063f-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 11 08:30:43 crc kubenswrapper[4860]: I1211 08:30:43.556416 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 11 08:30:43 crc kubenswrapper[4860]: I1211 08:30:43.604737 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a3d65995-e403-4e92-b7aa-b39b8084d1bc" path="/var/lib/kubelet/pods/a3d65995-e403-4e92-b7aa-b39b8084d1bc/volumes" Dec 11 08:30:43 crc kubenswrapper[4860]: I1211 08:30:43.710367 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 11 08:30:43 crc kubenswrapper[4860]: I1211 08:30:43.832928 4860 generic.go:334] "Generic (PLEG): container finished" podID="83082bff-37e1-4ac1-9074-ae9d69ed66b0" containerID="239173e5da0a533df7671655d48736e1bc38b5098483280871fd5ef02669df4c" exitCode=0 Dec 11 08:30:43 crc kubenswrapper[4860]: I1211 08:30:43.833032 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-776755b658-9fwr9" Dec 11 08:30:43 crc kubenswrapper[4860]: I1211 08:30:43.834070 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 11 08:30:43 crc kubenswrapper[4860]: I1211 08:30:43.834252 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"83082bff-37e1-4ac1-9074-ae9d69ed66b0","Type":"ContainerDied","Data":"239173e5da0a533df7671655d48736e1bc38b5098483280871fd5ef02669df4c"} Dec 11 08:30:43 crc kubenswrapper[4860]: I1211 08:30:43.834283 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"83082bff-37e1-4ac1-9074-ae9d69ed66b0","Type":"ContainerDied","Data":"df498fa320b29fcbeb2fb09668b03f3a7564a2703477e78ff4f432d27746d0ba"} Dec 11 08:30:43 crc kubenswrapper[4860]: I1211 08:30:43.834303 4860 scope.go:117] "RemoveContainer" containerID="239173e5da0a533df7671655d48736e1bc38b5098483280871fd5ef02669df4c" Dec 11 08:30:43 crc kubenswrapper[4860]: I1211 08:30:43.853350 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/83082bff-37e1-4ac1-9074-ae9d69ed66b0-scripts\") pod \"83082bff-37e1-4ac1-9074-ae9d69ed66b0\" (UID: \"83082bff-37e1-4ac1-9074-ae9d69ed66b0\") " Dec 11 08:30:43 crc kubenswrapper[4860]: I1211 08:30:43.853410 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"83082bff-37e1-4ac1-9074-ae9d69ed66b0\" (UID: \"83082bff-37e1-4ac1-9074-ae9d69ed66b0\") " Dec 11 08:30:43 crc kubenswrapper[4860]: I1211 08:30:43.853441 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/83082bff-37e1-4ac1-9074-ae9d69ed66b0-httpd-run\") pod \"83082bff-37e1-4ac1-9074-ae9d69ed66b0\" (UID: \"83082bff-37e1-4ac1-9074-ae9d69ed66b0\") " Dec 11 08:30:43 crc kubenswrapper[4860]: I1211 08:30:43.853503 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/83082bff-37e1-4ac1-9074-ae9d69ed66b0-logs\") pod \"83082bff-37e1-4ac1-9074-ae9d69ed66b0\" (UID: \"83082bff-37e1-4ac1-9074-ae9d69ed66b0\") " Dec 11 08:30:43 crc kubenswrapper[4860]: I1211 08:30:43.853597 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83082bff-37e1-4ac1-9074-ae9d69ed66b0-config-data\") pod \"83082bff-37e1-4ac1-9074-ae9d69ed66b0\" (UID: \"83082bff-37e1-4ac1-9074-ae9d69ed66b0\") " Dec 11 08:30:43 crc kubenswrapper[4860]: I1211 08:30:43.853619 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83082bff-37e1-4ac1-9074-ae9d69ed66b0-combined-ca-bundle\") pod \"83082bff-37e1-4ac1-9074-ae9d69ed66b0\" (UID: \"83082bff-37e1-4ac1-9074-ae9d69ed66b0\") " Dec 11 08:30:43 crc kubenswrapper[4860]: I1211 08:30:43.853768 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cwwwd\" (UniqueName: \"kubernetes.io/projected/83082bff-37e1-4ac1-9074-ae9d69ed66b0-kube-api-access-cwwwd\") pod \"83082bff-37e1-4ac1-9074-ae9d69ed66b0\" (UID: \"83082bff-37e1-4ac1-9074-ae9d69ed66b0\") " Dec 11 08:30:43 crc kubenswrapper[4860]: I1211 08:30:43.853787 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/83082bff-37e1-4ac1-9074-ae9d69ed66b0-internal-tls-certs\") pod \"83082bff-37e1-4ac1-9074-ae9d69ed66b0\" (UID: \"83082bff-37e1-4ac1-9074-ae9d69ed66b0\") " Dec 11 08:30:43 crc kubenswrapper[4860]: I1211 08:30:43.858008 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/83082bff-37e1-4ac1-9074-ae9d69ed66b0-logs" (OuterVolumeSpecName: "logs") pod "83082bff-37e1-4ac1-9074-ae9d69ed66b0" (UID: "83082bff-37e1-4ac1-9074-ae9d69ed66b0"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:30:43 crc kubenswrapper[4860]: I1211 08:30:43.858348 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/83082bff-37e1-4ac1-9074-ae9d69ed66b0-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "83082bff-37e1-4ac1-9074-ae9d69ed66b0" (UID: "83082bff-37e1-4ac1-9074-ae9d69ed66b0"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:30:43 crc kubenswrapper[4860]: I1211 08:30:43.913591 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-776755b658-9fwr9"] Dec 11 08:30:43 crc kubenswrapper[4860]: I1211 08:30:43.920851 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83082bff-37e1-4ac1-9074-ae9d69ed66b0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "83082bff-37e1-4ac1-9074-ae9d69ed66b0" (UID: "83082bff-37e1-4ac1-9074-ae9d69ed66b0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:30:43 crc kubenswrapper[4860]: I1211 08:30:43.937999 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "glance") pod "83082bff-37e1-4ac1-9074-ae9d69ed66b0" (UID: "83082bff-37e1-4ac1-9074-ae9d69ed66b0"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 11 08:30:43 crc kubenswrapper[4860]: I1211 08:30:43.941389 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-776755b658-9fwr9"] Dec 11 08:30:43 crc kubenswrapper[4860]: I1211 08:30:43.943652 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83082bff-37e1-4ac1-9074-ae9d69ed66b0-scripts" (OuterVolumeSpecName: "scripts") pod "83082bff-37e1-4ac1-9074-ae9d69ed66b0" (UID: "83082bff-37e1-4ac1-9074-ae9d69ed66b0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:30:43 crc kubenswrapper[4860]: I1211 08:30:43.945348 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83082bff-37e1-4ac1-9074-ae9d69ed66b0-kube-api-access-cwwwd" (OuterVolumeSpecName: "kube-api-access-cwwwd") pod "83082bff-37e1-4ac1-9074-ae9d69ed66b0" (UID: "83082bff-37e1-4ac1-9074-ae9d69ed66b0"). InnerVolumeSpecName "kube-api-access-cwwwd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:30:43 crc kubenswrapper[4860]: I1211 08:30:43.959779 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cwwwd\" (UniqueName: \"kubernetes.io/projected/83082bff-37e1-4ac1-9074-ae9d69ed66b0-kube-api-access-cwwwd\") on node \"crc\" DevicePath \"\"" Dec 11 08:30:43 crc kubenswrapper[4860]: I1211 08:30:43.959824 4860 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/83082bff-37e1-4ac1-9074-ae9d69ed66b0-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 08:30:43 crc kubenswrapper[4860]: I1211 08:30:43.959864 4860 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Dec 11 08:30:43 crc kubenswrapper[4860]: I1211 08:30:43.959875 4860 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/83082bff-37e1-4ac1-9074-ae9d69ed66b0-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 11 08:30:43 crc kubenswrapper[4860]: I1211 08:30:43.959890 4860 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/83082bff-37e1-4ac1-9074-ae9d69ed66b0-logs\") on node \"crc\" DevicePath \"\"" Dec 11 08:30:43 crc kubenswrapper[4860]: I1211 08:30:43.959899 4860 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83082bff-37e1-4ac1-9074-ae9d69ed66b0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 08:30:43 crc kubenswrapper[4860]: I1211 08:30:43.969840 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83082bff-37e1-4ac1-9074-ae9d69ed66b0-config-data" (OuterVolumeSpecName: "config-data") pod "83082bff-37e1-4ac1-9074-ae9d69ed66b0" (UID: "83082bff-37e1-4ac1-9074-ae9d69ed66b0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:30:43 crc kubenswrapper[4860]: I1211 08:30:43.982078 4860 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Dec 11 08:30:43 crc kubenswrapper[4860]: I1211 08:30:43.990159 4860 scope.go:117] "RemoveContainer" containerID="80eab5ca76a36476cefed22689137813096f4b92f167de2ab1d721595040852f" Dec 11 08:30:43 crc kubenswrapper[4860]: I1211 08:30:43.996239 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83082bff-37e1-4ac1-9074-ae9d69ed66b0-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "83082bff-37e1-4ac1-9074-ae9d69ed66b0" (UID: "83082bff-37e1-4ac1-9074-ae9d69ed66b0"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:30:44 crc kubenswrapper[4860]: I1211 08:30:44.047866 4860 scope.go:117] "RemoveContainer" containerID="239173e5da0a533df7671655d48736e1bc38b5098483280871fd5ef02669df4c" Dec 11 08:30:44 crc kubenswrapper[4860]: E1211 08:30:44.050544 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"239173e5da0a533df7671655d48736e1bc38b5098483280871fd5ef02669df4c\": container with ID starting with 239173e5da0a533df7671655d48736e1bc38b5098483280871fd5ef02669df4c not found: ID does not exist" containerID="239173e5da0a533df7671655d48736e1bc38b5098483280871fd5ef02669df4c" Dec 11 08:30:44 crc kubenswrapper[4860]: I1211 08:30:44.050584 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"239173e5da0a533df7671655d48736e1bc38b5098483280871fd5ef02669df4c"} err="failed to get container status \"239173e5da0a533df7671655d48736e1bc38b5098483280871fd5ef02669df4c\": rpc error: code = NotFound desc = could not find container \"239173e5da0a533df7671655d48736e1bc38b5098483280871fd5ef02669df4c\": container with ID starting with 239173e5da0a533df7671655d48736e1bc38b5098483280871fd5ef02669df4c not found: ID does not exist" Dec 11 08:30:44 crc kubenswrapper[4860]: I1211 08:30:44.050612 4860 scope.go:117] "RemoveContainer" containerID="80eab5ca76a36476cefed22689137813096f4b92f167de2ab1d721595040852f" Dec 11 08:30:44 crc kubenswrapper[4860]: E1211 08:30:44.054499 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"80eab5ca76a36476cefed22689137813096f4b92f167de2ab1d721595040852f\": container with ID starting with 80eab5ca76a36476cefed22689137813096f4b92f167de2ab1d721595040852f not found: ID does not exist" containerID="80eab5ca76a36476cefed22689137813096f4b92f167de2ab1d721595040852f" Dec 11 08:30:44 crc kubenswrapper[4860]: I1211 08:30:44.054712 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80eab5ca76a36476cefed22689137813096f4b92f167de2ab1d721595040852f"} err="failed to get container status \"80eab5ca76a36476cefed22689137813096f4b92f167de2ab1d721595040852f\": rpc error: code = NotFound desc = could not find container \"80eab5ca76a36476cefed22689137813096f4b92f167de2ab1d721595040852f\": container with ID starting with 80eab5ca76a36476cefed22689137813096f4b92f167de2ab1d721595040852f not found: ID does not exist" Dec 11 08:30:44 crc kubenswrapper[4860]: I1211 08:30:44.058189 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-kb24v"] Dec 11 08:30:44 crc kubenswrapper[4860]: E1211 08:30:44.058631 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83082bff-37e1-4ac1-9074-ae9d69ed66b0" containerName="glance-httpd" Dec 11 08:30:44 crc kubenswrapper[4860]: I1211 08:30:44.058659 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="83082bff-37e1-4ac1-9074-ae9d69ed66b0" containerName="glance-httpd" Dec 11 08:30:44 crc kubenswrapper[4860]: E1211 08:30:44.058673 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83082bff-37e1-4ac1-9074-ae9d69ed66b0" containerName="glance-log" Dec 11 08:30:44 crc kubenswrapper[4860]: I1211 08:30:44.058679 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="83082bff-37e1-4ac1-9074-ae9d69ed66b0" containerName="glance-log" Dec 11 08:30:44 crc kubenswrapper[4860]: I1211 08:30:44.058878 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="83082bff-37e1-4ac1-9074-ae9d69ed66b0" containerName="glance-log" Dec 11 08:30:44 crc kubenswrapper[4860]: I1211 08:30:44.058897 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="83082bff-37e1-4ac1-9074-ae9d69ed66b0" containerName="glance-httpd" Dec 11 08:30:44 crc kubenswrapper[4860]: I1211 08:30:44.060493 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-kb24v" Dec 11 08:30:44 crc kubenswrapper[4860]: I1211 08:30:44.067172 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Dec 11 08:30:44 crc kubenswrapper[4860]: I1211 08:30:44.067348 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 11 08:30:44 crc kubenswrapper[4860]: I1211 08:30:44.067389 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-w88hh" Dec 11 08:30:44 crc kubenswrapper[4860]: I1211 08:30:44.069428 4860 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83082bff-37e1-4ac1-9074-ae9d69ed66b0-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 08:30:44 crc kubenswrapper[4860]: I1211 08:30:44.069454 4860 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/83082bff-37e1-4ac1-9074-ae9d69ed66b0-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 11 08:30:44 crc kubenswrapper[4860]: I1211 08:30:44.069466 4860 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Dec 11 08:30:44 crc kubenswrapper[4860]: I1211 08:30:44.071331 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-kb24v"] Dec 11 08:30:44 crc kubenswrapper[4860]: I1211 08:30:44.170767 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 11 08:30:44 crc kubenswrapper[4860]: I1211 08:30:44.171176 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff7bf9ff-e357-4af7-b708-2eb6298b85a8-config-data\") pod \"nova-cell0-conductor-db-sync-kb24v\" (UID: \"ff7bf9ff-e357-4af7-b708-2eb6298b85a8\") " pod="openstack/nova-cell0-conductor-db-sync-kb24v" Dec 11 08:30:44 crc kubenswrapper[4860]: I1211 08:30:44.171260 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-249h9\" (UniqueName: \"kubernetes.io/projected/ff7bf9ff-e357-4af7-b708-2eb6298b85a8-kube-api-access-249h9\") pod \"nova-cell0-conductor-db-sync-kb24v\" (UID: \"ff7bf9ff-e357-4af7-b708-2eb6298b85a8\") " pod="openstack/nova-cell0-conductor-db-sync-kb24v" Dec 11 08:30:44 crc kubenswrapper[4860]: I1211 08:30:44.171331 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ff7bf9ff-e357-4af7-b708-2eb6298b85a8-scripts\") pod \"nova-cell0-conductor-db-sync-kb24v\" (UID: \"ff7bf9ff-e357-4af7-b708-2eb6298b85a8\") " pod="openstack/nova-cell0-conductor-db-sync-kb24v" Dec 11 08:30:44 crc kubenswrapper[4860]: I1211 08:30:44.171489 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff7bf9ff-e357-4af7-b708-2eb6298b85a8-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-kb24v\" (UID: \"ff7bf9ff-e357-4af7-b708-2eb6298b85a8\") " pod="openstack/nova-cell0-conductor-db-sync-kb24v" Dec 11 08:30:44 crc kubenswrapper[4860]: I1211 08:30:44.187656 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 11 08:30:44 crc kubenswrapper[4860]: I1211 08:30:44.197044 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 11 08:30:44 crc kubenswrapper[4860]: I1211 08:30:44.212404 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 11 08:30:44 crc kubenswrapper[4860]: I1211 08:30:44.214469 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 11 08:30:44 crc kubenswrapper[4860]: I1211 08:30:44.216589 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 11 08:30:44 crc kubenswrapper[4860]: I1211 08:30:44.216844 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 11 08:30:44 crc kubenswrapper[4860]: I1211 08:30:44.240137 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 11 08:30:44 crc kubenswrapper[4860]: I1211 08:30:44.273903 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gv6sg\" (UniqueName: \"kubernetes.io/projected/b7bf9af0-6051-4d90-8918-9fd00cb869cd-kube-api-access-gv6sg\") pod \"glance-default-internal-api-0\" (UID: \"b7bf9af0-6051-4d90-8918-9fd00cb869cd\") " pod="openstack/glance-default-internal-api-0" Dec 11 08:30:44 crc kubenswrapper[4860]: I1211 08:30:44.273963 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-249h9\" (UniqueName: \"kubernetes.io/projected/ff7bf9ff-e357-4af7-b708-2eb6298b85a8-kube-api-access-249h9\") pod \"nova-cell0-conductor-db-sync-kb24v\" (UID: \"ff7bf9ff-e357-4af7-b708-2eb6298b85a8\") " pod="openstack/nova-cell0-conductor-db-sync-kb24v" Dec 11 08:30:44 crc kubenswrapper[4860]: I1211 08:30:44.274012 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b7bf9af0-6051-4d90-8918-9fd00cb869cd-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"b7bf9af0-6051-4d90-8918-9fd00cb869cd\") " pod="openstack/glance-default-internal-api-0" Dec 11 08:30:44 crc kubenswrapper[4860]: I1211 08:30:44.274044 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b7bf9af0-6051-4d90-8918-9fd00cb869cd-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"b7bf9af0-6051-4d90-8918-9fd00cb869cd\") " pod="openstack/glance-default-internal-api-0" Dec 11 08:30:44 crc kubenswrapper[4860]: I1211 08:30:44.274082 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ff7bf9ff-e357-4af7-b708-2eb6298b85a8-scripts\") pod \"nova-cell0-conductor-db-sync-kb24v\" (UID: \"ff7bf9ff-e357-4af7-b708-2eb6298b85a8\") " pod="openstack/nova-cell0-conductor-db-sync-kb24v" Dec 11 08:30:44 crc kubenswrapper[4860]: I1211 08:30:44.274113 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"b7bf9af0-6051-4d90-8918-9fd00cb869cd\") " pod="openstack/glance-default-internal-api-0" Dec 11 08:30:44 crc kubenswrapper[4860]: I1211 08:30:44.274151 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b7bf9af0-6051-4d90-8918-9fd00cb869cd-scripts\") pod \"glance-default-internal-api-0\" (UID: \"b7bf9af0-6051-4d90-8918-9fd00cb869cd\") " pod="openstack/glance-default-internal-api-0" Dec 11 08:30:44 crc kubenswrapper[4860]: I1211 08:30:44.274173 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7bf9af0-6051-4d90-8918-9fd00cb869cd-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"b7bf9af0-6051-4d90-8918-9fd00cb869cd\") " pod="openstack/glance-default-internal-api-0" Dec 11 08:30:44 crc kubenswrapper[4860]: I1211 08:30:44.274303 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b7bf9af0-6051-4d90-8918-9fd00cb869cd-logs\") pod \"glance-default-internal-api-0\" (UID: \"b7bf9af0-6051-4d90-8918-9fd00cb869cd\") " pod="openstack/glance-default-internal-api-0" Dec 11 08:30:44 crc kubenswrapper[4860]: I1211 08:30:44.274331 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b7bf9af0-6051-4d90-8918-9fd00cb869cd-config-data\") pod \"glance-default-internal-api-0\" (UID: \"b7bf9af0-6051-4d90-8918-9fd00cb869cd\") " pod="openstack/glance-default-internal-api-0" Dec 11 08:30:44 crc kubenswrapper[4860]: I1211 08:30:44.274355 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff7bf9ff-e357-4af7-b708-2eb6298b85a8-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-kb24v\" (UID: \"ff7bf9ff-e357-4af7-b708-2eb6298b85a8\") " pod="openstack/nova-cell0-conductor-db-sync-kb24v" Dec 11 08:30:44 crc kubenswrapper[4860]: I1211 08:30:44.274399 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff7bf9ff-e357-4af7-b708-2eb6298b85a8-config-data\") pod \"nova-cell0-conductor-db-sync-kb24v\" (UID: \"ff7bf9ff-e357-4af7-b708-2eb6298b85a8\") " pod="openstack/nova-cell0-conductor-db-sync-kb24v" Dec 11 08:30:44 crc kubenswrapper[4860]: I1211 08:30:44.279031 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff7bf9ff-e357-4af7-b708-2eb6298b85a8-config-data\") pod \"nova-cell0-conductor-db-sync-kb24v\" (UID: \"ff7bf9ff-e357-4af7-b708-2eb6298b85a8\") " pod="openstack/nova-cell0-conductor-db-sync-kb24v" Dec 11 08:30:44 crc kubenswrapper[4860]: I1211 08:30:44.279894 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ff7bf9ff-e357-4af7-b708-2eb6298b85a8-scripts\") pod \"nova-cell0-conductor-db-sync-kb24v\" (UID: \"ff7bf9ff-e357-4af7-b708-2eb6298b85a8\") " pod="openstack/nova-cell0-conductor-db-sync-kb24v" Dec 11 08:30:44 crc kubenswrapper[4860]: I1211 08:30:44.280008 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff7bf9ff-e357-4af7-b708-2eb6298b85a8-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-kb24v\" (UID: \"ff7bf9ff-e357-4af7-b708-2eb6298b85a8\") " pod="openstack/nova-cell0-conductor-db-sync-kb24v" Dec 11 08:30:44 crc kubenswrapper[4860]: I1211 08:30:44.304225 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-249h9\" (UniqueName: \"kubernetes.io/projected/ff7bf9ff-e357-4af7-b708-2eb6298b85a8-kube-api-access-249h9\") pod \"nova-cell0-conductor-db-sync-kb24v\" (UID: \"ff7bf9ff-e357-4af7-b708-2eb6298b85a8\") " pod="openstack/nova-cell0-conductor-db-sync-kb24v" Dec 11 08:30:44 crc kubenswrapper[4860]: I1211 08:30:44.375749 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gv6sg\" (UniqueName: \"kubernetes.io/projected/b7bf9af0-6051-4d90-8918-9fd00cb869cd-kube-api-access-gv6sg\") pod \"glance-default-internal-api-0\" (UID: \"b7bf9af0-6051-4d90-8918-9fd00cb869cd\") " pod="openstack/glance-default-internal-api-0" Dec 11 08:30:44 crc kubenswrapper[4860]: I1211 08:30:44.375816 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b7bf9af0-6051-4d90-8918-9fd00cb869cd-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"b7bf9af0-6051-4d90-8918-9fd00cb869cd\") " pod="openstack/glance-default-internal-api-0" Dec 11 08:30:44 crc kubenswrapper[4860]: I1211 08:30:44.375841 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b7bf9af0-6051-4d90-8918-9fd00cb869cd-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"b7bf9af0-6051-4d90-8918-9fd00cb869cd\") " pod="openstack/glance-default-internal-api-0" Dec 11 08:30:44 crc kubenswrapper[4860]: I1211 08:30:44.375870 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"b7bf9af0-6051-4d90-8918-9fd00cb869cd\") " pod="openstack/glance-default-internal-api-0" Dec 11 08:30:44 crc kubenswrapper[4860]: I1211 08:30:44.375894 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b7bf9af0-6051-4d90-8918-9fd00cb869cd-scripts\") pod \"glance-default-internal-api-0\" (UID: \"b7bf9af0-6051-4d90-8918-9fd00cb869cd\") " pod="openstack/glance-default-internal-api-0" Dec 11 08:30:44 crc kubenswrapper[4860]: I1211 08:30:44.375909 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7bf9af0-6051-4d90-8918-9fd00cb869cd-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"b7bf9af0-6051-4d90-8918-9fd00cb869cd\") " pod="openstack/glance-default-internal-api-0" Dec 11 08:30:44 crc kubenswrapper[4860]: I1211 08:30:44.375992 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b7bf9af0-6051-4d90-8918-9fd00cb869cd-logs\") pod \"glance-default-internal-api-0\" (UID: \"b7bf9af0-6051-4d90-8918-9fd00cb869cd\") " pod="openstack/glance-default-internal-api-0" Dec 11 08:30:44 crc kubenswrapper[4860]: I1211 08:30:44.376009 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b7bf9af0-6051-4d90-8918-9fd00cb869cd-config-data\") pod \"glance-default-internal-api-0\" (UID: \"b7bf9af0-6051-4d90-8918-9fd00cb869cd\") " pod="openstack/glance-default-internal-api-0" Dec 11 08:30:44 crc kubenswrapper[4860]: I1211 08:30:44.377802 4860 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"b7bf9af0-6051-4d90-8918-9fd00cb869cd\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/glance-default-internal-api-0" Dec 11 08:30:44 crc kubenswrapper[4860]: I1211 08:30:44.378424 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b7bf9af0-6051-4d90-8918-9fd00cb869cd-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"b7bf9af0-6051-4d90-8918-9fd00cb869cd\") " pod="openstack/glance-default-internal-api-0" Dec 11 08:30:44 crc kubenswrapper[4860]: I1211 08:30:44.380670 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b7bf9af0-6051-4d90-8918-9fd00cb869cd-config-data\") pod \"glance-default-internal-api-0\" (UID: \"b7bf9af0-6051-4d90-8918-9fd00cb869cd\") " pod="openstack/glance-default-internal-api-0" Dec 11 08:30:44 crc kubenswrapper[4860]: I1211 08:30:44.380921 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b7bf9af0-6051-4d90-8918-9fd00cb869cd-logs\") pod \"glance-default-internal-api-0\" (UID: \"b7bf9af0-6051-4d90-8918-9fd00cb869cd\") " pod="openstack/glance-default-internal-api-0" Dec 11 08:30:44 crc kubenswrapper[4860]: I1211 08:30:44.387580 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7bf9af0-6051-4d90-8918-9fd00cb869cd-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"b7bf9af0-6051-4d90-8918-9fd00cb869cd\") " pod="openstack/glance-default-internal-api-0" Dec 11 08:30:44 crc kubenswrapper[4860]: I1211 08:30:44.392443 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b7bf9af0-6051-4d90-8918-9fd00cb869cd-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"b7bf9af0-6051-4d90-8918-9fd00cb869cd\") " pod="openstack/glance-default-internal-api-0" Dec 11 08:30:44 crc kubenswrapper[4860]: I1211 08:30:44.395408 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-kb24v" Dec 11 08:30:44 crc kubenswrapper[4860]: I1211 08:30:44.396248 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b7bf9af0-6051-4d90-8918-9fd00cb869cd-scripts\") pod \"glance-default-internal-api-0\" (UID: \"b7bf9af0-6051-4d90-8918-9fd00cb869cd\") " pod="openstack/glance-default-internal-api-0" Dec 11 08:30:44 crc kubenswrapper[4860]: I1211 08:30:44.429276 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gv6sg\" (UniqueName: \"kubernetes.io/projected/b7bf9af0-6051-4d90-8918-9fd00cb869cd-kube-api-access-gv6sg\") pod \"glance-default-internal-api-0\" (UID: \"b7bf9af0-6051-4d90-8918-9fd00cb869cd\") " pod="openstack/glance-default-internal-api-0" Dec 11 08:30:44 crc kubenswrapper[4860]: I1211 08:30:44.492463 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"b7bf9af0-6051-4d90-8918-9fd00cb869cd\") " pod="openstack/glance-default-internal-api-0" Dec 11 08:30:44 crc kubenswrapper[4860]: I1211 08:30:44.552988 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 11 08:30:44 crc kubenswrapper[4860]: I1211 08:30:44.847277 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f51c05dd-1881-4644-9a40-7e6e2174c591","Type":"ContainerStarted","Data":"fba3f221fcb908ad23e49f61db1251649ff4e52d00bbe28b5eda6d0f88b1af81"} Dec 11 08:30:44 crc kubenswrapper[4860]: I1211 08:30:44.995693 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-kb24v"] Dec 11 08:30:45 crc kubenswrapper[4860]: W1211 08:30:45.000331 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podff7bf9ff_e357_4af7_b708_2eb6298b85a8.slice/crio-0d38e50b018a6f6589d87260420b91c51c0f2073b23b0d050697dc03421f6245 WatchSource:0}: Error finding container 0d38e50b018a6f6589d87260420b91c51c0f2073b23b0d050697dc03421f6245: Status 404 returned error can't find the container with id 0d38e50b018a6f6589d87260420b91c51c0f2073b23b0d050697dc03421f6245 Dec 11 08:30:45 crc kubenswrapper[4860]: W1211 08:30:45.210262 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb7bf9af0_6051_4d90_8918_9fd00cb869cd.slice/crio-022e94a74513eb2296d249c67e91f9b29749af16534ba4fbbd2848af54b9ab6e WatchSource:0}: Error finding container 022e94a74513eb2296d249c67e91f9b29749af16534ba4fbbd2848af54b9ab6e: Status 404 returned error can't find the container with id 022e94a74513eb2296d249c67e91f9b29749af16534ba4fbbd2848af54b9ab6e Dec 11 08:30:45 crc kubenswrapper[4860]: I1211 08:30:45.212011 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 11 08:30:45 crc kubenswrapper[4860]: I1211 08:30:45.593944 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="83082bff-37e1-4ac1-9074-ae9d69ed66b0" path="/var/lib/kubelet/pods/83082bff-37e1-4ac1-9074-ae9d69ed66b0/volumes" Dec 11 08:30:45 crc kubenswrapper[4860]: I1211 08:30:45.596938 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f38e033-5243-4e65-a7e0-29f1282b063f" path="/var/lib/kubelet/pods/8f38e033-5243-4e65-a7e0-29f1282b063f/volumes" Dec 11 08:30:45 crc kubenswrapper[4860]: I1211 08:30:45.862801 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"b7bf9af0-6051-4d90-8918-9fd00cb869cd","Type":"ContainerStarted","Data":"022e94a74513eb2296d249c67e91f9b29749af16534ba4fbbd2848af54b9ab6e"} Dec 11 08:30:45 crc kubenswrapper[4860]: I1211 08:30:45.868803 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-kb24v" event={"ID":"ff7bf9ff-e357-4af7-b708-2eb6298b85a8","Type":"ContainerStarted","Data":"0d38e50b018a6f6589d87260420b91c51c0f2073b23b0d050697dc03421f6245"} Dec 11 08:30:45 crc kubenswrapper[4860]: I1211 08:30:45.884526 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f51c05dd-1881-4644-9a40-7e6e2174c591","Type":"ContainerStarted","Data":"f7ef67abbd8e1252f05bab0c2498f6eb1b6ec70940aeb498f5d03d8a8a3d2d1f"} Dec 11 08:30:46 crc kubenswrapper[4860]: I1211 08:30:46.898687 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f51c05dd-1881-4644-9a40-7e6e2174c591","Type":"ContainerStarted","Data":"f5d362e8b054c2545574b6cfcfbd3d51fc863ab0b58e9eb0ffbd48fe89fa3753"} Dec 11 08:30:46 crc kubenswrapper[4860]: I1211 08:30:46.903075 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"b7bf9af0-6051-4d90-8918-9fd00cb869cd","Type":"ContainerStarted","Data":"f59b01871832c1a6711d880ab82ef9b6d89e5f67fd5ae9869833a5fca7f9dbb7"} Dec 11 08:30:46 crc kubenswrapper[4860]: I1211 08:30:46.903124 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"b7bf9af0-6051-4d90-8918-9fd00cb869cd","Type":"ContainerStarted","Data":"dcf174a8128a55e0f4108e9f8997b98f93242b163df4d8b4b2c28677ad207dee"} Dec 11 08:30:46 crc kubenswrapper[4860]: I1211 08:30:46.928475 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.928449766 podStartE2EDuration="4.928449766s" podCreationTimestamp="2025-12-11 08:30:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:30:46.921182087 +0000 UTC m=+1179.649701152" watchObservedRunningTime="2025-12-11 08:30:46.928449766 +0000 UTC m=+1179.656968821" Dec 11 08:30:46 crc kubenswrapper[4860]: I1211 08:30:46.954489 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=2.954461228 podStartE2EDuration="2.954461228s" podCreationTimestamp="2025-12-11 08:30:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:30:46.948775508 +0000 UTC m=+1179.677294573" watchObservedRunningTime="2025-12-11 08:30:46.954461228 +0000 UTC m=+1179.682980283" Dec 11 08:30:53 crc kubenswrapper[4860]: I1211 08:30:53.558470 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 11 08:30:53 crc kubenswrapper[4860]: I1211 08:30:53.559123 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 11 08:30:53 crc kubenswrapper[4860]: I1211 08:30:53.600691 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 11 08:30:53 crc kubenswrapper[4860]: I1211 08:30:53.601297 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 11 08:30:53 crc kubenswrapper[4860]: I1211 08:30:53.995759 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-kb24v" event={"ID":"ff7bf9ff-e357-4af7-b708-2eb6298b85a8","Type":"ContainerStarted","Data":"ba86ec1b211d7519b3b1d47ede401a6153052bcb53b200f1e27b5cbe89e1a916"} Dec 11 08:30:53 crc kubenswrapper[4860]: I1211 08:30:53.996307 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 11 08:30:53 crc kubenswrapper[4860]: I1211 08:30:53.996327 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 11 08:30:54 crc kubenswrapper[4860]: I1211 08:30:54.554409 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 11 08:30:54 crc kubenswrapper[4860]: I1211 08:30:54.554524 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 11 08:30:54 crc kubenswrapper[4860]: I1211 08:30:54.591683 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 11 08:30:54 crc kubenswrapper[4860]: I1211 08:30:54.603359 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 11 08:30:54 crc kubenswrapper[4860]: I1211 08:30:54.625950 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-kb24v" podStartSLOduration=2.8211799490000002 podStartE2EDuration="10.625931185s" podCreationTimestamp="2025-12-11 08:30:44 +0000 UTC" firstStartedPulling="2025-12-11 08:30:45.001771737 +0000 UTC m=+1177.730290792" lastFinishedPulling="2025-12-11 08:30:52.806522973 +0000 UTC m=+1185.535042028" observedRunningTime="2025-12-11 08:30:54.010316169 +0000 UTC m=+1186.738835244" watchObservedRunningTime="2025-12-11 08:30:54.625931185 +0000 UTC m=+1187.354450240" Dec 11 08:30:55 crc kubenswrapper[4860]: I1211 08:30:55.006431 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 11 08:30:55 crc kubenswrapper[4860]: I1211 08:30:55.006858 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 11 08:30:56 crc kubenswrapper[4860]: I1211 08:30:56.178956 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 11 08:30:56 crc kubenswrapper[4860]: I1211 08:30:56.179348 4860 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 11 08:30:56 crc kubenswrapper[4860]: I1211 08:30:56.190308 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 11 08:30:57 crc kubenswrapper[4860]: I1211 08:30:57.037497 4860 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 11 08:30:57 crc kubenswrapper[4860]: I1211 08:30:57.037535 4860 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 11 08:30:57 crc kubenswrapper[4860]: I1211 08:30:57.132948 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 11 08:30:57 crc kubenswrapper[4860]: I1211 08:30:57.138952 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 11 08:30:58 crc kubenswrapper[4860]: I1211 08:30:58.993181 4860 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="fb9e2dd0-591f-4a0e-88aa-5f21285fbc77" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Dec 11 08:31:04 crc kubenswrapper[4860]: I1211 08:31:04.115887 4860 generic.go:334] "Generic (PLEG): container finished" podID="ff7bf9ff-e357-4af7-b708-2eb6298b85a8" containerID="ba86ec1b211d7519b3b1d47ede401a6153052bcb53b200f1e27b5cbe89e1a916" exitCode=0 Dec 11 08:31:04 crc kubenswrapper[4860]: I1211 08:31:04.115995 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-kb24v" event={"ID":"ff7bf9ff-e357-4af7-b708-2eb6298b85a8","Type":"ContainerDied","Data":"ba86ec1b211d7519b3b1d47ede401a6153052bcb53b200f1e27b5cbe89e1a916"} Dec 11 08:31:05 crc kubenswrapper[4860]: I1211 08:31:05.479447 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-kb24v" Dec 11 08:31:05 crc kubenswrapper[4860]: I1211 08:31:05.551302 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff7bf9ff-e357-4af7-b708-2eb6298b85a8-combined-ca-bundle\") pod \"ff7bf9ff-e357-4af7-b708-2eb6298b85a8\" (UID: \"ff7bf9ff-e357-4af7-b708-2eb6298b85a8\") " Dec 11 08:31:05 crc kubenswrapper[4860]: I1211 08:31:05.551381 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249h9\" (UniqueName: \"kubernetes.io/projected/ff7bf9ff-e357-4af7-b708-2eb6298b85a8-kube-api-access-249h9\") pod \"ff7bf9ff-e357-4af7-b708-2eb6298b85a8\" (UID: \"ff7bf9ff-e357-4af7-b708-2eb6298b85a8\") " Dec 11 08:31:05 crc kubenswrapper[4860]: I1211 08:31:05.551511 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ff7bf9ff-e357-4af7-b708-2eb6298b85a8-scripts\") pod \"ff7bf9ff-e357-4af7-b708-2eb6298b85a8\" (UID: \"ff7bf9ff-e357-4af7-b708-2eb6298b85a8\") " Dec 11 08:31:05 crc kubenswrapper[4860]: I1211 08:31:05.551632 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff7bf9ff-e357-4af7-b708-2eb6298b85a8-config-data\") pod \"ff7bf9ff-e357-4af7-b708-2eb6298b85a8\" (UID: \"ff7bf9ff-e357-4af7-b708-2eb6298b85a8\") " Dec 11 08:31:05 crc kubenswrapper[4860]: I1211 08:31:05.557174 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff7bf9ff-e357-4af7-b708-2eb6298b85a8-kube-api-access-249h9" (OuterVolumeSpecName: "kube-api-access-249h9") pod "ff7bf9ff-e357-4af7-b708-2eb6298b85a8" (UID: "ff7bf9ff-e357-4af7-b708-2eb6298b85a8"). InnerVolumeSpecName "kube-api-access-249h9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:31:05 crc kubenswrapper[4860]: I1211 08:31:05.557591 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff7bf9ff-e357-4af7-b708-2eb6298b85a8-scripts" (OuterVolumeSpecName: "scripts") pod "ff7bf9ff-e357-4af7-b708-2eb6298b85a8" (UID: "ff7bf9ff-e357-4af7-b708-2eb6298b85a8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:31:05 crc kubenswrapper[4860]: I1211 08:31:05.582148 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff7bf9ff-e357-4af7-b708-2eb6298b85a8-config-data" (OuterVolumeSpecName: "config-data") pod "ff7bf9ff-e357-4af7-b708-2eb6298b85a8" (UID: "ff7bf9ff-e357-4af7-b708-2eb6298b85a8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:31:05 crc kubenswrapper[4860]: I1211 08:31:05.587244 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff7bf9ff-e357-4af7-b708-2eb6298b85a8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ff7bf9ff-e357-4af7-b708-2eb6298b85a8" (UID: "ff7bf9ff-e357-4af7-b708-2eb6298b85a8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:31:05 crc kubenswrapper[4860]: I1211 08:31:05.655261 4860 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff7bf9ff-e357-4af7-b708-2eb6298b85a8-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 08:31:05 crc kubenswrapper[4860]: I1211 08:31:05.655303 4860 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff7bf9ff-e357-4af7-b708-2eb6298b85a8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 08:31:05 crc kubenswrapper[4860]: I1211 08:31:05.655319 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249h9\" (UniqueName: \"kubernetes.io/projected/ff7bf9ff-e357-4af7-b708-2eb6298b85a8-kube-api-access-249h9\") on node \"crc\" DevicePath \"\"" Dec 11 08:31:05 crc kubenswrapper[4860]: I1211 08:31:05.655333 4860 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ff7bf9ff-e357-4af7-b708-2eb6298b85a8-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 08:31:06 crc kubenswrapper[4860]: I1211 08:31:06.141562 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-kb24v" event={"ID":"ff7bf9ff-e357-4af7-b708-2eb6298b85a8","Type":"ContainerDied","Data":"0d38e50b018a6f6589d87260420b91c51c0f2073b23b0d050697dc03421f6245"} Dec 11 08:31:06 crc kubenswrapper[4860]: I1211 08:31:06.141994 4860 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0d38e50b018a6f6589d87260420b91c51c0f2073b23b0d050697dc03421f6245" Dec 11 08:31:06 crc kubenswrapper[4860]: I1211 08:31:06.141610 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-kb24v" Dec 11 08:31:06 crc kubenswrapper[4860]: I1211 08:31:06.145372 4860 generic.go:334] "Generic (PLEG): container finished" podID="fb9e2dd0-591f-4a0e-88aa-5f21285fbc77" containerID="0654f036c46ff9fe2261a835fd570095454a3493304ad137b4dc5ee75c6a22cb" exitCode=137 Dec 11 08:31:06 crc kubenswrapper[4860]: I1211 08:31:06.145402 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fb9e2dd0-591f-4a0e-88aa-5f21285fbc77","Type":"ContainerDied","Data":"0654f036c46ff9fe2261a835fd570095454a3493304ad137b4dc5ee75c6a22cb"} Dec 11 08:31:06 crc kubenswrapper[4860]: I1211 08:31:06.244819 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 11 08:31:06 crc kubenswrapper[4860]: E1211 08:31:06.245784 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff7bf9ff-e357-4af7-b708-2eb6298b85a8" containerName="nova-cell0-conductor-db-sync" Dec 11 08:31:06 crc kubenswrapper[4860]: I1211 08:31:06.245925 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff7bf9ff-e357-4af7-b708-2eb6298b85a8" containerName="nova-cell0-conductor-db-sync" Dec 11 08:31:06 crc kubenswrapper[4860]: I1211 08:31:06.246422 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff7bf9ff-e357-4af7-b708-2eb6298b85a8" containerName="nova-cell0-conductor-db-sync" Dec 11 08:31:06 crc kubenswrapper[4860]: I1211 08:31:06.247842 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 11 08:31:06 crc kubenswrapper[4860]: I1211 08:31:06.250172 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-w88hh" Dec 11 08:31:06 crc kubenswrapper[4860]: I1211 08:31:06.250870 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 11 08:31:06 crc kubenswrapper[4860]: I1211 08:31:06.257577 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 11 08:31:06 crc kubenswrapper[4860]: I1211 08:31:06.374192 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cchk6\" (UniqueName: \"kubernetes.io/projected/581da06a-7d8d-48dd-a5d9-d95ec1539a9f-kube-api-access-cchk6\") pod \"nova-cell0-conductor-0\" (UID: \"581da06a-7d8d-48dd-a5d9-d95ec1539a9f\") " pod="openstack/nova-cell0-conductor-0" Dec 11 08:31:06 crc kubenswrapper[4860]: I1211 08:31:06.374288 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/581da06a-7d8d-48dd-a5d9-d95ec1539a9f-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"581da06a-7d8d-48dd-a5d9-d95ec1539a9f\") " pod="openstack/nova-cell0-conductor-0" Dec 11 08:31:06 crc kubenswrapper[4860]: I1211 08:31:06.374357 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/581da06a-7d8d-48dd-a5d9-d95ec1539a9f-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"581da06a-7d8d-48dd-a5d9-d95ec1539a9f\") " pod="openstack/nova-cell0-conductor-0" Dec 11 08:31:06 crc kubenswrapper[4860]: I1211 08:31:06.481487 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cchk6\" (UniqueName: \"kubernetes.io/projected/581da06a-7d8d-48dd-a5d9-d95ec1539a9f-kube-api-access-cchk6\") pod \"nova-cell0-conductor-0\" (UID: \"581da06a-7d8d-48dd-a5d9-d95ec1539a9f\") " pod="openstack/nova-cell0-conductor-0" Dec 11 08:31:06 crc kubenswrapper[4860]: I1211 08:31:06.481613 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/581da06a-7d8d-48dd-a5d9-d95ec1539a9f-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"581da06a-7d8d-48dd-a5d9-d95ec1539a9f\") " pod="openstack/nova-cell0-conductor-0" Dec 11 08:31:06 crc kubenswrapper[4860]: I1211 08:31:06.481772 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/581da06a-7d8d-48dd-a5d9-d95ec1539a9f-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"581da06a-7d8d-48dd-a5d9-d95ec1539a9f\") " pod="openstack/nova-cell0-conductor-0" Dec 11 08:31:06 crc kubenswrapper[4860]: I1211 08:31:06.487615 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/581da06a-7d8d-48dd-a5d9-d95ec1539a9f-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"581da06a-7d8d-48dd-a5d9-d95ec1539a9f\") " pod="openstack/nova-cell0-conductor-0" Dec 11 08:31:06 crc kubenswrapper[4860]: I1211 08:31:06.491756 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/581da06a-7d8d-48dd-a5d9-d95ec1539a9f-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"581da06a-7d8d-48dd-a5d9-d95ec1539a9f\") " pod="openstack/nova-cell0-conductor-0" Dec 11 08:31:06 crc kubenswrapper[4860]: I1211 08:31:06.498344 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cchk6\" (UniqueName: \"kubernetes.io/projected/581da06a-7d8d-48dd-a5d9-d95ec1539a9f-kube-api-access-cchk6\") pod \"nova-cell0-conductor-0\" (UID: \"581da06a-7d8d-48dd-a5d9-d95ec1539a9f\") " pod="openstack/nova-cell0-conductor-0" Dec 11 08:31:06 crc kubenswrapper[4860]: I1211 08:31:06.578611 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 11 08:31:06 crc kubenswrapper[4860]: I1211 08:31:06.675774 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 11 08:31:06 crc kubenswrapper[4860]: I1211 08:31:06.787270 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fb9e2dd0-591f-4a0e-88aa-5f21285fbc77-scripts\") pod \"fb9e2dd0-591f-4a0e-88aa-5f21285fbc77\" (UID: \"fb9e2dd0-591f-4a0e-88aa-5f21285fbc77\") " Dec 11 08:31:06 crc kubenswrapper[4860]: I1211 08:31:06.787410 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb9e2dd0-591f-4a0e-88aa-5f21285fbc77-ceilometer-tls-certs\") pod \"fb9e2dd0-591f-4a0e-88aa-5f21285fbc77\" (UID: \"fb9e2dd0-591f-4a0e-88aa-5f21285fbc77\") " Dec 11 08:31:06 crc kubenswrapper[4860]: I1211 08:31:06.787457 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb9e2dd0-591f-4a0e-88aa-5f21285fbc77-config-data\") pod \"fb9e2dd0-591f-4a0e-88aa-5f21285fbc77\" (UID: \"fb9e2dd0-591f-4a0e-88aa-5f21285fbc77\") " Dec 11 08:31:06 crc kubenswrapper[4860]: I1211 08:31:06.787525 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2c7rl\" (UniqueName: \"kubernetes.io/projected/fb9e2dd0-591f-4a0e-88aa-5f21285fbc77-kube-api-access-2c7rl\") pod \"fb9e2dd0-591f-4a0e-88aa-5f21285fbc77\" (UID: \"fb9e2dd0-591f-4a0e-88aa-5f21285fbc77\") " Dec 11 08:31:06 crc kubenswrapper[4860]: I1211 08:31:06.787557 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fb9e2dd0-591f-4a0e-88aa-5f21285fbc77-sg-core-conf-yaml\") pod \"fb9e2dd0-591f-4a0e-88aa-5f21285fbc77\" (UID: \"fb9e2dd0-591f-4a0e-88aa-5f21285fbc77\") " Dec 11 08:31:06 crc kubenswrapper[4860]: I1211 08:31:06.787602 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fb9e2dd0-591f-4a0e-88aa-5f21285fbc77-log-httpd\") pod \"fb9e2dd0-591f-4a0e-88aa-5f21285fbc77\" (UID: \"fb9e2dd0-591f-4a0e-88aa-5f21285fbc77\") " Dec 11 08:31:06 crc kubenswrapper[4860]: I1211 08:31:06.787769 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fb9e2dd0-591f-4a0e-88aa-5f21285fbc77-run-httpd\") pod \"fb9e2dd0-591f-4a0e-88aa-5f21285fbc77\" (UID: \"fb9e2dd0-591f-4a0e-88aa-5f21285fbc77\") " Dec 11 08:31:06 crc kubenswrapper[4860]: I1211 08:31:06.787814 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb9e2dd0-591f-4a0e-88aa-5f21285fbc77-combined-ca-bundle\") pod \"fb9e2dd0-591f-4a0e-88aa-5f21285fbc77\" (UID: \"fb9e2dd0-591f-4a0e-88aa-5f21285fbc77\") " Dec 11 08:31:06 crc kubenswrapper[4860]: I1211 08:31:06.788260 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fb9e2dd0-591f-4a0e-88aa-5f21285fbc77-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "fb9e2dd0-591f-4a0e-88aa-5f21285fbc77" (UID: "fb9e2dd0-591f-4a0e-88aa-5f21285fbc77"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:31:06 crc kubenswrapper[4860]: I1211 08:31:06.788395 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fb9e2dd0-591f-4a0e-88aa-5f21285fbc77-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "fb9e2dd0-591f-4a0e-88aa-5f21285fbc77" (UID: "fb9e2dd0-591f-4a0e-88aa-5f21285fbc77"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:31:06 crc kubenswrapper[4860]: I1211 08:31:06.792874 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fb9e2dd0-591f-4a0e-88aa-5f21285fbc77-kube-api-access-2c7rl" (OuterVolumeSpecName: "kube-api-access-2c7rl") pod "fb9e2dd0-591f-4a0e-88aa-5f21285fbc77" (UID: "fb9e2dd0-591f-4a0e-88aa-5f21285fbc77"). InnerVolumeSpecName "kube-api-access-2c7rl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:31:06 crc kubenswrapper[4860]: I1211 08:31:06.806503 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb9e2dd0-591f-4a0e-88aa-5f21285fbc77-scripts" (OuterVolumeSpecName: "scripts") pod "fb9e2dd0-591f-4a0e-88aa-5f21285fbc77" (UID: "fb9e2dd0-591f-4a0e-88aa-5f21285fbc77"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:31:06 crc kubenswrapper[4860]: I1211 08:31:06.817936 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb9e2dd0-591f-4a0e-88aa-5f21285fbc77-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "fb9e2dd0-591f-4a0e-88aa-5f21285fbc77" (UID: "fb9e2dd0-591f-4a0e-88aa-5f21285fbc77"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:31:06 crc kubenswrapper[4860]: I1211 08:31:06.862249 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb9e2dd0-591f-4a0e-88aa-5f21285fbc77-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "fb9e2dd0-591f-4a0e-88aa-5f21285fbc77" (UID: "fb9e2dd0-591f-4a0e-88aa-5f21285fbc77"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:31:06 crc kubenswrapper[4860]: I1211 08:31:06.876223 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb9e2dd0-591f-4a0e-88aa-5f21285fbc77-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fb9e2dd0-591f-4a0e-88aa-5f21285fbc77" (UID: "fb9e2dd0-591f-4a0e-88aa-5f21285fbc77"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:31:06 crc kubenswrapper[4860]: I1211 08:31:06.890228 4860 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fb9e2dd0-591f-4a0e-88aa-5f21285fbc77-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 11 08:31:06 crc kubenswrapper[4860]: I1211 08:31:06.890261 4860 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb9e2dd0-591f-4a0e-88aa-5f21285fbc77-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 08:31:06 crc kubenswrapper[4860]: I1211 08:31:06.890272 4860 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fb9e2dd0-591f-4a0e-88aa-5f21285fbc77-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 08:31:06 crc kubenswrapper[4860]: I1211 08:31:06.890280 4860 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb9e2dd0-591f-4a0e-88aa-5f21285fbc77-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 11 08:31:06 crc kubenswrapper[4860]: I1211 08:31:06.890291 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2c7rl\" (UniqueName: \"kubernetes.io/projected/fb9e2dd0-591f-4a0e-88aa-5f21285fbc77-kube-api-access-2c7rl\") on node \"crc\" DevicePath \"\"" Dec 11 08:31:06 crc kubenswrapper[4860]: I1211 08:31:06.890304 4860 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fb9e2dd0-591f-4a0e-88aa-5f21285fbc77-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 11 08:31:06 crc kubenswrapper[4860]: I1211 08:31:06.890316 4860 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fb9e2dd0-591f-4a0e-88aa-5f21285fbc77-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 11 08:31:06 crc kubenswrapper[4860]: I1211 08:31:06.890723 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb9e2dd0-591f-4a0e-88aa-5f21285fbc77-config-data" (OuterVolumeSpecName: "config-data") pod "fb9e2dd0-591f-4a0e-88aa-5f21285fbc77" (UID: "fb9e2dd0-591f-4a0e-88aa-5f21285fbc77"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:31:06 crc kubenswrapper[4860]: I1211 08:31:06.992339 4860 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb9e2dd0-591f-4a0e-88aa-5f21285fbc77-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 08:31:07 crc kubenswrapper[4860]: I1211 08:31:07.028222 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 11 08:31:07 crc kubenswrapper[4860]: W1211 08:31:07.032792 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod581da06a_7d8d_48dd_a5d9_d95ec1539a9f.slice/crio-c08a0106cdcc3215d14a26f026119e349f4fb308e5c9d87e6305c892e278b3bc WatchSource:0}: Error finding container c08a0106cdcc3215d14a26f026119e349f4fb308e5c9d87e6305c892e278b3bc: Status 404 returned error can't find the container with id c08a0106cdcc3215d14a26f026119e349f4fb308e5c9d87e6305c892e278b3bc Dec 11 08:31:07 crc kubenswrapper[4860]: I1211 08:31:07.154257 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"581da06a-7d8d-48dd-a5d9-d95ec1539a9f","Type":"ContainerStarted","Data":"c08a0106cdcc3215d14a26f026119e349f4fb308e5c9d87e6305c892e278b3bc"} Dec 11 08:31:07 crc kubenswrapper[4860]: I1211 08:31:07.155815 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fb9e2dd0-591f-4a0e-88aa-5f21285fbc77","Type":"ContainerDied","Data":"a496c61f077dcb30e98e70f519be52f2cb0fe7bff1cedf2fa450405d885f360b"} Dec 11 08:31:07 crc kubenswrapper[4860]: I1211 08:31:07.155847 4860 scope.go:117] "RemoveContainer" containerID="0654f036c46ff9fe2261a835fd570095454a3493304ad137b4dc5ee75c6a22cb" Dec 11 08:31:07 crc kubenswrapper[4860]: I1211 08:31:07.155968 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 11 08:31:07 crc kubenswrapper[4860]: I1211 08:31:07.197286 4860 scope.go:117] "RemoveContainer" containerID="bc42015dc1ea63f1842e6450d911e1e49b5ea23e5bf2c6aaeeab303b302b119a" Dec 11 08:31:07 crc kubenswrapper[4860]: I1211 08:31:07.216453 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 11 08:31:07 crc kubenswrapper[4860]: I1211 08:31:07.222687 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 11 08:31:07 crc kubenswrapper[4860]: I1211 08:31:07.234389 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 11 08:31:07 crc kubenswrapper[4860]: E1211 08:31:07.234999 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb9e2dd0-591f-4a0e-88aa-5f21285fbc77" containerName="proxy-httpd" Dec 11 08:31:07 crc kubenswrapper[4860]: I1211 08:31:07.235020 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb9e2dd0-591f-4a0e-88aa-5f21285fbc77" containerName="proxy-httpd" Dec 11 08:31:07 crc kubenswrapper[4860]: E1211 08:31:07.235039 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb9e2dd0-591f-4a0e-88aa-5f21285fbc77" containerName="sg-core" Dec 11 08:31:07 crc kubenswrapper[4860]: I1211 08:31:07.235048 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb9e2dd0-591f-4a0e-88aa-5f21285fbc77" containerName="sg-core" Dec 11 08:31:07 crc kubenswrapper[4860]: E1211 08:31:07.235071 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb9e2dd0-591f-4a0e-88aa-5f21285fbc77" containerName="ceilometer-notification-agent" Dec 11 08:31:07 crc kubenswrapper[4860]: I1211 08:31:07.235079 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb9e2dd0-591f-4a0e-88aa-5f21285fbc77" containerName="ceilometer-notification-agent" Dec 11 08:31:07 crc kubenswrapper[4860]: E1211 08:31:07.235100 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb9e2dd0-591f-4a0e-88aa-5f21285fbc77" containerName="ceilometer-central-agent" Dec 11 08:31:07 crc kubenswrapper[4860]: I1211 08:31:07.235108 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb9e2dd0-591f-4a0e-88aa-5f21285fbc77" containerName="ceilometer-central-agent" Dec 11 08:31:07 crc kubenswrapper[4860]: I1211 08:31:07.235521 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb9e2dd0-591f-4a0e-88aa-5f21285fbc77" containerName="proxy-httpd" Dec 11 08:31:07 crc kubenswrapper[4860]: I1211 08:31:07.235550 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb9e2dd0-591f-4a0e-88aa-5f21285fbc77" containerName="ceilometer-central-agent" Dec 11 08:31:07 crc kubenswrapper[4860]: I1211 08:31:07.235572 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb9e2dd0-591f-4a0e-88aa-5f21285fbc77" containerName="sg-core" Dec 11 08:31:07 crc kubenswrapper[4860]: I1211 08:31:07.235591 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb9e2dd0-591f-4a0e-88aa-5f21285fbc77" containerName="ceilometer-notification-agent" Dec 11 08:31:07 crc kubenswrapper[4860]: I1211 08:31:07.243654 4860 scope.go:117] "RemoveContainer" containerID="967f589ab90b87704934b82341d0412adecad46c59feec2439df3412a407ac73" Dec 11 08:31:07 crc kubenswrapper[4860]: I1211 08:31:07.247099 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 11 08:31:07 crc kubenswrapper[4860]: I1211 08:31:07.248676 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 11 08:31:07 crc kubenswrapper[4860]: I1211 08:31:07.258616 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 11 08:31:07 crc kubenswrapper[4860]: I1211 08:31:07.258782 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 11 08:31:07 crc kubenswrapper[4860]: I1211 08:31:07.259161 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 11 08:31:07 crc kubenswrapper[4860]: I1211 08:31:07.295109 4860 scope.go:117] "RemoveContainer" containerID="2333689d78c9ee84e0452445e3f70919ddfeeccadb977eb6bd41e9ac35f23fd0" Dec 11 08:31:07 crc kubenswrapper[4860]: I1211 08:31:07.296860 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/47b450dd-42c9-428c-aa47-a971d2dfa3c5-scripts\") pod \"ceilometer-0\" (UID: \"47b450dd-42c9-428c-aa47-a971d2dfa3c5\") " pod="openstack/ceilometer-0" Dec 11 08:31:07 crc kubenswrapper[4860]: I1211 08:31:07.296906 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47b450dd-42c9-428c-aa47-a971d2dfa3c5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"47b450dd-42c9-428c-aa47-a971d2dfa3c5\") " pod="openstack/ceilometer-0" Dec 11 08:31:07 crc kubenswrapper[4860]: I1211 08:31:07.297094 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/47b450dd-42c9-428c-aa47-a971d2dfa3c5-log-httpd\") pod \"ceilometer-0\" (UID: \"47b450dd-42c9-428c-aa47-a971d2dfa3c5\") " pod="openstack/ceilometer-0" Dec 11 08:31:07 crc kubenswrapper[4860]: I1211 08:31:07.297128 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/47b450dd-42c9-428c-aa47-a971d2dfa3c5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"47b450dd-42c9-428c-aa47-a971d2dfa3c5\") " pod="openstack/ceilometer-0" Dec 11 08:31:07 crc kubenswrapper[4860]: I1211 08:31:07.297418 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/47b450dd-42c9-428c-aa47-a971d2dfa3c5-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"47b450dd-42c9-428c-aa47-a971d2dfa3c5\") " pod="openstack/ceilometer-0" Dec 11 08:31:07 crc kubenswrapper[4860]: I1211 08:31:07.297495 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9z89t\" (UniqueName: \"kubernetes.io/projected/47b450dd-42c9-428c-aa47-a971d2dfa3c5-kube-api-access-9z89t\") pod \"ceilometer-0\" (UID: \"47b450dd-42c9-428c-aa47-a971d2dfa3c5\") " pod="openstack/ceilometer-0" Dec 11 08:31:07 crc kubenswrapper[4860]: I1211 08:31:07.297786 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47b450dd-42c9-428c-aa47-a971d2dfa3c5-config-data\") pod \"ceilometer-0\" (UID: \"47b450dd-42c9-428c-aa47-a971d2dfa3c5\") " pod="openstack/ceilometer-0" Dec 11 08:31:07 crc kubenswrapper[4860]: I1211 08:31:07.297847 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/47b450dd-42c9-428c-aa47-a971d2dfa3c5-run-httpd\") pod \"ceilometer-0\" (UID: \"47b450dd-42c9-428c-aa47-a971d2dfa3c5\") " pod="openstack/ceilometer-0" Dec 11 08:31:07 crc kubenswrapper[4860]: E1211 08:31:07.357445 4860 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfb9e2dd0_591f_4a0e_88aa_5f21285fbc77.slice\": RecentStats: unable to find data in memory cache]" Dec 11 08:31:07 crc kubenswrapper[4860]: I1211 08:31:07.401545 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47b450dd-42c9-428c-aa47-a971d2dfa3c5-config-data\") pod \"ceilometer-0\" (UID: \"47b450dd-42c9-428c-aa47-a971d2dfa3c5\") " pod="openstack/ceilometer-0" Dec 11 08:31:07 crc kubenswrapper[4860]: I1211 08:31:07.401872 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/47b450dd-42c9-428c-aa47-a971d2dfa3c5-run-httpd\") pod \"ceilometer-0\" (UID: \"47b450dd-42c9-428c-aa47-a971d2dfa3c5\") " pod="openstack/ceilometer-0" Dec 11 08:31:07 crc kubenswrapper[4860]: I1211 08:31:07.402031 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/47b450dd-42c9-428c-aa47-a971d2dfa3c5-scripts\") pod \"ceilometer-0\" (UID: \"47b450dd-42c9-428c-aa47-a971d2dfa3c5\") " pod="openstack/ceilometer-0" Dec 11 08:31:07 crc kubenswrapper[4860]: I1211 08:31:07.402062 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47b450dd-42c9-428c-aa47-a971d2dfa3c5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"47b450dd-42c9-428c-aa47-a971d2dfa3c5\") " pod="openstack/ceilometer-0" Dec 11 08:31:07 crc kubenswrapper[4860]: I1211 08:31:07.402289 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/47b450dd-42c9-428c-aa47-a971d2dfa3c5-log-httpd\") pod \"ceilometer-0\" (UID: \"47b450dd-42c9-428c-aa47-a971d2dfa3c5\") " pod="openstack/ceilometer-0" Dec 11 08:31:07 crc kubenswrapper[4860]: I1211 08:31:07.402327 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/47b450dd-42c9-428c-aa47-a971d2dfa3c5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"47b450dd-42c9-428c-aa47-a971d2dfa3c5\") " pod="openstack/ceilometer-0" Dec 11 08:31:07 crc kubenswrapper[4860]: I1211 08:31:07.402446 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/47b450dd-42c9-428c-aa47-a971d2dfa3c5-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"47b450dd-42c9-428c-aa47-a971d2dfa3c5\") " pod="openstack/ceilometer-0" Dec 11 08:31:07 crc kubenswrapper[4860]: I1211 08:31:07.402473 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9z89t\" (UniqueName: \"kubernetes.io/projected/47b450dd-42c9-428c-aa47-a971d2dfa3c5-kube-api-access-9z89t\") pod \"ceilometer-0\" (UID: \"47b450dd-42c9-428c-aa47-a971d2dfa3c5\") " pod="openstack/ceilometer-0" Dec 11 08:31:07 crc kubenswrapper[4860]: I1211 08:31:07.402987 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/47b450dd-42c9-428c-aa47-a971d2dfa3c5-run-httpd\") pod \"ceilometer-0\" (UID: \"47b450dd-42c9-428c-aa47-a971d2dfa3c5\") " pod="openstack/ceilometer-0" Dec 11 08:31:07 crc kubenswrapper[4860]: I1211 08:31:07.403436 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/47b450dd-42c9-428c-aa47-a971d2dfa3c5-log-httpd\") pod \"ceilometer-0\" (UID: \"47b450dd-42c9-428c-aa47-a971d2dfa3c5\") " pod="openstack/ceilometer-0" Dec 11 08:31:07 crc kubenswrapper[4860]: I1211 08:31:07.407254 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/47b450dd-42c9-428c-aa47-a971d2dfa3c5-scripts\") pod \"ceilometer-0\" (UID: \"47b450dd-42c9-428c-aa47-a971d2dfa3c5\") " pod="openstack/ceilometer-0" Dec 11 08:31:07 crc kubenswrapper[4860]: I1211 08:31:07.408157 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/47b450dd-42c9-428c-aa47-a971d2dfa3c5-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"47b450dd-42c9-428c-aa47-a971d2dfa3c5\") " pod="openstack/ceilometer-0" Dec 11 08:31:07 crc kubenswrapper[4860]: I1211 08:31:07.408851 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/47b450dd-42c9-428c-aa47-a971d2dfa3c5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"47b450dd-42c9-428c-aa47-a971d2dfa3c5\") " pod="openstack/ceilometer-0" Dec 11 08:31:07 crc kubenswrapper[4860]: I1211 08:31:07.409448 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47b450dd-42c9-428c-aa47-a971d2dfa3c5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"47b450dd-42c9-428c-aa47-a971d2dfa3c5\") " pod="openstack/ceilometer-0" Dec 11 08:31:07 crc kubenswrapper[4860]: I1211 08:31:07.410506 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47b450dd-42c9-428c-aa47-a971d2dfa3c5-config-data\") pod \"ceilometer-0\" (UID: \"47b450dd-42c9-428c-aa47-a971d2dfa3c5\") " pod="openstack/ceilometer-0" Dec 11 08:31:07 crc kubenswrapper[4860]: I1211 08:31:07.420449 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9z89t\" (UniqueName: \"kubernetes.io/projected/47b450dd-42c9-428c-aa47-a971d2dfa3c5-kube-api-access-9z89t\") pod \"ceilometer-0\" (UID: \"47b450dd-42c9-428c-aa47-a971d2dfa3c5\") " pod="openstack/ceilometer-0" Dec 11 08:31:07 crc kubenswrapper[4860]: I1211 08:31:07.582422 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 11 08:31:07 crc kubenswrapper[4860]: I1211 08:31:07.605893 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fb9e2dd0-591f-4a0e-88aa-5f21285fbc77" path="/var/lib/kubelet/pods/fb9e2dd0-591f-4a0e-88aa-5f21285fbc77/volumes" Dec 11 08:31:08 crc kubenswrapper[4860]: I1211 08:31:08.084763 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 11 08:31:08 crc kubenswrapper[4860]: W1211 08:31:08.091621 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod47b450dd_42c9_428c_aa47_a971d2dfa3c5.slice/crio-8b435cff8d1221c4dc29857a7afa255bf2214be17b34ae8ce7b90506c02b5094 WatchSource:0}: Error finding container 8b435cff8d1221c4dc29857a7afa255bf2214be17b34ae8ce7b90506c02b5094: Status 404 returned error can't find the container with id 8b435cff8d1221c4dc29857a7afa255bf2214be17b34ae8ce7b90506c02b5094 Dec 11 08:31:08 crc kubenswrapper[4860]: I1211 08:31:08.169096 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"581da06a-7d8d-48dd-a5d9-d95ec1539a9f","Type":"ContainerStarted","Data":"8609f06d910af8f644a0da16029043520feb3d7f69ff305789d399e9cee4c6d0"} Dec 11 08:31:08 crc kubenswrapper[4860]: I1211 08:31:08.169257 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Dec 11 08:31:08 crc kubenswrapper[4860]: I1211 08:31:08.170056 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"47b450dd-42c9-428c-aa47-a971d2dfa3c5","Type":"ContainerStarted","Data":"8b435cff8d1221c4dc29857a7afa255bf2214be17b34ae8ce7b90506c02b5094"} Dec 11 08:31:08 crc kubenswrapper[4860]: I1211 08:31:08.184389 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.184353999 podStartE2EDuration="2.184353999s" podCreationTimestamp="2025-12-11 08:31:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:31:08.184144362 +0000 UTC m=+1200.912663447" watchObservedRunningTime="2025-12-11 08:31:08.184353999 +0000 UTC m=+1200.912873054" Dec 11 08:31:08 crc kubenswrapper[4860]: I1211 08:31:08.795216 4860 patch_prober.go:28] interesting pod/machine-config-daemon-99qgp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 08:31:08 crc kubenswrapper[4860]: I1211 08:31:08.795632 4860 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 08:31:10 crc kubenswrapper[4860]: I1211 08:31:10.204379 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"47b450dd-42c9-428c-aa47-a971d2dfa3c5","Type":"ContainerStarted","Data":"c2ea26488b5d7472bd8221a6fff096ef6aef5589626b5e5b3f29566e841cfb3a"} Dec 11 08:31:11 crc kubenswrapper[4860]: I1211 08:31:11.216388 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"47b450dd-42c9-428c-aa47-a971d2dfa3c5","Type":"ContainerStarted","Data":"be79c2475bfd46873eb93a3a1f2a092d8cff2b6213a2d7e5015c7d8fa6d222d9"} Dec 11 08:31:14 crc kubenswrapper[4860]: I1211 08:31:14.249968 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"47b450dd-42c9-428c-aa47-a971d2dfa3c5","Type":"ContainerStarted","Data":"9f5fcb630d5bb7b7347be31a08fc385034a9bc7d82ee521a19e39962890ea047"} Dec 11 08:31:16 crc kubenswrapper[4860]: I1211 08:31:16.276551 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"47b450dd-42c9-428c-aa47-a971d2dfa3c5","Type":"ContainerStarted","Data":"98f36f332db8260b7bb17975f0dd0b38806bc4f0217970243d7c9ea4b2839ede"} Dec 11 08:31:16 crc kubenswrapper[4860]: I1211 08:31:16.277627 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 11 08:31:16 crc kubenswrapper[4860]: I1211 08:31:16.618862 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Dec 11 08:31:16 crc kubenswrapper[4860]: I1211 08:31:16.645101 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.412759762 podStartE2EDuration="9.645078117s" podCreationTimestamp="2025-12-11 08:31:07 +0000 UTC" firstStartedPulling="2025-12-11 08:31:08.094037776 +0000 UTC m=+1200.822556831" lastFinishedPulling="2025-12-11 08:31:15.326356131 +0000 UTC m=+1208.054875186" observedRunningTime="2025-12-11 08:31:16.302582458 +0000 UTC m=+1209.031101533" watchObservedRunningTime="2025-12-11 08:31:16.645078117 +0000 UTC m=+1209.373597182" Dec 11 08:31:17 crc kubenswrapper[4860]: I1211 08:31:17.136168 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-nvt7h"] Dec 11 08:31:17 crc kubenswrapper[4860]: I1211 08:31:17.141749 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-nvt7h" Dec 11 08:31:17 crc kubenswrapper[4860]: I1211 08:31:17.144977 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Dec 11 08:31:17 crc kubenswrapper[4860]: I1211 08:31:17.145224 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Dec 11 08:31:17 crc kubenswrapper[4860]: I1211 08:31:17.154454 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-nvt7h"] Dec 11 08:31:17 crc kubenswrapper[4860]: I1211 08:31:17.249571 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/983cbde3-46fe-4643-b213-c74900941587-config-data\") pod \"nova-cell0-cell-mapping-nvt7h\" (UID: \"983cbde3-46fe-4643-b213-c74900941587\") " pod="openstack/nova-cell0-cell-mapping-nvt7h" Dec 11 08:31:17 crc kubenswrapper[4860]: I1211 08:31:17.249639 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pnq45\" (UniqueName: \"kubernetes.io/projected/983cbde3-46fe-4643-b213-c74900941587-kube-api-access-pnq45\") pod \"nova-cell0-cell-mapping-nvt7h\" (UID: \"983cbde3-46fe-4643-b213-c74900941587\") " pod="openstack/nova-cell0-cell-mapping-nvt7h" Dec 11 08:31:17 crc kubenswrapper[4860]: I1211 08:31:17.250311 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/983cbde3-46fe-4643-b213-c74900941587-scripts\") pod \"nova-cell0-cell-mapping-nvt7h\" (UID: \"983cbde3-46fe-4643-b213-c74900941587\") " pod="openstack/nova-cell0-cell-mapping-nvt7h" Dec 11 08:31:17 crc kubenswrapper[4860]: I1211 08:31:17.250378 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/983cbde3-46fe-4643-b213-c74900941587-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-nvt7h\" (UID: \"983cbde3-46fe-4643-b213-c74900941587\") " pod="openstack/nova-cell0-cell-mapping-nvt7h" Dec 11 08:31:17 crc kubenswrapper[4860]: I1211 08:31:17.346038 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 11 08:31:17 crc kubenswrapper[4860]: I1211 08:31:17.349470 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 11 08:31:17 crc kubenswrapper[4860]: I1211 08:31:17.354475 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/983cbde3-46fe-4643-b213-c74900941587-scripts\") pod \"nova-cell0-cell-mapping-nvt7h\" (UID: \"983cbde3-46fe-4643-b213-c74900941587\") " pod="openstack/nova-cell0-cell-mapping-nvt7h" Dec 11 08:31:17 crc kubenswrapper[4860]: I1211 08:31:17.354804 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/983cbde3-46fe-4643-b213-c74900941587-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-nvt7h\" (UID: \"983cbde3-46fe-4643-b213-c74900941587\") " pod="openstack/nova-cell0-cell-mapping-nvt7h" Dec 11 08:31:17 crc kubenswrapper[4860]: I1211 08:31:17.354998 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/983cbde3-46fe-4643-b213-c74900941587-config-data\") pod \"nova-cell0-cell-mapping-nvt7h\" (UID: \"983cbde3-46fe-4643-b213-c74900941587\") " pod="openstack/nova-cell0-cell-mapping-nvt7h" Dec 11 08:31:17 crc kubenswrapper[4860]: I1211 08:31:17.355124 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pnq45\" (UniqueName: \"kubernetes.io/projected/983cbde3-46fe-4643-b213-c74900941587-kube-api-access-pnq45\") pod \"nova-cell0-cell-mapping-nvt7h\" (UID: \"983cbde3-46fe-4643-b213-c74900941587\") " pod="openstack/nova-cell0-cell-mapping-nvt7h" Dec 11 08:31:17 crc kubenswrapper[4860]: I1211 08:31:17.361799 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/983cbde3-46fe-4643-b213-c74900941587-scripts\") pod \"nova-cell0-cell-mapping-nvt7h\" (UID: \"983cbde3-46fe-4643-b213-c74900941587\") " pod="openstack/nova-cell0-cell-mapping-nvt7h" Dec 11 08:31:17 crc kubenswrapper[4860]: I1211 08:31:17.366730 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 11 08:31:17 crc kubenswrapper[4860]: I1211 08:31:17.367421 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/983cbde3-46fe-4643-b213-c74900941587-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-nvt7h\" (UID: \"983cbde3-46fe-4643-b213-c74900941587\") " pod="openstack/nova-cell0-cell-mapping-nvt7h" Dec 11 08:31:17 crc kubenswrapper[4860]: I1211 08:31:17.375312 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/983cbde3-46fe-4643-b213-c74900941587-config-data\") pod \"nova-cell0-cell-mapping-nvt7h\" (UID: \"983cbde3-46fe-4643-b213-c74900941587\") " pod="openstack/nova-cell0-cell-mapping-nvt7h" Dec 11 08:31:17 crc kubenswrapper[4860]: I1211 08:31:17.378935 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 11 08:31:17 crc kubenswrapper[4860]: I1211 08:31:17.380602 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 11 08:31:17 crc kubenswrapper[4860]: I1211 08:31:17.410732 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 11 08:31:17 crc kubenswrapper[4860]: I1211 08:31:17.417257 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pnq45\" (UniqueName: \"kubernetes.io/projected/983cbde3-46fe-4643-b213-c74900941587-kube-api-access-pnq45\") pod \"nova-cell0-cell-mapping-nvt7h\" (UID: \"983cbde3-46fe-4643-b213-c74900941587\") " pod="openstack/nova-cell0-cell-mapping-nvt7h" Dec 11 08:31:17 crc kubenswrapper[4860]: I1211 08:31:17.418111 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 11 08:31:17 crc kubenswrapper[4860]: I1211 08:31:17.439765 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 11 08:31:17 crc kubenswrapper[4860]: I1211 08:31:17.470520 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 11 08:31:17 crc kubenswrapper[4860]: I1211 08:31:17.470754 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-nvt7h" Dec 11 08:31:17 crc kubenswrapper[4860]: I1211 08:31:17.471929 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 11 08:31:17 crc kubenswrapper[4860]: I1211 08:31:17.472384 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ggrt5\" (UniqueName: \"kubernetes.io/projected/2ee186b9-2dbe-4492-89ff-4962cd4c942a-kube-api-access-ggrt5\") pod \"nova-scheduler-0\" (UID: \"2ee186b9-2dbe-4492-89ff-4962cd4c942a\") " pod="openstack/nova-scheduler-0" Dec 11 08:31:17 crc kubenswrapper[4860]: I1211 08:31:17.472465 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ee186b9-2dbe-4492-89ff-4962cd4c942a-config-data\") pod \"nova-scheduler-0\" (UID: \"2ee186b9-2dbe-4492-89ff-4962cd4c942a\") " pod="openstack/nova-scheduler-0" Dec 11 08:31:17 crc kubenswrapper[4860]: I1211 08:31:17.472488 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59f08d16-b65e-4758-b17f-471dcb23781b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"59f08d16-b65e-4758-b17f-471dcb23781b\") " pod="openstack/nova-api-0" Dec 11 08:31:17 crc kubenswrapper[4860]: I1211 08:31:17.472530 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59f08d16-b65e-4758-b17f-471dcb23781b-config-data\") pod \"nova-api-0\" (UID: \"59f08d16-b65e-4758-b17f-471dcb23781b\") " pod="openstack/nova-api-0" Dec 11 08:31:17 crc kubenswrapper[4860]: I1211 08:31:17.472554 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ee186b9-2dbe-4492-89ff-4962cd4c942a-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"2ee186b9-2dbe-4492-89ff-4962cd4c942a\") " pod="openstack/nova-scheduler-0" Dec 11 08:31:17 crc kubenswrapper[4860]: I1211 08:31:17.472633 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/59f08d16-b65e-4758-b17f-471dcb23781b-logs\") pod \"nova-api-0\" (UID: \"59f08d16-b65e-4758-b17f-471dcb23781b\") " pod="openstack/nova-api-0" Dec 11 08:31:17 crc kubenswrapper[4860]: I1211 08:31:17.472670 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-77fg9\" (UniqueName: \"kubernetes.io/projected/59f08d16-b65e-4758-b17f-471dcb23781b-kube-api-access-77fg9\") pod \"nova-api-0\" (UID: \"59f08d16-b65e-4758-b17f-471dcb23781b\") " pod="openstack/nova-api-0" Dec 11 08:31:17 crc kubenswrapper[4860]: I1211 08:31:17.485924 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 11 08:31:17 crc kubenswrapper[4860]: I1211 08:31:17.500083 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 11 08:31:17 crc kubenswrapper[4860]: I1211 08:31:17.574696 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 11 08:31:17 crc kubenswrapper[4860]: I1211 08:31:17.630472 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ee186b9-2dbe-4492-89ff-4962cd4c942a-config-data\") pod \"nova-scheduler-0\" (UID: \"2ee186b9-2dbe-4492-89ff-4962cd4c942a\") " pod="openstack/nova-scheduler-0" Dec 11 08:31:17 crc kubenswrapper[4860]: I1211 08:31:17.630988 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59f08d16-b65e-4758-b17f-471dcb23781b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"59f08d16-b65e-4758-b17f-471dcb23781b\") " pod="openstack/nova-api-0" Dec 11 08:31:17 crc kubenswrapper[4860]: I1211 08:31:17.631052 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3954f800-6e02-4387-bd5f-c9dd575ce3c8-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"3954f800-6e02-4387-bd5f-c9dd575ce3c8\") " pod="openstack/nova-cell1-novncproxy-0" Dec 11 08:31:17 crc kubenswrapper[4860]: I1211 08:31:17.631121 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59f08d16-b65e-4758-b17f-471dcb23781b-config-data\") pod \"nova-api-0\" (UID: \"59f08d16-b65e-4758-b17f-471dcb23781b\") " pod="openstack/nova-api-0" Dec 11 08:31:17 crc kubenswrapper[4860]: I1211 08:31:17.631357 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3954f800-6e02-4387-bd5f-c9dd575ce3c8-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"3954f800-6e02-4387-bd5f-c9dd575ce3c8\") " pod="openstack/nova-cell1-novncproxy-0" Dec 11 08:31:17 crc kubenswrapper[4860]: I1211 08:31:17.631401 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ee186b9-2dbe-4492-89ff-4962cd4c942a-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"2ee186b9-2dbe-4492-89ff-4962cd4c942a\") " pod="openstack/nova-scheduler-0" Dec 11 08:31:17 crc kubenswrapper[4860]: I1211 08:31:17.631558 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/59f08d16-b65e-4758-b17f-471dcb23781b-logs\") pod \"nova-api-0\" (UID: \"59f08d16-b65e-4758-b17f-471dcb23781b\") " pod="openstack/nova-api-0" Dec 11 08:31:17 crc kubenswrapper[4860]: I1211 08:31:17.638061 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-77fg9\" (UniqueName: \"kubernetes.io/projected/59f08d16-b65e-4758-b17f-471dcb23781b-kube-api-access-77fg9\") pod \"nova-api-0\" (UID: \"59f08d16-b65e-4758-b17f-471dcb23781b\") " pod="openstack/nova-api-0" Dec 11 08:31:17 crc kubenswrapper[4860]: I1211 08:31:17.638256 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d4t22\" (UniqueName: \"kubernetes.io/projected/3954f800-6e02-4387-bd5f-c9dd575ce3c8-kube-api-access-d4t22\") pod \"nova-cell1-novncproxy-0\" (UID: \"3954f800-6e02-4387-bd5f-c9dd575ce3c8\") " pod="openstack/nova-cell1-novncproxy-0" Dec 11 08:31:17 crc kubenswrapper[4860]: I1211 08:31:17.638357 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ggrt5\" (UniqueName: \"kubernetes.io/projected/2ee186b9-2dbe-4492-89ff-4962cd4c942a-kube-api-access-ggrt5\") pod \"nova-scheduler-0\" (UID: \"2ee186b9-2dbe-4492-89ff-4962cd4c942a\") " pod="openstack/nova-scheduler-0" Dec 11 08:31:17 crc kubenswrapper[4860]: I1211 08:31:17.659981 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/59f08d16-b65e-4758-b17f-471dcb23781b-logs\") pod \"nova-api-0\" (UID: \"59f08d16-b65e-4758-b17f-471dcb23781b\") " pod="openstack/nova-api-0" Dec 11 08:31:17 crc kubenswrapper[4860]: I1211 08:31:17.716395 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ee186b9-2dbe-4492-89ff-4962cd4c942a-config-data\") pod \"nova-scheduler-0\" (UID: \"2ee186b9-2dbe-4492-89ff-4962cd4c942a\") " pod="openstack/nova-scheduler-0" Dec 11 08:31:17 crc kubenswrapper[4860]: I1211 08:31:17.724470 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59f08d16-b65e-4758-b17f-471dcb23781b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"59f08d16-b65e-4758-b17f-471dcb23781b\") " pod="openstack/nova-api-0" Dec 11 08:31:17 crc kubenswrapper[4860]: I1211 08:31:17.734044 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59f08d16-b65e-4758-b17f-471dcb23781b-config-data\") pod \"nova-api-0\" (UID: \"59f08d16-b65e-4758-b17f-471dcb23781b\") " pod="openstack/nova-api-0" Dec 11 08:31:17 crc kubenswrapper[4860]: I1211 08:31:17.744744 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-77fg9\" (UniqueName: \"kubernetes.io/projected/59f08d16-b65e-4758-b17f-471dcb23781b-kube-api-access-77fg9\") pod \"nova-api-0\" (UID: \"59f08d16-b65e-4758-b17f-471dcb23781b\") " pod="openstack/nova-api-0" Dec 11 08:31:17 crc kubenswrapper[4860]: I1211 08:31:17.753014 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 11 08:31:17 crc kubenswrapper[4860]: I1211 08:31:17.756261 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ee186b9-2dbe-4492-89ff-4962cd4c942a-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"2ee186b9-2dbe-4492-89ff-4962cd4c942a\") " pod="openstack/nova-scheduler-0" Dec 11 08:31:17 crc kubenswrapper[4860]: I1211 08:31:17.757780 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3954f800-6e02-4387-bd5f-c9dd575ce3c8-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"3954f800-6e02-4387-bd5f-c9dd575ce3c8\") " pod="openstack/nova-cell1-novncproxy-0" Dec 11 08:31:17 crc kubenswrapper[4860]: I1211 08:31:17.757926 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bqt6f\" (UniqueName: \"kubernetes.io/projected/1747e5bf-b214-43a8-8276-8c2badd571ab-kube-api-access-bqt6f\") pod \"nova-metadata-0\" (UID: \"1747e5bf-b214-43a8-8276-8c2badd571ab\") " pod="openstack/nova-metadata-0" Dec 11 08:31:17 crc kubenswrapper[4860]: I1211 08:31:17.758065 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3954f800-6e02-4387-bd5f-c9dd575ce3c8-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"3954f800-6e02-4387-bd5f-c9dd575ce3c8\") " pod="openstack/nova-cell1-novncproxy-0" Dec 11 08:31:17 crc kubenswrapper[4860]: I1211 08:31:17.758166 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1747e5bf-b214-43a8-8276-8c2badd571ab-config-data\") pod \"nova-metadata-0\" (UID: \"1747e5bf-b214-43a8-8276-8c2badd571ab\") " pod="openstack/nova-metadata-0" Dec 11 08:31:17 crc kubenswrapper[4860]: I1211 08:31:17.758495 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d4t22\" (UniqueName: \"kubernetes.io/projected/3954f800-6e02-4387-bd5f-c9dd575ce3c8-kube-api-access-d4t22\") pod \"nova-cell1-novncproxy-0\" (UID: \"3954f800-6e02-4387-bd5f-c9dd575ce3c8\") " pod="openstack/nova-cell1-novncproxy-0" Dec 11 08:31:17 crc kubenswrapper[4860]: I1211 08:31:17.762978 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 11 08:31:17 crc kubenswrapper[4860]: I1211 08:31:17.763422 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ggrt5\" (UniqueName: \"kubernetes.io/projected/2ee186b9-2dbe-4492-89ff-4962cd4c942a-kube-api-access-ggrt5\") pod \"nova-scheduler-0\" (UID: \"2ee186b9-2dbe-4492-89ff-4962cd4c942a\") " pod="openstack/nova-scheduler-0" Dec 11 08:31:17 crc kubenswrapper[4860]: I1211 08:31:17.779681 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1747e5bf-b214-43a8-8276-8c2badd571ab-logs\") pod \"nova-metadata-0\" (UID: \"1747e5bf-b214-43a8-8276-8c2badd571ab\") " pod="openstack/nova-metadata-0" Dec 11 08:31:17 crc kubenswrapper[4860]: I1211 08:31:17.780134 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1747e5bf-b214-43a8-8276-8c2badd571ab-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"1747e5bf-b214-43a8-8276-8c2badd571ab\") " pod="openstack/nova-metadata-0" Dec 11 08:31:17 crc kubenswrapper[4860]: I1211 08:31:17.779968 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3954f800-6e02-4387-bd5f-c9dd575ce3c8-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"3954f800-6e02-4387-bd5f-c9dd575ce3c8\") " pod="openstack/nova-cell1-novncproxy-0" Dec 11 08:31:17 crc kubenswrapper[4860]: I1211 08:31:17.792491 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3954f800-6e02-4387-bd5f-c9dd575ce3c8-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"3954f800-6e02-4387-bd5f-c9dd575ce3c8\") " pod="openstack/nova-cell1-novncproxy-0" Dec 11 08:31:17 crc kubenswrapper[4860]: I1211 08:31:17.842433 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 11 08:31:17 crc kubenswrapper[4860]: I1211 08:31:17.842477 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-865f5d856f-wz6cz"] Dec 11 08:31:17 crc kubenswrapper[4860]: I1211 08:31:17.846144 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d4t22\" (UniqueName: \"kubernetes.io/projected/3954f800-6e02-4387-bd5f-c9dd575ce3c8-kube-api-access-d4t22\") pod \"nova-cell1-novncproxy-0\" (UID: \"3954f800-6e02-4387-bd5f-c9dd575ce3c8\") " pod="openstack/nova-cell1-novncproxy-0" Dec 11 08:31:17 crc kubenswrapper[4860]: I1211 08:31:17.846385 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-865f5d856f-wz6cz" Dec 11 08:31:17 crc kubenswrapper[4860]: I1211 08:31:17.856738 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-865f5d856f-wz6cz"] Dec 11 08:31:17 crc kubenswrapper[4860]: I1211 08:31:17.886153 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1747e5bf-b214-43a8-8276-8c2badd571ab-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"1747e5bf-b214-43a8-8276-8c2badd571ab\") " pod="openstack/nova-metadata-0" Dec 11 08:31:17 crc kubenswrapper[4860]: I1211 08:31:17.886261 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7e338f34-d693-481b-a8ed-eb8ecbdd5a3b-dns-svc\") pod \"dnsmasq-dns-865f5d856f-wz6cz\" (UID: \"7e338f34-d693-481b-a8ed-eb8ecbdd5a3b\") " pod="openstack/dnsmasq-dns-865f5d856f-wz6cz" Dec 11 08:31:17 crc kubenswrapper[4860]: I1211 08:31:17.886300 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7e338f34-d693-481b-a8ed-eb8ecbdd5a3b-ovsdbserver-nb\") pod \"dnsmasq-dns-865f5d856f-wz6cz\" (UID: \"7e338f34-d693-481b-a8ed-eb8ecbdd5a3b\") " pod="openstack/dnsmasq-dns-865f5d856f-wz6cz" Dec 11 08:31:17 crc kubenswrapper[4860]: I1211 08:31:17.886326 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bqt6f\" (UniqueName: \"kubernetes.io/projected/1747e5bf-b214-43a8-8276-8c2badd571ab-kube-api-access-bqt6f\") pod \"nova-metadata-0\" (UID: \"1747e5bf-b214-43a8-8276-8c2badd571ab\") " pod="openstack/nova-metadata-0" Dec 11 08:31:17 crc kubenswrapper[4860]: I1211 08:31:17.886371 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1747e5bf-b214-43a8-8276-8c2badd571ab-config-data\") pod \"nova-metadata-0\" (UID: \"1747e5bf-b214-43a8-8276-8c2badd571ab\") " pod="openstack/nova-metadata-0" Dec 11 08:31:17 crc kubenswrapper[4860]: I1211 08:31:17.886391 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7e338f34-d693-481b-a8ed-eb8ecbdd5a3b-ovsdbserver-sb\") pod \"dnsmasq-dns-865f5d856f-wz6cz\" (UID: \"7e338f34-d693-481b-a8ed-eb8ecbdd5a3b\") " pod="openstack/dnsmasq-dns-865f5d856f-wz6cz" Dec 11 08:31:17 crc kubenswrapper[4860]: I1211 08:31:17.886468 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7e338f34-d693-481b-a8ed-eb8ecbdd5a3b-dns-swift-storage-0\") pod \"dnsmasq-dns-865f5d856f-wz6cz\" (UID: \"7e338f34-d693-481b-a8ed-eb8ecbdd5a3b\") " pod="openstack/dnsmasq-dns-865f5d856f-wz6cz" Dec 11 08:31:17 crc kubenswrapper[4860]: I1211 08:31:17.886510 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4c9vd\" (UniqueName: \"kubernetes.io/projected/7e338f34-d693-481b-a8ed-eb8ecbdd5a3b-kube-api-access-4c9vd\") pod \"dnsmasq-dns-865f5d856f-wz6cz\" (UID: \"7e338f34-d693-481b-a8ed-eb8ecbdd5a3b\") " pod="openstack/dnsmasq-dns-865f5d856f-wz6cz" Dec 11 08:31:17 crc kubenswrapper[4860]: I1211 08:31:17.886603 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e338f34-d693-481b-a8ed-eb8ecbdd5a3b-config\") pod \"dnsmasq-dns-865f5d856f-wz6cz\" (UID: \"7e338f34-d693-481b-a8ed-eb8ecbdd5a3b\") " pod="openstack/dnsmasq-dns-865f5d856f-wz6cz" Dec 11 08:31:17 crc kubenswrapper[4860]: I1211 08:31:17.886627 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1747e5bf-b214-43a8-8276-8c2badd571ab-logs\") pod \"nova-metadata-0\" (UID: \"1747e5bf-b214-43a8-8276-8c2badd571ab\") " pod="openstack/nova-metadata-0" Dec 11 08:31:17 crc kubenswrapper[4860]: I1211 08:31:17.897932 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1747e5bf-b214-43a8-8276-8c2badd571ab-logs\") pod \"nova-metadata-0\" (UID: \"1747e5bf-b214-43a8-8276-8c2badd571ab\") " pod="openstack/nova-metadata-0" Dec 11 08:31:17 crc kubenswrapper[4860]: I1211 08:31:17.904847 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1747e5bf-b214-43a8-8276-8c2badd571ab-config-data\") pod \"nova-metadata-0\" (UID: \"1747e5bf-b214-43a8-8276-8c2badd571ab\") " pod="openstack/nova-metadata-0" Dec 11 08:31:17 crc kubenswrapper[4860]: I1211 08:31:17.940625 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1747e5bf-b214-43a8-8276-8c2badd571ab-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"1747e5bf-b214-43a8-8276-8c2badd571ab\") " pod="openstack/nova-metadata-0" Dec 11 08:31:17 crc kubenswrapper[4860]: I1211 08:31:17.956303 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bqt6f\" (UniqueName: \"kubernetes.io/projected/1747e5bf-b214-43a8-8276-8c2badd571ab-kube-api-access-bqt6f\") pod \"nova-metadata-0\" (UID: \"1747e5bf-b214-43a8-8276-8c2badd571ab\") " pod="openstack/nova-metadata-0" Dec 11 08:31:17 crc kubenswrapper[4860]: I1211 08:31:17.989391 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e338f34-d693-481b-a8ed-eb8ecbdd5a3b-config\") pod \"dnsmasq-dns-865f5d856f-wz6cz\" (UID: \"7e338f34-d693-481b-a8ed-eb8ecbdd5a3b\") " pod="openstack/dnsmasq-dns-865f5d856f-wz6cz" Dec 11 08:31:17 crc kubenswrapper[4860]: I1211 08:31:17.989542 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7e338f34-d693-481b-a8ed-eb8ecbdd5a3b-dns-svc\") pod \"dnsmasq-dns-865f5d856f-wz6cz\" (UID: \"7e338f34-d693-481b-a8ed-eb8ecbdd5a3b\") " pod="openstack/dnsmasq-dns-865f5d856f-wz6cz" Dec 11 08:31:17 crc kubenswrapper[4860]: I1211 08:31:17.989585 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7e338f34-d693-481b-a8ed-eb8ecbdd5a3b-ovsdbserver-nb\") pod \"dnsmasq-dns-865f5d856f-wz6cz\" (UID: \"7e338f34-d693-481b-a8ed-eb8ecbdd5a3b\") " pod="openstack/dnsmasq-dns-865f5d856f-wz6cz" Dec 11 08:31:17 crc kubenswrapper[4860]: I1211 08:31:17.992783 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7e338f34-d693-481b-a8ed-eb8ecbdd5a3b-ovsdbserver-sb\") pod \"dnsmasq-dns-865f5d856f-wz6cz\" (UID: \"7e338f34-d693-481b-a8ed-eb8ecbdd5a3b\") " pod="openstack/dnsmasq-dns-865f5d856f-wz6cz" Dec 11 08:31:17 crc kubenswrapper[4860]: I1211 08:31:17.992935 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7e338f34-d693-481b-a8ed-eb8ecbdd5a3b-dns-swift-storage-0\") pod \"dnsmasq-dns-865f5d856f-wz6cz\" (UID: \"7e338f34-d693-481b-a8ed-eb8ecbdd5a3b\") " pod="openstack/dnsmasq-dns-865f5d856f-wz6cz" Dec 11 08:31:17 crc kubenswrapper[4860]: I1211 08:31:17.992987 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4c9vd\" (UniqueName: \"kubernetes.io/projected/7e338f34-d693-481b-a8ed-eb8ecbdd5a3b-kube-api-access-4c9vd\") pod \"dnsmasq-dns-865f5d856f-wz6cz\" (UID: \"7e338f34-d693-481b-a8ed-eb8ecbdd5a3b\") " pod="openstack/dnsmasq-dns-865f5d856f-wz6cz" Dec 11 08:31:17 crc kubenswrapper[4860]: I1211 08:31:17.994318 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7e338f34-d693-481b-a8ed-eb8ecbdd5a3b-ovsdbserver-nb\") pod \"dnsmasq-dns-865f5d856f-wz6cz\" (UID: \"7e338f34-d693-481b-a8ed-eb8ecbdd5a3b\") " pod="openstack/dnsmasq-dns-865f5d856f-wz6cz" Dec 11 08:31:17 crc kubenswrapper[4860]: I1211 08:31:17.994495 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7e338f34-d693-481b-a8ed-eb8ecbdd5a3b-ovsdbserver-sb\") pod \"dnsmasq-dns-865f5d856f-wz6cz\" (UID: \"7e338f34-d693-481b-a8ed-eb8ecbdd5a3b\") " pod="openstack/dnsmasq-dns-865f5d856f-wz6cz" Dec 11 08:31:17 crc kubenswrapper[4860]: I1211 08:31:17.994944 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e338f34-d693-481b-a8ed-eb8ecbdd5a3b-config\") pod \"dnsmasq-dns-865f5d856f-wz6cz\" (UID: \"7e338f34-d693-481b-a8ed-eb8ecbdd5a3b\") " pod="openstack/dnsmasq-dns-865f5d856f-wz6cz" Dec 11 08:31:17 crc kubenswrapper[4860]: I1211 08:31:17.995249 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7e338f34-d693-481b-a8ed-eb8ecbdd5a3b-dns-svc\") pod \"dnsmasq-dns-865f5d856f-wz6cz\" (UID: \"7e338f34-d693-481b-a8ed-eb8ecbdd5a3b\") " pod="openstack/dnsmasq-dns-865f5d856f-wz6cz" Dec 11 08:31:17 crc kubenswrapper[4860]: I1211 08:31:17.995605 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7e338f34-d693-481b-a8ed-eb8ecbdd5a3b-dns-swift-storage-0\") pod \"dnsmasq-dns-865f5d856f-wz6cz\" (UID: \"7e338f34-d693-481b-a8ed-eb8ecbdd5a3b\") " pod="openstack/dnsmasq-dns-865f5d856f-wz6cz" Dec 11 08:31:18 crc kubenswrapper[4860]: I1211 08:31:18.015774 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 11 08:31:18 crc kubenswrapper[4860]: I1211 08:31:18.016364 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 11 08:31:18 crc kubenswrapper[4860]: I1211 08:31:18.020164 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 11 08:31:18 crc kubenswrapper[4860]: I1211 08:31:18.029533 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4c9vd\" (UniqueName: \"kubernetes.io/projected/7e338f34-d693-481b-a8ed-eb8ecbdd5a3b-kube-api-access-4c9vd\") pod \"dnsmasq-dns-865f5d856f-wz6cz\" (UID: \"7e338f34-d693-481b-a8ed-eb8ecbdd5a3b\") " pod="openstack/dnsmasq-dns-865f5d856f-wz6cz" Dec 11 08:31:18 crc kubenswrapper[4860]: I1211 08:31:18.153491 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 11 08:31:18 crc kubenswrapper[4860]: I1211 08:31:18.217656 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-865f5d856f-wz6cz" Dec 11 08:31:18 crc kubenswrapper[4860]: I1211 08:31:18.449807 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-nvt7h"] Dec 11 08:31:18 crc kubenswrapper[4860]: I1211 08:31:18.796892 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-bq4cn"] Dec 11 08:31:18 crc kubenswrapper[4860]: I1211 08:31:18.799201 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-bq4cn" Dec 11 08:31:18 crc kubenswrapper[4860]: I1211 08:31:18.802296 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 11 08:31:18 crc kubenswrapper[4860]: I1211 08:31:18.810692 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Dec 11 08:31:18 crc kubenswrapper[4860]: I1211 08:31:18.829223 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-bq4cn"] Dec 11 08:31:18 crc kubenswrapper[4860]: I1211 08:31:18.847487 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 11 08:31:18 crc kubenswrapper[4860]: W1211 08:31:18.922323 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2ee186b9_2dbe_4492_89ff_4962cd4c942a.slice/crio-0ae686847bf69d44e17110b43f7a37203a7e7c485760132db74d8a3daad8b9e3 WatchSource:0}: Error finding container 0ae686847bf69d44e17110b43f7a37203a7e7c485760132db74d8a3daad8b9e3: Status 404 returned error can't find the container with id 0ae686847bf69d44e17110b43f7a37203a7e7c485760132db74d8a3daad8b9e3 Dec 11 08:31:18 crc kubenswrapper[4860]: I1211 08:31:18.935786 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af3bd0b9-ab28-4e7b-8316-3206671fb97b-config-data\") pod \"nova-cell1-conductor-db-sync-bq4cn\" (UID: \"af3bd0b9-ab28-4e7b-8316-3206671fb97b\") " pod="openstack/nova-cell1-conductor-db-sync-bq4cn" Dec 11 08:31:18 crc kubenswrapper[4860]: I1211 08:31:18.935944 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4xcxb\" (UniqueName: \"kubernetes.io/projected/af3bd0b9-ab28-4e7b-8316-3206671fb97b-kube-api-access-4xcxb\") pod \"nova-cell1-conductor-db-sync-bq4cn\" (UID: \"af3bd0b9-ab28-4e7b-8316-3206671fb97b\") " pod="openstack/nova-cell1-conductor-db-sync-bq4cn" Dec 11 08:31:18 crc kubenswrapper[4860]: I1211 08:31:18.936011 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af3bd0b9-ab28-4e7b-8316-3206671fb97b-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-bq4cn\" (UID: \"af3bd0b9-ab28-4e7b-8316-3206671fb97b\") " pod="openstack/nova-cell1-conductor-db-sync-bq4cn" Dec 11 08:31:18 crc kubenswrapper[4860]: I1211 08:31:18.936099 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/af3bd0b9-ab28-4e7b-8316-3206671fb97b-scripts\") pod \"nova-cell1-conductor-db-sync-bq4cn\" (UID: \"af3bd0b9-ab28-4e7b-8316-3206671fb97b\") " pod="openstack/nova-cell1-conductor-db-sync-bq4cn" Dec 11 08:31:18 crc kubenswrapper[4860]: I1211 08:31:18.938975 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 11 08:31:19 crc kubenswrapper[4860]: I1211 08:31:19.038557 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4xcxb\" (UniqueName: \"kubernetes.io/projected/af3bd0b9-ab28-4e7b-8316-3206671fb97b-kube-api-access-4xcxb\") pod \"nova-cell1-conductor-db-sync-bq4cn\" (UID: \"af3bd0b9-ab28-4e7b-8316-3206671fb97b\") " pod="openstack/nova-cell1-conductor-db-sync-bq4cn" Dec 11 08:31:19 crc kubenswrapper[4860]: I1211 08:31:19.039209 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af3bd0b9-ab28-4e7b-8316-3206671fb97b-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-bq4cn\" (UID: \"af3bd0b9-ab28-4e7b-8316-3206671fb97b\") " pod="openstack/nova-cell1-conductor-db-sync-bq4cn" Dec 11 08:31:19 crc kubenswrapper[4860]: I1211 08:31:19.040608 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/af3bd0b9-ab28-4e7b-8316-3206671fb97b-scripts\") pod \"nova-cell1-conductor-db-sync-bq4cn\" (UID: \"af3bd0b9-ab28-4e7b-8316-3206671fb97b\") " pod="openstack/nova-cell1-conductor-db-sync-bq4cn" Dec 11 08:31:19 crc kubenswrapper[4860]: I1211 08:31:19.040999 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af3bd0b9-ab28-4e7b-8316-3206671fb97b-config-data\") pod \"nova-cell1-conductor-db-sync-bq4cn\" (UID: \"af3bd0b9-ab28-4e7b-8316-3206671fb97b\") " pod="openstack/nova-cell1-conductor-db-sync-bq4cn" Dec 11 08:31:19 crc kubenswrapper[4860]: I1211 08:31:19.044909 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/af3bd0b9-ab28-4e7b-8316-3206671fb97b-scripts\") pod \"nova-cell1-conductor-db-sync-bq4cn\" (UID: \"af3bd0b9-ab28-4e7b-8316-3206671fb97b\") " pod="openstack/nova-cell1-conductor-db-sync-bq4cn" Dec 11 08:31:19 crc kubenswrapper[4860]: I1211 08:31:19.044992 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af3bd0b9-ab28-4e7b-8316-3206671fb97b-config-data\") pod \"nova-cell1-conductor-db-sync-bq4cn\" (UID: \"af3bd0b9-ab28-4e7b-8316-3206671fb97b\") " pod="openstack/nova-cell1-conductor-db-sync-bq4cn" Dec 11 08:31:19 crc kubenswrapper[4860]: I1211 08:31:19.045035 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af3bd0b9-ab28-4e7b-8316-3206671fb97b-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-bq4cn\" (UID: \"af3bd0b9-ab28-4e7b-8316-3206671fb97b\") " pod="openstack/nova-cell1-conductor-db-sync-bq4cn" Dec 11 08:31:19 crc kubenswrapper[4860]: I1211 08:31:19.056804 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4xcxb\" (UniqueName: \"kubernetes.io/projected/af3bd0b9-ab28-4e7b-8316-3206671fb97b-kube-api-access-4xcxb\") pod \"nova-cell1-conductor-db-sync-bq4cn\" (UID: \"af3bd0b9-ab28-4e7b-8316-3206671fb97b\") " pod="openstack/nova-cell1-conductor-db-sync-bq4cn" Dec 11 08:31:19 crc kubenswrapper[4860]: W1211 08:31:19.117272 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3954f800_6e02_4387_bd5f_c9dd575ce3c8.slice/crio-03d1b78bda444d6d3b54c9908434f2a3d90042a8510c403e9827c13cd23fd807 WatchSource:0}: Error finding container 03d1b78bda444d6d3b54c9908434f2a3d90042a8510c403e9827c13cd23fd807: Status 404 returned error can't find the container with id 03d1b78bda444d6d3b54c9908434f2a3d90042a8510c403e9827c13cd23fd807 Dec 11 08:31:19 crc kubenswrapper[4860]: I1211 08:31:19.124929 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 11 08:31:19 crc kubenswrapper[4860]: W1211 08:31:19.135169 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1747e5bf_b214_43a8_8276_8c2badd571ab.slice/crio-d2c9b360d0d6f8fbbdf6d3172b4cd1477fd2e25f41459862f7eae99a16037ac4 WatchSource:0}: Error finding container d2c9b360d0d6f8fbbdf6d3172b4cd1477fd2e25f41459862f7eae99a16037ac4: Status 404 returned error can't find the container with id d2c9b360d0d6f8fbbdf6d3172b4cd1477fd2e25f41459862f7eae99a16037ac4 Dec 11 08:31:19 crc kubenswrapper[4860]: I1211 08:31:19.135318 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-bq4cn" Dec 11 08:31:19 crc kubenswrapper[4860]: W1211 08:31:19.149618 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7e338f34_d693_481b_a8ed_eb8ecbdd5a3b.slice/crio-33f05c5474bbc667f0539ad9f823dea04ac0f1a4e75a4e55f7e0c4f5c172a2ed WatchSource:0}: Error finding container 33f05c5474bbc667f0539ad9f823dea04ac0f1a4e75a4e55f7e0c4f5c172a2ed: Status 404 returned error can't find the container with id 33f05c5474bbc667f0539ad9f823dea04ac0f1a4e75a4e55f7e0c4f5c172a2ed Dec 11 08:31:19 crc kubenswrapper[4860]: I1211 08:31:19.159423 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 11 08:31:19 crc kubenswrapper[4860]: I1211 08:31:19.172242 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-865f5d856f-wz6cz"] Dec 11 08:31:19 crc kubenswrapper[4860]: I1211 08:31:19.339447 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-nvt7h" event={"ID":"983cbde3-46fe-4643-b213-c74900941587","Type":"ContainerStarted","Data":"b29ada59900b63fcf4b78fcd929f9f91ae19df8e2b3a8506766956c7d67e078c"} Dec 11 08:31:19 crc kubenswrapper[4860]: I1211 08:31:19.340073 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-nvt7h" event={"ID":"983cbde3-46fe-4643-b213-c74900941587","Type":"ContainerStarted","Data":"9e6956fde9407ef3fe083d5ce4f3f47237f2eeb965f8471d7a4794756151253b"} Dec 11 08:31:19 crc kubenswrapper[4860]: I1211 08:31:19.347545 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1747e5bf-b214-43a8-8276-8c2badd571ab","Type":"ContainerStarted","Data":"d2c9b360d0d6f8fbbdf6d3172b4cd1477fd2e25f41459862f7eae99a16037ac4"} Dec 11 08:31:19 crc kubenswrapper[4860]: I1211 08:31:19.350635 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"59f08d16-b65e-4758-b17f-471dcb23781b","Type":"ContainerStarted","Data":"9ff5149f18d31af8a19b8040918f69c098dd1646825505c9219be72eee0b7fee"} Dec 11 08:31:19 crc kubenswrapper[4860]: I1211 08:31:19.353275 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"3954f800-6e02-4387-bd5f-c9dd575ce3c8","Type":"ContainerStarted","Data":"03d1b78bda444d6d3b54c9908434f2a3d90042a8510c403e9827c13cd23fd807"} Dec 11 08:31:19 crc kubenswrapper[4860]: I1211 08:31:19.357600 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-865f5d856f-wz6cz" event={"ID":"7e338f34-d693-481b-a8ed-eb8ecbdd5a3b","Type":"ContainerStarted","Data":"33f05c5474bbc667f0539ad9f823dea04ac0f1a4e75a4e55f7e0c4f5c172a2ed"} Dec 11 08:31:19 crc kubenswrapper[4860]: I1211 08:31:19.372052 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"2ee186b9-2dbe-4492-89ff-4962cd4c942a","Type":"ContainerStarted","Data":"0ae686847bf69d44e17110b43f7a37203a7e7c485760132db74d8a3daad8b9e3"} Dec 11 08:31:19 crc kubenswrapper[4860]: I1211 08:31:19.386520 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-nvt7h" podStartSLOduration=2.386494042 podStartE2EDuration="2.386494042s" podCreationTimestamp="2025-12-11 08:31:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:31:19.372084427 +0000 UTC m=+1212.100603482" watchObservedRunningTime="2025-12-11 08:31:19.386494042 +0000 UTC m=+1212.115013117" Dec 11 08:31:19 crc kubenswrapper[4860]: I1211 08:31:19.709531 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-bq4cn"] Dec 11 08:31:20 crc kubenswrapper[4860]: I1211 08:31:20.406841 4860 generic.go:334] "Generic (PLEG): container finished" podID="7e338f34-d693-481b-a8ed-eb8ecbdd5a3b" containerID="b72ee3f3e1ae8d023bd56cc1da250ac39f56b858ce74442ccc4f0e77c28e7c20" exitCode=0 Dec 11 08:31:20 crc kubenswrapper[4860]: I1211 08:31:20.406937 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-865f5d856f-wz6cz" event={"ID":"7e338f34-d693-481b-a8ed-eb8ecbdd5a3b","Type":"ContainerDied","Data":"b72ee3f3e1ae8d023bd56cc1da250ac39f56b858ce74442ccc4f0e77c28e7c20"} Dec 11 08:31:20 crc kubenswrapper[4860]: I1211 08:31:20.412344 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-bq4cn" event={"ID":"af3bd0b9-ab28-4e7b-8316-3206671fb97b","Type":"ContainerStarted","Data":"25fa35ea439bdf563989e6595788235e05f9c66929274bdfa3aa467f9c6dac01"} Dec 11 08:31:20 crc kubenswrapper[4860]: I1211 08:31:20.412393 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-bq4cn" event={"ID":"af3bd0b9-ab28-4e7b-8316-3206671fb97b","Type":"ContainerStarted","Data":"dbca1f64ace3a04a5f620953a8495bdfb3d8047e555166b4f1aaed502a742007"} Dec 11 08:31:20 crc kubenswrapper[4860]: I1211 08:31:20.492412 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-bq4cn" podStartSLOduration=2.492376984 podStartE2EDuration="2.492376984s" podCreationTimestamp="2025-12-11 08:31:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:31:20.454171438 +0000 UTC m=+1213.182690493" watchObservedRunningTime="2025-12-11 08:31:20.492376984 +0000 UTC m=+1213.220896039" Dec 11 08:31:21 crc kubenswrapper[4860]: I1211 08:31:21.428415 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 11 08:31:21 crc kubenswrapper[4860]: I1211 08:31:21.435795 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-865f5d856f-wz6cz" event={"ID":"7e338f34-d693-481b-a8ed-eb8ecbdd5a3b","Type":"ContainerStarted","Data":"b291a77f964230ad3604e0f8070421f9d71a6daf17b5e1350871337ab4bd9f19"} Dec 11 08:31:21 crc kubenswrapper[4860]: I1211 08:31:21.435947 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-865f5d856f-wz6cz" Dec 11 08:31:21 crc kubenswrapper[4860]: I1211 08:31:21.441790 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 11 08:31:21 crc kubenswrapper[4860]: I1211 08:31:21.462841 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-865f5d856f-wz6cz" podStartSLOduration=4.462807279 podStartE2EDuration="4.462807279s" podCreationTimestamp="2025-12-11 08:31:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:31:21.459327719 +0000 UTC m=+1214.187846794" watchObservedRunningTime="2025-12-11 08:31:21.462807279 +0000 UTC m=+1214.191326334" Dec 11 08:31:23 crc kubenswrapper[4860]: I1211 08:31:23.463163 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1747e5bf-b214-43a8-8276-8c2badd571ab","Type":"ContainerStarted","Data":"d9680e100d2b8c7ca46ba193a30f6785027443e460f3dde0db28424733778d75"} Dec 11 08:31:23 crc kubenswrapper[4860]: I1211 08:31:23.463685 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1747e5bf-b214-43a8-8276-8c2badd571ab","Type":"ContainerStarted","Data":"e0ca50f57c13291025fb7966a5d4309c6cd0f863e7d5488cc1c068dc2aaead6b"} Dec 11 08:31:23 crc kubenswrapper[4860]: I1211 08:31:23.463410 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="1747e5bf-b214-43a8-8276-8c2badd571ab" containerName="nova-metadata-metadata" containerID="cri-o://d9680e100d2b8c7ca46ba193a30f6785027443e460f3dde0db28424733778d75" gracePeriod=30 Dec 11 08:31:23 crc kubenswrapper[4860]: I1211 08:31:23.463307 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="1747e5bf-b214-43a8-8276-8c2badd571ab" containerName="nova-metadata-log" containerID="cri-o://e0ca50f57c13291025fb7966a5d4309c6cd0f863e7d5488cc1c068dc2aaead6b" gracePeriod=30 Dec 11 08:31:23 crc kubenswrapper[4860]: I1211 08:31:23.465913 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"59f08d16-b65e-4758-b17f-471dcb23781b","Type":"ContainerStarted","Data":"513c5f1935162d62bb01b78f733a9f9eb73e25f7cb0e6739aafc410d1b82df65"} Dec 11 08:31:23 crc kubenswrapper[4860]: I1211 08:31:23.465990 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"59f08d16-b65e-4758-b17f-471dcb23781b","Type":"ContainerStarted","Data":"b91d2fecd9207ff2bc6ae98d9326d1848747386c2259481d5a8b51b88bd731e1"} Dec 11 08:31:23 crc kubenswrapper[4860]: I1211 08:31:23.469426 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"3954f800-6e02-4387-bd5f-c9dd575ce3c8","Type":"ContainerStarted","Data":"b861a25799833453fb1b060105a06855f4c458ede12f385f939e4a216002e9bd"} Dec 11 08:31:23 crc kubenswrapper[4860]: I1211 08:31:23.469592 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="3954f800-6e02-4387-bd5f-c9dd575ce3c8" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://b861a25799833453fb1b060105a06855f4c458ede12f385f939e4a216002e9bd" gracePeriod=30 Dec 11 08:31:23 crc kubenswrapper[4860]: I1211 08:31:23.473776 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"2ee186b9-2dbe-4492-89ff-4962cd4c942a","Type":"ContainerStarted","Data":"7506728391e138c7cb3f179f605e565bb04063c28ba7fa01c16af44f73c687a2"} Dec 11 08:31:23 crc kubenswrapper[4860]: I1211 08:31:23.492963 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.126923072 podStartE2EDuration="6.492938677s" podCreationTimestamp="2025-12-11 08:31:17 +0000 UTC" firstStartedPulling="2025-12-11 08:31:19.190283085 +0000 UTC m=+1211.918802150" lastFinishedPulling="2025-12-11 08:31:22.5562987 +0000 UTC m=+1215.284817755" observedRunningTime="2025-12-11 08:31:23.48798046 +0000 UTC m=+1216.216499535" watchObservedRunningTime="2025-12-11 08:31:23.492938677 +0000 UTC m=+1216.221457732" Dec 11 08:31:23 crc kubenswrapper[4860]: I1211 08:31:23.526205 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.827810223 podStartE2EDuration="6.526185117s" podCreationTimestamp="2025-12-11 08:31:17 +0000 UTC" firstStartedPulling="2025-12-11 08:31:18.821559337 +0000 UTC m=+1211.550078392" lastFinishedPulling="2025-12-11 08:31:22.519934231 +0000 UTC m=+1215.248453286" observedRunningTime="2025-12-11 08:31:23.51773841 +0000 UTC m=+1216.246257495" watchObservedRunningTime="2025-12-11 08:31:23.526185117 +0000 UTC m=+1216.254704162" Dec 11 08:31:23 crc kubenswrapper[4860]: I1211 08:31:23.559688 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=3.165606134 podStartE2EDuration="6.559668225s" podCreationTimestamp="2025-12-11 08:31:17 +0000 UTC" firstStartedPulling="2025-12-11 08:31:19.125844999 +0000 UTC m=+1211.854364054" lastFinishedPulling="2025-12-11 08:31:22.5199071 +0000 UTC m=+1215.248426145" observedRunningTime="2025-12-11 08:31:23.536326307 +0000 UTC m=+1216.264845392" watchObservedRunningTime="2025-12-11 08:31:23.559668225 +0000 UTC m=+1216.288187280" Dec 11 08:31:23 crc kubenswrapper[4860]: I1211 08:31:23.561654 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.975234549 podStartE2EDuration="6.561630567s" podCreationTimestamp="2025-12-11 08:31:17 +0000 UTC" firstStartedPulling="2025-12-11 08:31:18.933584225 +0000 UTC m=+1211.662103280" lastFinishedPulling="2025-12-11 08:31:22.519980243 +0000 UTC m=+1215.248499298" observedRunningTime="2025-12-11 08:31:23.556129073 +0000 UTC m=+1216.284648138" watchObservedRunningTime="2025-12-11 08:31:23.561630567 +0000 UTC m=+1216.290149622" Dec 11 08:31:24 crc kubenswrapper[4860]: I1211 08:31:24.172776 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 11 08:31:24 crc kubenswrapper[4860]: I1211 08:31:24.305465 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1747e5bf-b214-43a8-8276-8c2badd571ab-config-data\") pod \"1747e5bf-b214-43a8-8276-8c2badd571ab\" (UID: \"1747e5bf-b214-43a8-8276-8c2badd571ab\") " Dec 11 08:31:24 crc kubenswrapper[4860]: I1211 08:31:24.305745 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1747e5bf-b214-43a8-8276-8c2badd571ab-combined-ca-bundle\") pod \"1747e5bf-b214-43a8-8276-8c2badd571ab\" (UID: \"1747e5bf-b214-43a8-8276-8c2badd571ab\") " Dec 11 08:31:24 crc kubenswrapper[4860]: I1211 08:31:24.305820 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1747e5bf-b214-43a8-8276-8c2badd571ab-logs\") pod \"1747e5bf-b214-43a8-8276-8c2badd571ab\" (UID: \"1747e5bf-b214-43a8-8276-8c2badd571ab\") " Dec 11 08:31:24 crc kubenswrapper[4860]: I1211 08:31:24.305911 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bqt6f\" (UniqueName: \"kubernetes.io/projected/1747e5bf-b214-43a8-8276-8c2badd571ab-kube-api-access-bqt6f\") pod \"1747e5bf-b214-43a8-8276-8c2badd571ab\" (UID: \"1747e5bf-b214-43a8-8276-8c2badd571ab\") " Dec 11 08:31:24 crc kubenswrapper[4860]: I1211 08:31:24.306359 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1747e5bf-b214-43a8-8276-8c2badd571ab-logs" (OuterVolumeSpecName: "logs") pod "1747e5bf-b214-43a8-8276-8c2badd571ab" (UID: "1747e5bf-b214-43a8-8276-8c2badd571ab"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:31:24 crc kubenswrapper[4860]: I1211 08:31:24.306769 4860 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1747e5bf-b214-43a8-8276-8c2badd571ab-logs\") on node \"crc\" DevicePath \"\"" Dec 11 08:31:24 crc kubenswrapper[4860]: I1211 08:31:24.311928 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1747e5bf-b214-43a8-8276-8c2badd571ab-kube-api-access-bqt6f" (OuterVolumeSpecName: "kube-api-access-bqt6f") pod "1747e5bf-b214-43a8-8276-8c2badd571ab" (UID: "1747e5bf-b214-43a8-8276-8c2badd571ab"). InnerVolumeSpecName "kube-api-access-bqt6f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:31:24 crc kubenswrapper[4860]: I1211 08:31:24.341491 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1747e5bf-b214-43a8-8276-8c2badd571ab-config-data" (OuterVolumeSpecName: "config-data") pod "1747e5bf-b214-43a8-8276-8c2badd571ab" (UID: "1747e5bf-b214-43a8-8276-8c2badd571ab"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:31:24 crc kubenswrapper[4860]: I1211 08:31:24.341715 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1747e5bf-b214-43a8-8276-8c2badd571ab-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1747e5bf-b214-43a8-8276-8c2badd571ab" (UID: "1747e5bf-b214-43a8-8276-8c2badd571ab"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:31:24 crc kubenswrapper[4860]: I1211 08:31:24.408267 4860 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1747e5bf-b214-43a8-8276-8c2badd571ab-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 08:31:24 crc kubenswrapper[4860]: I1211 08:31:24.408302 4860 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1747e5bf-b214-43a8-8276-8c2badd571ab-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 08:31:24 crc kubenswrapper[4860]: I1211 08:31:24.408317 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bqt6f\" (UniqueName: \"kubernetes.io/projected/1747e5bf-b214-43a8-8276-8c2badd571ab-kube-api-access-bqt6f\") on node \"crc\" DevicePath \"\"" Dec 11 08:31:24 crc kubenswrapper[4860]: I1211 08:31:24.486010 4860 generic.go:334] "Generic (PLEG): container finished" podID="1747e5bf-b214-43a8-8276-8c2badd571ab" containerID="d9680e100d2b8c7ca46ba193a30f6785027443e460f3dde0db28424733778d75" exitCode=0 Dec 11 08:31:24 crc kubenswrapper[4860]: I1211 08:31:24.486052 4860 generic.go:334] "Generic (PLEG): container finished" podID="1747e5bf-b214-43a8-8276-8c2badd571ab" containerID="e0ca50f57c13291025fb7966a5d4309c6cd0f863e7d5488cc1c068dc2aaead6b" exitCode=143 Dec 11 08:31:24 crc kubenswrapper[4860]: I1211 08:31:24.486080 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 11 08:31:24 crc kubenswrapper[4860]: I1211 08:31:24.486142 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1747e5bf-b214-43a8-8276-8c2badd571ab","Type":"ContainerDied","Data":"d9680e100d2b8c7ca46ba193a30f6785027443e460f3dde0db28424733778d75"} Dec 11 08:31:24 crc kubenswrapper[4860]: I1211 08:31:24.486183 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1747e5bf-b214-43a8-8276-8c2badd571ab","Type":"ContainerDied","Data":"e0ca50f57c13291025fb7966a5d4309c6cd0f863e7d5488cc1c068dc2aaead6b"} Dec 11 08:31:24 crc kubenswrapper[4860]: I1211 08:31:24.486198 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1747e5bf-b214-43a8-8276-8c2badd571ab","Type":"ContainerDied","Data":"d2c9b360d0d6f8fbbdf6d3172b4cd1477fd2e25f41459862f7eae99a16037ac4"} Dec 11 08:31:24 crc kubenswrapper[4860]: I1211 08:31:24.486218 4860 scope.go:117] "RemoveContainer" containerID="d9680e100d2b8c7ca46ba193a30f6785027443e460f3dde0db28424733778d75" Dec 11 08:31:24 crc kubenswrapper[4860]: I1211 08:31:24.523662 4860 scope.go:117] "RemoveContainer" containerID="e0ca50f57c13291025fb7966a5d4309c6cd0f863e7d5488cc1c068dc2aaead6b" Dec 11 08:31:24 crc kubenswrapper[4860]: I1211 08:31:24.535067 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 11 08:31:24 crc kubenswrapper[4860]: I1211 08:31:24.550915 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 11 08:31:24 crc kubenswrapper[4860]: I1211 08:31:24.568986 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 11 08:31:24 crc kubenswrapper[4860]: E1211 08:31:24.569513 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1747e5bf-b214-43a8-8276-8c2badd571ab" containerName="nova-metadata-log" Dec 11 08:31:24 crc kubenswrapper[4860]: I1211 08:31:24.569554 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="1747e5bf-b214-43a8-8276-8c2badd571ab" containerName="nova-metadata-log" Dec 11 08:31:24 crc kubenswrapper[4860]: E1211 08:31:24.569602 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1747e5bf-b214-43a8-8276-8c2badd571ab" containerName="nova-metadata-metadata" Dec 11 08:31:24 crc kubenswrapper[4860]: I1211 08:31:24.569608 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="1747e5bf-b214-43a8-8276-8c2badd571ab" containerName="nova-metadata-metadata" Dec 11 08:31:24 crc kubenswrapper[4860]: I1211 08:31:24.569821 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="1747e5bf-b214-43a8-8276-8c2badd571ab" containerName="nova-metadata-log" Dec 11 08:31:24 crc kubenswrapper[4860]: I1211 08:31:24.569856 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="1747e5bf-b214-43a8-8276-8c2badd571ab" containerName="nova-metadata-metadata" Dec 11 08:31:24 crc kubenswrapper[4860]: I1211 08:31:24.570967 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 11 08:31:24 crc kubenswrapper[4860]: I1211 08:31:24.574029 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 11 08:31:24 crc kubenswrapper[4860]: I1211 08:31:24.574287 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 11 08:31:24 crc kubenswrapper[4860]: I1211 08:31:24.589249 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 11 08:31:24 crc kubenswrapper[4860]: I1211 08:31:24.639437 4860 scope.go:117] "RemoveContainer" containerID="d9680e100d2b8c7ca46ba193a30f6785027443e460f3dde0db28424733778d75" Dec 11 08:31:24 crc kubenswrapper[4860]: E1211 08:31:24.640005 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d9680e100d2b8c7ca46ba193a30f6785027443e460f3dde0db28424733778d75\": container with ID starting with d9680e100d2b8c7ca46ba193a30f6785027443e460f3dde0db28424733778d75 not found: ID does not exist" containerID="d9680e100d2b8c7ca46ba193a30f6785027443e460f3dde0db28424733778d75" Dec 11 08:31:24 crc kubenswrapper[4860]: I1211 08:31:24.640051 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d9680e100d2b8c7ca46ba193a30f6785027443e460f3dde0db28424733778d75"} err="failed to get container status \"d9680e100d2b8c7ca46ba193a30f6785027443e460f3dde0db28424733778d75\": rpc error: code = NotFound desc = could not find container \"d9680e100d2b8c7ca46ba193a30f6785027443e460f3dde0db28424733778d75\": container with ID starting with d9680e100d2b8c7ca46ba193a30f6785027443e460f3dde0db28424733778d75 not found: ID does not exist" Dec 11 08:31:24 crc kubenswrapper[4860]: I1211 08:31:24.640084 4860 scope.go:117] "RemoveContainer" containerID="e0ca50f57c13291025fb7966a5d4309c6cd0f863e7d5488cc1c068dc2aaead6b" Dec 11 08:31:24 crc kubenswrapper[4860]: E1211 08:31:24.640702 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e0ca50f57c13291025fb7966a5d4309c6cd0f863e7d5488cc1c068dc2aaead6b\": container with ID starting with e0ca50f57c13291025fb7966a5d4309c6cd0f863e7d5488cc1c068dc2aaead6b not found: ID does not exist" containerID="e0ca50f57c13291025fb7966a5d4309c6cd0f863e7d5488cc1c068dc2aaead6b" Dec 11 08:31:24 crc kubenswrapper[4860]: I1211 08:31:24.640806 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e0ca50f57c13291025fb7966a5d4309c6cd0f863e7d5488cc1c068dc2aaead6b"} err="failed to get container status \"e0ca50f57c13291025fb7966a5d4309c6cd0f863e7d5488cc1c068dc2aaead6b\": rpc error: code = NotFound desc = could not find container \"e0ca50f57c13291025fb7966a5d4309c6cd0f863e7d5488cc1c068dc2aaead6b\": container with ID starting with e0ca50f57c13291025fb7966a5d4309c6cd0f863e7d5488cc1c068dc2aaead6b not found: ID does not exist" Dec 11 08:31:24 crc kubenswrapper[4860]: I1211 08:31:24.640841 4860 scope.go:117] "RemoveContainer" containerID="d9680e100d2b8c7ca46ba193a30f6785027443e460f3dde0db28424733778d75" Dec 11 08:31:24 crc kubenswrapper[4860]: I1211 08:31:24.641160 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d9680e100d2b8c7ca46ba193a30f6785027443e460f3dde0db28424733778d75"} err="failed to get container status \"d9680e100d2b8c7ca46ba193a30f6785027443e460f3dde0db28424733778d75\": rpc error: code = NotFound desc = could not find container \"d9680e100d2b8c7ca46ba193a30f6785027443e460f3dde0db28424733778d75\": container with ID starting with d9680e100d2b8c7ca46ba193a30f6785027443e460f3dde0db28424733778d75 not found: ID does not exist" Dec 11 08:31:24 crc kubenswrapper[4860]: I1211 08:31:24.641194 4860 scope.go:117] "RemoveContainer" containerID="e0ca50f57c13291025fb7966a5d4309c6cd0f863e7d5488cc1c068dc2aaead6b" Dec 11 08:31:24 crc kubenswrapper[4860]: I1211 08:31:24.642514 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e0ca50f57c13291025fb7966a5d4309c6cd0f863e7d5488cc1c068dc2aaead6b"} err="failed to get container status \"e0ca50f57c13291025fb7966a5d4309c6cd0f863e7d5488cc1c068dc2aaead6b\": rpc error: code = NotFound desc = could not find container \"e0ca50f57c13291025fb7966a5d4309c6cd0f863e7d5488cc1c068dc2aaead6b\": container with ID starting with e0ca50f57c13291025fb7966a5d4309c6cd0f863e7d5488cc1c068dc2aaead6b not found: ID does not exist" Dec 11 08:31:24 crc kubenswrapper[4860]: I1211 08:31:24.715250 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/96c72d04-577b-4ed2-b9fe-d7319c5d56a6-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"96c72d04-577b-4ed2-b9fe-d7319c5d56a6\") " pod="openstack/nova-metadata-0" Dec 11 08:31:24 crc kubenswrapper[4860]: I1211 08:31:24.715372 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96c72d04-577b-4ed2-b9fe-d7319c5d56a6-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"96c72d04-577b-4ed2-b9fe-d7319c5d56a6\") " pod="openstack/nova-metadata-0" Dec 11 08:31:24 crc kubenswrapper[4860]: I1211 08:31:24.715528 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/96c72d04-577b-4ed2-b9fe-d7319c5d56a6-config-data\") pod \"nova-metadata-0\" (UID: \"96c72d04-577b-4ed2-b9fe-d7319c5d56a6\") " pod="openstack/nova-metadata-0" Dec 11 08:31:24 crc kubenswrapper[4860]: I1211 08:31:24.715575 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/96c72d04-577b-4ed2-b9fe-d7319c5d56a6-logs\") pod \"nova-metadata-0\" (UID: \"96c72d04-577b-4ed2-b9fe-d7319c5d56a6\") " pod="openstack/nova-metadata-0" Dec 11 08:31:24 crc kubenswrapper[4860]: I1211 08:31:24.715918 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xhvtj\" (UniqueName: \"kubernetes.io/projected/96c72d04-577b-4ed2-b9fe-d7319c5d56a6-kube-api-access-xhvtj\") pod \"nova-metadata-0\" (UID: \"96c72d04-577b-4ed2-b9fe-d7319c5d56a6\") " pod="openstack/nova-metadata-0" Dec 11 08:31:24 crc kubenswrapper[4860]: I1211 08:31:24.817477 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/96c72d04-577b-4ed2-b9fe-d7319c5d56a6-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"96c72d04-577b-4ed2-b9fe-d7319c5d56a6\") " pod="openstack/nova-metadata-0" Dec 11 08:31:24 crc kubenswrapper[4860]: I1211 08:31:24.817559 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96c72d04-577b-4ed2-b9fe-d7319c5d56a6-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"96c72d04-577b-4ed2-b9fe-d7319c5d56a6\") " pod="openstack/nova-metadata-0" Dec 11 08:31:24 crc kubenswrapper[4860]: I1211 08:31:24.817613 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/96c72d04-577b-4ed2-b9fe-d7319c5d56a6-config-data\") pod \"nova-metadata-0\" (UID: \"96c72d04-577b-4ed2-b9fe-d7319c5d56a6\") " pod="openstack/nova-metadata-0" Dec 11 08:31:24 crc kubenswrapper[4860]: I1211 08:31:24.817634 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/96c72d04-577b-4ed2-b9fe-d7319c5d56a6-logs\") pod \"nova-metadata-0\" (UID: \"96c72d04-577b-4ed2-b9fe-d7319c5d56a6\") " pod="openstack/nova-metadata-0" Dec 11 08:31:24 crc kubenswrapper[4860]: I1211 08:31:24.817691 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xhvtj\" (UniqueName: \"kubernetes.io/projected/96c72d04-577b-4ed2-b9fe-d7319c5d56a6-kube-api-access-xhvtj\") pod \"nova-metadata-0\" (UID: \"96c72d04-577b-4ed2-b9fe-d7319c5d56a6\") " pod="openstack/nova-metadata-0" Dec 11 08:31:24 crc kubenswrapper[4860]: I1211 08:31:24.818182 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/96c72d04-577b-4ed2-b9fe-d7319c5d56a6-logs\") pod \"nova-metadata-0\" (UID: \"96c72d04-577b-4ed2-b9fe-d7319c5d56a6\") " pod="openstack/nova-metadata-0" Dec 11 08:31:24 crc kubenswrapper[4860]: I1211 08:31:24.821876 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/96c72d04-577b-4ed2-b9fe-d7319c5d56a6-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"96c72d04-577b-4ed2-b9fe-d7319c5d56a6\") " pod="openstack/nova-metadata-0" Dec 11 08:31:24 crc kubenswrapper[4860]: I1211 08:31:24.821999 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96c72d04-577b-4ed2-b9fe-d7319c5d56a6-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"96c72d04-577b-4ed2-b9fe-d7319c5d56a6\") " pod="openstack/nova-metadata-0" Dec 11 08:31:24 crc kubenswrapper[4860]: I1211 08:31:24.822511 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/96c72d04-577b-4ed2-b9fe-d7319c5d56a6-config-data\") pod \"nova-metadata-0\" (UID: \"96c72d04-577b-4ed2-b9fe-d7319c5d56a6\") " pod="openstack/nova-metadata-0" Dec 11 08:31:24 crc kubenswrapper[4860]: I1211 08:31:24.836311 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xhvtj\" (UniqueName: \"kubernetes.io/projected/96c72d04-577b-4ed2-b9fe-d7319c5d56a6-kube-api-access-xhvtj\") pod \"nova-metadata-0\" (UID: \"96c72d04-577b-4ed2-b9fe-d7319c5d56a6\") " pod="openstack/nova-metadata-0" Dec 11 08:31:24 crc kubenswrapper[4860]: I1211 08:31:24.925049 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 11 08:31:25 crc kubenswrapper[4860]: I1211 08:31:25.377720 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 11 08:31:25 crc kubenswrapper[4860]: W1211 08:31:25.379249 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod96c72d04_577b_4ed2_b9fe_d7319c5d56a6.slice/crio-fba0421ef1552c6d4640b58c84903639d9e85a927178eb72056cf0c174a13d3a WatchSource:0}: Error finding container fba0421ef1552c6d4640b58c84903639d9e85a927178eb72056cf0c174a13d3a: Status 404 returned error can't find the container with id fba0421ef1552c6d4640b58c84903639d9e85a927178eb72056cf0c174a13d3a Dec 11 08:31:25 crc kubenswrapper[4860]: I1211 08:31:25.502431 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"96c72d04-577b-4ed2-b9fe-d7319c5d56a6","Type":"ContainerStarted","Data":"fba0421ef1552c6d4640b58c84903639d9e85a927178eb72056cf0c174a13d3a"} Dec 11 08:31:25 crc kubenswrapper[4860]: I1211 08:31:25.601030 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1747e5bf-b214-43a8-8276-8c2badd571ab" path="/var/lib/kubelet/pods/1747e5bf-b214-43a8-8276-8c2badd571ab/volumes" Dec 11 08:31:26 crc kubenswrapper[4860]: I1211 08:31:26.519898 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"96c72d04-577b-4ed2-b9fe-d7319c5d56a6","Type":"ContainerStarted","Data":"14f6f9a01ac6e9725c9b1bade90deaab2ab59447d0821c52bbfed989797caa75"} Dec 11 08:31:26 crc kubenswrapper[4860]: I1211 08:31:26.520291 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"96c72d04-577b-4ed2-b9fe-d7319c5d56a6","Type":"ContainerStarted","Data":"ffea859d11da0e998a8570da5a7edb29c1c59d7fd1d23e319c83bfbdfc654520"} Dec 11 08:31:26 crc kubenswrapper[4860]: I1211 08:31:26.545201 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.545178261 podStartE2EDuration="2.545178261s" podCreationTimestamp="2025-12-11 08:31:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:31:26.538979776 +0000 UTC m=+1219.267498851" watchObservedRunningTime="2025-12-11 08:31:26.545178261 +0000 UTC m=+1219.273697316" Dec 11 08:31:27 crc kubenswrapper[4860]: I1211 08:31:27.531894 4860 generic.go:334] "Generic (PLEG): container finished" podID="983cbde3-46fe-4643-b213-c74900941587" containerID="b29ada59900b63fcf4b78fcd929f9f91ae19df8e2b3a8506766956c7d67e078c" exitCode=0 Dec 11 08:31:27 crc kubenswrapper[4860]: I1211 08:31:27.532049 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-nvt7h" event={"ID":"983cbde3-46fe-4643-b213-c74900941587","Type":"ContainerDied","Data":"b29ada59900b63fcf4b78fcd929f9f91ae19df8e2b3a8506766956c7d67e078c"} Dec 11 08:31:28 crc kubenswrapper[4860]: I1211 08:31:28.016969 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 11 08:31:28 crc kubenswrapper[4860]: I1211 08:31:28.017158 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 11 08:31:28 crc kubenswrapper[4860]: I1211 08:31:28.017203 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 11 08:31:28 crc kubenswrapper[4860]: I1211 08:31:28.020841 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 11 08:31:28 crc kubenswrapper[4860]: I1211 08:31:28.020887 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 11 08:31:28 crc kubenswrapper[4860]: I1211 08:31:28.048382 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 11 08:31:28 crc kubenswrapper[4860]: I1211 08:31:28.219855 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-865f5d856f-wz6cz" Dec 11 08:31:28 crc kubenswrapper[4860]: I1211 08:31:28.326983 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bb4fc677f-mcn2v"] Dec 11 08:31:28 crc kubenswrapper[4860]: I1211 08:31:28.327546 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6bb4fc677f-mcn2v" podUID="8ad63b0c-4ee1-4a57-9116-a8fed2dae193" containerName="dnsmasq-dns" containerID="cri-o://e0a65b2d44cd87deeec4155c7f61b065474241e76538ddbd2f6cb854d27d25d3" gracePeriod=10 Dec 11 08:31:28 crc kubenswrapper[4860]: I1211 08:31:28.547764 4860 generic.go:334] "Generic (PLEG): container finished" podID="8ad63b0c-4ee1-4a57-9116-a8fed2dae193" containerID="e0a65b2d44cd87deeec4155c7f61b065474241e76538ddbd2f6cb854d27d25d3" exitCode=0 Dec 11 08:31:28 crc kubenswrapper[4860]: I1211 08:31:28.547830 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bb4fc677f-mcn2v" event={"ID":"8ad63b0c-4ee1-4a57-9116-a8fed2dae193","Type":"ContainerDied","Data":"e0a65b2d44cd87deeec4155c7f61b065474241e76538ddbd2f6cb854d27d25d3"} Dec 11 08:31:28 crc kubenswrapper[4860]: I1211 08:31:28.549733 4860 generic.go:334] "Generic (PLEG): container finished" podID="af3bd0b9-ab28-4e7b-8316-3206671fb97b" containerID="25fa35ea439bdf563989e6595788235e05f9c66929274bdfa3aa467f9c6dac01" exitCode=0 Dec 11 08:31:28 crc kubenswrapper[4860]: I1211 08:31:28.550326 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-bq4cn" event={"ID":"af3bd0b9-ab28-4e7b-8316-3206671fb97b","Type":"ContainerDied","Data":"25fa35ea439bdf563989e6595788235e05f9c66929274bdfa3aa467f9c6dac01"} Dec 11 08:31:28 crc kubenswrapper[4860]: E1211 08:31:28.584324 4860 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8ad63b0c_4ee1_4a57_9116_a8fed2dae193.slice/crio-conmon-e0a65b2d44cd87deeec4155c7f61b065474241e76538ddbd2f6cb854d27d25d3.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8ad63b0c_4ee1_4a57_9116_a8fed2dae193.slice/crio-e0a65b2d44cd87deeec4155c7f61b065474241e76538ddbd2f6cb854d27d25d3.scope\": RecentStats: unable to find data in memory cache]" Dec 11 08:31:28 crc kubenswrapper[4860]: I1211 08:31:28.616998 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 11 08:31:28 crc kubenswrapper[4860]: I1211 08:31:28.895185 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bb4fc677f-mcn2v" Dec 11 08:31:29 crc kubenswrapper[4860]: I1211 08:31:29.007462 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8ad63b0c-4ee1-4a57-9116-a8fed2dae193-ovsdbserver-nb\") pod \"8ad63b0c-4ee1-4a57-9116-a8fed2dae193\" (UID: \"8ad63b0c-4ee1-4a57-9116-a8fed2dae193\") " Dec 11 08:31:29 crc kubenswrapper[4860]: I1211 08:31:29.007539 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8ad63b0c-4ee1-4a57-9116-a8fed2dae193-config\") pod \"8ad63b0c-4ee1-4a57-9116-a8fed2dae193\" (UID: \"8ad63b0c-4ee1-4a57-9116-a8fed2dae193\") " Dec 11 08:31:29 crc kubenswrapper[4860]: I1211 08:31:29.007579 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4qtwf\" (UniqueName: \"kubernetes.io/projected/8ad63b0c-4ee1-4a57-9116-a8fed2dae193-kube-api-access-4qtwf\") pod \"8ad63b0c-4ee1-4a57-9116-a8fed2dae193\" (UID: \"8ad63b0c-4ee1-4a57-9116-a8fed2dae193\") " Dec 11 08:31:29 crc kubenswrapper[4860]: I1211 08:31:29.007631 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8ad63b0c-4ee1-4a57-9116-a8fed2dae193-ovsdbserver-sb\") pod \"8ad63b0c-4ee1-4a57-9116-a8fed2dae193\" (UID: \"8ad63b0c-4ee1-4a57-9116-a8fed2dae193\") " Dec 11 08:31:29 crc kubenswrapper[4860]: I1211 08:31:29.007802 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8ad63b0c-4ee1-4a57-9116-a8fed2dae193-dns-swift-storage-0\") pod \"8ad63b0c-4ee1-4a57-9116-a8fed2dae193\" (UID: \"8ad63b0c-4ee1-4a57-9116-a8fed2dae193\") " Dec 11 08:31:29 crc kubenswrapper[4860]: I1211 08:31:29.007853 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8ad63b0c-4ee1-4a57-9116-a8fed2dae193-dns-svc\") pod \"8ad63b0c-4ee1-4a57-9116-a8fed2dae193\" (UID: \"8ad63b0c-4ee1-4a57-9116-a8fed2dae193\") " Dec 11 08:31:29 crc kubenswrapper[4860]: I1211 08:31:29.011469 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-nvt7h" Dec 11 08:31:29 crc kubenswrapper[4860]: I1211 08:31:29.016854 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ad63b0c-4ee1-4a57-9116-a8fed2dae193-kube-api-access-4qtwf" (OuterVolumeSpecName: "kube-api-access-4qtwf") pod "8ad63b0c-4ee1-4a57-9116-a8fed2dae193" (UID: "8ad63b0c-4ee1-4a57-9116-a8fed2dae193"). InnerVolumeSpecName "kube-api-access-4qtwf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:31:29 crc kubenswrapper[4860]: I1211 08:31:29.070174 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8ad63b0c-4ee1-4a57-9116-a8fed2dae193-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "8ad63b0c-4ee1-4a57-9116-a8fed2dae193" (UID: "8ad63b0c-4ee1-4a57-9116-a8fed2dae193"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:31:29 crc kubenswrapper[4860]: I1211 08:31:29.075000 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8ad63b0c-4ee1-4a57-9116-a8fed2dae193-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "8ad63b0c-4ee1-4a57-9116-a8fed2dae193" (UID: "8ad63b0c-4ee1-4a57-9116-a8fed2dae193"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:31:29 crc kubenswrapper[4860]: I1211 08:31:29.093261 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8ad63b0c-4ee1-4a57-9116-a8fed2dae193-config" (OuterVolumeSpecName: "config") pod "8ad63b0c-4ee1-4a57-9116-a8fed2dae193" (UID: "8ad63b0c-4ee1-4a57-9116-a8fed2dae193"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:31:29 crc kubenswrapper[4860]: I1211 08:31:29.094518 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8ad63b0c-4ee1-4a57-9116-a8fed2dae193-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8ad63b0c-4ee1-4a57-9116-a8fed2dae193" (UID: "8ad63b0c-4ee1-4a57-9116-a8fed2dae193"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:31:29 crc kubenswrapper[4860]: I1211 08:31:29.099831 4860 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="59f08d16-b65e-4758-b17f-471dcb23781b" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.183:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 11 08:31:29 crc kubenswrapper[4860]: I1211 08:31:29.099880 4860 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="59f08d16-b65e-4758-b17f-471dcb23781b" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.183:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 11 08:31:29 crc kubenswrapper[4860]: I1211 08:31:29.104037 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8ad63b0c-4ee1-4a57-9116-a8fed2dae193-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "8ad63b0c-4ee1-4a57-9116-a8fed2dae193" (UID: "8ad63b0c-4ee1-4a57-9116-a8fed2dae193"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:31:29 crc kubenswrapper[4860]: I1211 08:31:29.109840 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/983cbde3-46fe-4643-b213-c74900941587-combined-ca-bundle\") pod \"983cbde3-46fe-4643-b213-c74900941587\" (UID: \"983cbde3-46fe-4643-b213-c74900941587\") " Dec 11 08:31:29 crc kubenswrapper[4860]: I1211 08:31:29.109970 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/983cbde3-46fe-4643-b213-c74900941587-scripts\") pod \"983cbde3-46fe-4643-b213-c74900941587\" (UID: \"983cbde3-46fe-4643-b213-c74900941587\") " Dec 11 08:31:29 crc kubenswrapper[4860]: I1211 08:31:29.110007 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pnq45\" (UniqueName: \"kubernetes.io/projected/983cbde3-46fe-4643-b213-c74900941587-kube-api-access-pnq45\") pod \"983cbde3-46fe-4643-b213-c74900941587\" (UID: \"983cbde3-46fe-4643-b213-c74900941587\") " Dec 11 08:31:29 crc kubenswrapper[4860]: I1211 08:31:29.110072 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/983cbde3-46fe-4643-b213-c74900941587-config-data\") pod \"983cbde3-46fe-4643-b213-c74900941587\" (UID: \"983cbde3-46fe-4643-b213-c74900941587\") " Dec 11 08:31:29 crc kubenswrapper[4860]: I1211 08:31:29.110797 4860 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8ad63b0c-4ee1-4a57-9116-a8fed2dae193-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 11 08:31:29 crc kubenswrapper[4860]: I1211 08:31:29.110822 4860 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8ad63b0c-4ee1-4a57-9116-a8fed2dae193-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 11 08:31:29 crc kubenswrapper[4860]: I1211 08:31:29.110835 4860 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8ad63b0c-4ee1-4a57-9116-a8fed2dae193-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 11 08:31:29 crc kubenswrapper[4860]: I1211 08:31:29.110846 4860 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8ad63b0c-4ee1-4a57-9116-a8fed2dae193-config\") on node \"crc\" DevicePath \"\"" Dec 11 08:31:29 crc kubenswrapper[4860]: I1211 08:31:29.110857 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4qtwf\" (UniqueName: \"kubernetes.io/projected/8ad63b0c-4ee1-4a57-9116-a8fed2dae193-kube-api-access-4qtwf\") on node \"crc\" DevicePath \"\"" Dec 11 08:31:29 crc kubenswrapper[4860]: I1211 08:31:29.110867 4860 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8ad63b0c-4ee1-4a57-9116-a8fed2dae193-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 11 08:31:29 crc kubenswrapper[4860]: I1211 08:31:29.113382 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/983cbde3-46fe-4643-b213-c74900941587-kube-api-access-pnq45" (OuterVolumeSpecName: "kube-api-access-pnq45") pod "983cbde3-46fe-4643-b213-c74900941587" (UID: "983cbde3-46fe-4643-b213-c74900941587"). InnerVolumeSpecName "kube-api-access-pnq45". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:31:29 crc kubenswrapper[4860]: I1211 08:31:29.114786 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/983cbde3-46fe-4643-b213-c74900941587-scripts" (OuterVolumeSpecName: "scripts") pod "983cbde3-46fe-4643-b213-c74900941587" (UID: "983cbde3-46fe-4643-b213-c74900941587"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:31:29 crc kubenswrapper[4860]: I1211 08:31:29.141845 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/983cbde3-46fe-4643-b213-c74900941587-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "983cbde3-46fe-4643-b213-c74900941587" (UID: "983cbde3-46fe-4643-b213-c74900941587"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:31:29 crc kubenswrapper[4860]: I1211 08:31:29.147894 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/983cbde3-46fe-4643-b213-c74900941587-config-data" (OuterVolumeSpecName: "config-data") pod "983cbde3-46fe-4643-b213-c74900941587" (UID: "983cbde3-46fe-4643-b213-c74900941587"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:31:29 crc kubenswrapper[4860]: I1211 08:31:29.212926 4860 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/983cbde3-46fe-4643-b213-c74900941587-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 08:31:29 crc kubenswrapper[4860]: I1211 08:31:29.212973 4860 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/983cbde3-46fe-4643-b213-c74900941587-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 08:31:29 crc kubenswrapper[4860]: I1211 08:31:29.212989 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pnq45\" (UniqueName: \"kubernetes.io/projected/983cbde3-46fe-4643-b213-c74900941587-kube-api-access-pnq45\") on node \"crc\" DevicePath \"\"" Dec 11 08:31:29 crc kubenswrapper[4860]: I1211 08:31:29.213002 4860 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/983cbde3-46fe-4643-b213-c74900941587-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 08:31:29 crc kubenswrapper[4860]: I1211 08:31:29.559679 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-nvt7h" event={"ID":"983cbde3-46fe-4643-b213-c74900941587","Type":"ContainerDied","Data":"9e6956fde9407ef3fe083d5ce4f3f47237f2eeb965f8471d7a4794756151253b"} Dec 11 08:31:29 crc kubenswrapper[4860]: I1211 08:31:29.559984 4860 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9e6956fde9407ef3fe083d5ce4f3f47237f2eeb965f8471d7a4794756151253b" Dec 11 08:31:29 crc kubenswrapper[4860]: I1211 08:31:29.560065 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-nvt7h" Dec 11 08:31:29 crc kubenswrapper[4860]: I1211 08:31:29.563978 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bb4fc677f-mcn2v" event={"ID":"8ad63b0c-4ee1-4a57-9116-a8fed2dae193","Type":"ContainerDied","Data":"441f8a1c9c0169f1a8f3ea5b7181e356c2b01e36ab18ac1e1249a121676b54ce"} Dec 11 08:31:29 crc kubenswrapper[4860]: I1211 08:31:29.564004 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bb4fc677f-mcn2v" Dec 11 08:31:29 crc kubenswrapper[4860]: I1211 08:31:29.564090 4860 scope.go:117] "RemoveContainer" containerID="e0a65b2d44cd87deeec4155c7f61b065474241e76538ddbd2f6cb854d27d25d3" Dec 11 08:31:29 crc kubenswrapper[4860]: I1211 08:31:29.630656 4860 scope.go:117] "RemoveContainer" containerID="98e328ef2d1f54bfc06f425726808d1fccb7c13163c8851db70e87e836ba8ec5" Dec 11 08:31:29 crc kubenswrapper[4860]: I1211 08:31:29.644511 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bb4fc677f-mcn2v"] Dec 11 08:31:29 crc kubenswrapper[4860]: I1211 08:31:29.663224 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6bb4fc677f-mcn2v"] Dec 11 08:31:29 crc kubenswrapper[4860]: I1211 08:31:29.766728 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 11 08:31:29 crc kubenswrapper[4860]: I1211 08:31:29.767139 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="59f08d16-b65e-4758-b17f-471dcb23781b" containerName="nova-api-log" containerID="cri-o://b91d2fecd9207ff2bc6ae98d9326d1848747386c2259481d5a8b51b88bd731e1" gracePeriod=30 Dec 11 08:31:29 crc kubenswrapper[4860]: I1211 08:31:29.767804 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="59f08d16-b65e-4758-b17f-471dcb23781b" containerName="nova-api-api" containerID="cri-o://513c5f1935162d62bb01b78f733a9f9eb73e25f7cb0e6739aafc410d1b82df65" gracePeriod=30 Dec 11 08:31:29 crc kubenswrapper[4860]: I1211 08:31:29.797539 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 11 08:31:29 crc kubenswrapper[4860]: I1211 08:31:29.840524 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 11 08:31:29 crc kubenswrapper[4860]: I1211 08:31:29.840895 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="96c72d04-577b-4ed2-b9fe-d7319c5d56a6" containerName="nova-metadata-log" containerID="cri-o://ffea859d11da0e998a8570da5a7edb29c1c59d7fd1d23e319c83bfbdfc654520" gracePeriod=30 Dec 11 08:31:29 crc kubenswrapper[4860]: I1211 08:31:29.841098 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="96c72d04-577b-4ed2-b9fe-d7319c5d56a6" containerName="nova-metadata-metadata" containerID="cri-o://14f6f9a01ac6e9725c9b1bade90deaab2ab59447d0821c52bbfed989797caa75" gracePeriod=30 Dec 11 08:31:29 crc kubenswrapper[4860]: I1211 08:31:29.933943 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 11 08:31:29 crc kubenswrapper[4860]: I1211 08:31:29.934002 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 11 08:31:30 crc kubenswrapper[4860]: I1211 08:31:30.083189 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-bq4cn" Dec 11 08:31:30 crc kubenswrapper[4860]: I1211 08:31:30.243882 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/af3bd0b9-ab28-4e7b-8316-3206671fb97b-scripts\") pod \"af3bd0b9-ab28-4e7b-8316-3206671fb97b\" (UID: \"af3bd0b9-ab28-4e7b-8316-3206671fb97b\") " Dec 11 08:31:30 crc kubenswrapper[4860]: I1211 08:31:30.243934 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af3bd0b9-ab28-4e7b-8316-3206671fb97b-config-data\") pod \"af3bd0b9-ab28-4e7b-8316-3206671fb97b\" (UID: \"af3bd0b9-ab28-4e7b-8316-3206671fb97b\") " Dec 11 08:31:30 crc kubenswrapper[4860]: I1211 08:31:30.244061 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af3bd0b9-ab28-4e7b-8316-3206671fb97b-combined-ca-bundle\") pod \"af3bd0b9-ab28-4e7b-8316-3206671fb97b\" (UID: \"af3bd0b9-ab28-4e7b-8316-3206671fb97b\") " Dec 11 08:31:30 crc kubenswrapper[4860]: I1211 08:31:30.244096 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4xcxb\" (UniqueName: \"kubernetes.io/projected/af3bd0b9-ab28-4e7b-8316-3206671fb97b-kube-api-access-4xcxb\") pod \"af3bd0b9-ab28-4e7b-8316-3206671fb97b\" (UID: \"af3bd0b9-ab28-4e7b-8316-3206671fb97b\") " Dec 11 08:31:30 crc kubenswrapper[4860]: I1211 08:31:30.264018 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af3bd0b9-ab28-4e7b-8316-3206671fb97b-kube-api-access-4xcxb" (OuterVolumeSpecName: "kube-api-access-4xcxb") pod "af3bd0b9-ab28-4e7b-8316-3206671fb97b" (UID: "af3bd0b9-ab28-4e7b-8316-3206671fb97b"). InnerVolumeSpecName "kube-api-access-4xcxb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:31:30 crc kubenswrapper[4860]: I1211 08:31:30.269828 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af3bd0b9-ab28-4e7b-8316-3206671fb97b-scripts" (OuterVolumeSpecName: "scripts") pod "af3bd0b9-ab28-4e7b-8316-3206671fb97b" (UID: "af3bd0b9-ab28-4e7b-8316-3206671fb97b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:31:30 crc kubenswrapper[4860]: I1211 08:31:30.347152 4860 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/af3bd0b9-ab28-4e7b-8316-3206671fb97b-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 08:31:30 crc kubenswrapper[4860]: I1211 08:31:30.347186 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4xcxb\" (UniqueName: \"kubernetes.io/projected/af3bd0b9-ab28-4e7b-8316-3206671fb97b-kube-api-access-4xcxb\") on node \"crc\" DevicePath \"\"" Dec 11 08:31:30 crc kubenswrapper[4860]: I1211 08:31:30.347282 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af3bd0b9-ab28-4e7b-8316-3206671fb97b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "af3bd0b9-ab28-4e7b-8316-3206671fb97b" (UID: "af3bd0b9-ab28-4e7b-8316-3206671fb97b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:31:30 crc kubenswrapper[4860]: I1211 08:31:30.385904 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af3bd0b9-ab28-4e7b-8316-3206671fb97b-config-data" (OuterVolumeSpecName: "config-data") pod "af3bd0b9-ab28-4e7b-8316-3206671fb97b" (UID: "af3bd0b9-ab28-4e7b-8316-3206671fb97b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:31:30 crc kubenswrapper[4860]: I1211 08:31:30.457041 4860 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af3bd0b9-ab28-4e7b-8316-3206671fb97b-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 08:31:30 crc kubenswrapper[4860]: I1211 08:31:30.457079 4860 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af3bd0b9-ab28-4e7b-8316-3206671fb97b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 08:31:30 crc kubenswrapper[4860]: I1211 08:31:30.576068 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-bq4cn" event={"ID":"af3bd0b9-ab28-4e7b-8316-3206671fb97b","Type":"ContainerDied","Data":"dbca1f64ace3a04a5f620953a8495bdfb3d8047e555166b4f1aaed502a742007"} Dec 11 08:31:30 crc kubenswrapper[4860]: I1211 08:31:30.576112 4860 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dbca1f64ace3a04a5f620953a8495bdfb3d8047e555166b4f1aaed502a742007" Dec 11 08:31:30 crc kubenswrapper[4860]: I1211 08:31:30.576217 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-bq4cn" Dec 11 08:31:30 crc kubenswrapper[4860]: I1211 08:31:30.599486 4860 generic.go:334] "Generic (PLEG): container finished" podID="96c72d04-577b-4ed2-b9fe-d7319c5d56a6" containerID="14f6f9a01ac6e9725c9b1bade90deaab2ab59447d0821c52bbfed989797caa75" exitCode=0 Dec 11 08:31:30 crc kubenswrapper[4860]: I1211 08:31:30.599512 4860 generic.go:334] "Generic (PLEG): container finished" podID="96c72d04-577b-4ed2-b9fe-d7319c5d56a6" containerID="ffea859d11da0e998a8570da5a7edb29c1c59d7fd1d23e319c83bfbdfc654520" exitCode=143 Dec 11 08:31:30 crc kubenswrapper[4860]: I1211 08:31:30.599573 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"96c72d04-577b-4ed2-b9fe-d7319c5d56a6","Type":"ContainerDied","Data":"14f6f9a01ac6e9725c9b1bade90deaab2ab59447d0821c52bbfed989797caa75"} Dec 11 08:31:30 crc kubenswrapper[4860]: I1211 08:31:30.599604 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"96c72d04-577b-4ed2-b9fe-d7319c5d56a6","Type":"ContainerDied","Data":"ffea859d11da0e998a8570da5a7edb29c1c59d7fd1d23e319c83bfbdfc654520"} Dec 11 08:31:30 crc kubenswrapper[4860]: I1211 08:31:30.599618 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"96c72d04-577b-4ed2-b9fe-d7319c5d56a6","Type":"ContainerDied","Data":"fba0421ef1552c6d4640b58c84903639d9e85a927178eb72056cf0c174a13d3a"} Dec 11 08:31:30 crc kubenswrapper[4860]: I1211 08:31:30.599627 4860 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fba0421ef1552c6d4640b58c84903639d9e85a927178eb72056cf0c174a13d3a" Dec 11 08:31:30 crc kubenswrapper[4860]: I1211 08:31:30.604875 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 11 08:31:30 crc kubenswrapper[4860]: I1211 08:31:30.606571 4860 generic.go:334] "Generic (PLEG): container finished" podID="59f08d16-b65e-4758-b17f-471dcb23781b" containerID="b91d2fecd9207ff2bc6ae98d9326d1848747386c2259481d5a8b51b88bd731e1" exitCode=143 Dec 11 08:31:30 crc kubenswrapper[4860]: I1211 08:31:30.606757 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="2ee186b9-2dbe-4492-89ff-4962cd4c942a" containerName="nova-scheduler-scheduler" containerID="cri-o://7506728391e138c7cb3f179f605e565bb04063c28ba7fa01c16af44f73c687a2" gracePeriod=30 Dec 11 08:31:30 crc kubenswrapper[4860]: I1211 08:31:30.607015 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"59f08d16-b65e-4758-b17f-471dcb23781b","Type":"ContainerDied","Data":"b91d2fecd9207ff2bc6ae98d9326d1848747386c2259481d5a8b51b88bd731e1"} Dec 11 08:31:30 crc kubenswrapper[4860]: I1211 08:31:30.680601 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 11 08:31:30 crc kubenswrapper[4860]: E1211 08:31:30.681632 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ad63b0c-4ee1-4a57-9116-a8fed2dae193" containerName="dnsmasq-dns" Dec 11 08:31:30 crc kubenswrapper[4860]: I1211 08:31:30.681668 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ad63b0c-4ee1-4a57-9116-a8fed2dae193" containerName="dnsmasq-dns" Dec 11 08:31:30 crc kubenswrapper[4860]: E1211 08:31:30.681693 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96c72d04-577b-4ed2-b9fe-d7319c5d56a6" containerName="nova-metadata-metadata" Dec 11 08:31:30 crc kubenswrapper[4860]: I1211 08:31:30.681699 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="96c72d04-577b-4ed2-b9fe-d7319c5d56a6" containerName="nova-metadata-metadata" Dec 11 08:31:30 crc kubenswrapper[4860]: E1211 08:31:30.681711 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ad63b0c-4ee1-4a57-9116-a8fed2dae193" containerName="init" Dec 11 08:31:30 crc kubenswrapper[4860]: I1211 08:31:30.681719 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ad63b0c-4ee1-4a57-9116-a8fed2dae193" containerName="init" Dec 11 08:31:30 crc kubenswrapper[4860]: E1211 08:31:30.681731 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="983cbde3-46fe-4643-b213-c74900941587" containerName="nova-manage" Dec 11 08:31:30 crc kubenswrapper[4860]: I1211 08:31:30.681737 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="983cbde3-46fe-4643-b213-c74900941587" containerName="nova-manage" Dec 11 08:31:30 crc kubenswrapper[4860]: E1211 08:31:30.681758 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96c72d04-577b-4ed2-b9fe-d7319c5d56a6" containerName="nova-metadata-log" Dec 11 08:31:30 crc kubenswrapper[4860]: I1211 08:31:30.681765 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="96c72d04-577b-4ed2-b9fe-d7319c5d56a6" containerName="nova-metadata-log" Dec 11 08:31:30 crc kubenswrapper[4860]: E1211 08:31:30.681777 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af3bd0b9-ab28-4e7b-8316-3206671fb97b" containerName="nova-cell1-conductor-db-sync" Dec 11 08:31:30 crc kubenswrapper[4860]: I1211 08:31:30.681783 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="af3bd0b9-ab28-4e7b-8316-3206671fb97b" containerName="nova-cell1-conductor-db-sync" Dec 11 08:31:30 crc kubenswrapper[4860]: I1211 08:31:30.681958 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="96c72d04-577b-4ed2-b9fe-d7319c5d56a6" containerName="nova-metadata-log" Dec 11 08:31:30 crc kubenswrapper[4860]: I1211 08:31:30.681973 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="af3bd0b9-ab28-4e7b-8316-3206671fb97b" containerName="nova-cell1-conductor-db-sync" Dec 11 08:31:30 crc kubenswrapper[4860]: I1211 08:31:30.681985 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="96c72d04-577b-4ed2-b9fe-d7319c5d56a6" containerName="nova-metadata-metadata" Dec 11 08:31:30 crc kubenswrapper[4860]: I1211 08:31:30.681999 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="983cbde3-46fe-4643-b213-c74900941587" containerName="nova-manage" Dec 11 08:31:30 crc kubenswrapper[4860]: I1211 08:31:30.682009 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ad63b0c-4ee1-4a57-9116-a8fed2dae193" containerName="dnsmasq-dns" Dec 11 08:31:30 crc kubenswrapper[4860]: I1211 08:31:30.682768 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 11 08:31:30 crc kubenswrapper[4860]: I1211 08:31:30.686253 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 11 08:31:30 crc kubenswrapper[4860]: I1211 08:31:30.706532 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 11 08:31:30 crc kubenswrapper[4860]: I1211 08:31:30.763598 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xhvtj\" (UniqueName: \"kubernetes.io/projected/96c72d04-577b-4ed2-b9fe-d7319c5d56a6-kube-api-access-xhvtj\") pod \"96c72d04-577b-4ed2-b9fe-d7319c5d56a6\" (UID: \"96c72d04-577b-4ed2-b9fe-d7319c5d56a6\") " Dec 11 08:31:30 crc kubenswrapper[4860]: I1211 08:31:30.763671 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96c72d04-577b-4ed2-b9fe-d7319c5d56a6-combined-ca-bundle\") pod \"96c72d04-577b-4ed2-b9fe-d7319c5d56a6\" (UID: \"96c72d04-577b-4ed2-b9fe-d7319c5d56a6\") " Dec 11 08:31:30 crc kubenswrapper[4860]: I1211 08:31:30.763765 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/96c72d04-577b-4ed2-b9fe-d7319c5d56a6-logs\") pod \"96c72d04-577b-4ed2-b9fe-d7319c5d56a6\" (UID: \"96c72d04-577b-4ed2-b9fe-d7319c5d56a6\") " Dec 11 08:31:30 crc kubenswrapper[4860]: I1211 08:31:30.763829 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/96c72d04-577b-4ed2-b9fe-d7319c5d56a6-nova-metadata-tls-certs\") pod \"96c72d04-577b-4ed2-b9fe-d7319c5d56a6\" (UID: \"96c72d04-577b-4ed2-b9fe-d7319c5d56a6\") " Dec 11 08:31:30 crc kubenswrapper[4860]: I1211 08:31:30.763891 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/96c72d04-577b-4ed2-b9fe-d7319c5d56a6-config-data\") pod \"96c72d04-577b-4ed2-b9fe-d7319c5d56a6\" (UID: \"96c72d04-577b-4ed2-b9fe-d7319c5d56a6\") " Dec 11 08:31:30 crc kubenswrapper[4860]: I1211 08:31:30.764283 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/773668a5-7449-4ced-b8d0-509faab0eabb-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"773668a5-7449-4ced-b8d0-509faab0eabb\") " pod="openstack/nova-cell1-conductor-0" Dec 11 08:31:30 crc kubenswrapper[4860]: I1211 08:31:30.764390 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/773668a5-7449-4ced-b8d0-509faab0eabb-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"773668a5-7449-4ced-b8d0-509faab0eabb\") " pod="openstack/nova-cell1-conductor-0" Dec 11 08:31:30 crc kubenswrapper[4860]: I1211 08:31:30.764445 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-grdjp\" (UniqueName: \"kubernetes.io/projected/773668a5-7449-4ced-b8d0-509faab0eabb-kube-api-access-grdjp\") pod \"nova-cell1-conductor-0\" (UID: \"773668a5-7449-4ced-b8d0-509faab0eabb\") " pod="openstack/nova-cell1-conductor-0" Dec 11 08:31:30 crc kubenswrapper[4860]: I1211 08:31:30.764872 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/96c72d04-577b-4ed2-b9fe-d7319c5d56a6-logs" (OuterVolumeSpecName: "logs") pod "96c72d04-577b-4ed2-b9fe-d7319c5d56a6" (UID: "96c72d04-577b-4ed2-b9fe-d7319c5d56a6"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:31:30 crc kubenswrapper[4860]: I1211 08:31:30.768557 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96c72d04-577b-4ed2-b9fe-d7319c5d56a6-kube-api-access-xhvtj" (OuterVolumeSpecName: "kube-api-access-xhvtj") pod "96c72d04-577b-4ed2-b9fe-d7319c5d56a6" (UID: "96c72d04-577b-4ed2-b9fe-d7319c5d56a6"). InnerVolumeSpecName "kube-api-access-xhvtj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:31:30 crc kubenswrapper[4860]: I1211 08:31:30.809026 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96c72d04-577b-4ed2-b9fe-d7319c5d56a6-config-data" (OuterVolumeSpecName: "config-data") pod "96c72d04-577b-4ed2-b9fe-d7319c5d56a6" (UID: "96c72d04-577b-4ed2-b9fe-d7319c5d56a6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:31:30 crc kubenswrapper[4860]: I1211 08:31:30.814417 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96c72d04-577b-4ed2-b9fe-d7319c5d56a6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "96c72d04-577b-4ed2-b9fe-d7319c5d56a6" (UID: "96c72d04-577b-4ed2-b9fe-d7319c5d56a6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:31:30 crc kubenswrapper[4860]: I1211 08:31:30.841781 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96c72d04-577b-4ed2-b9fe-d7319c5d56a6-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "96c72d04-577b-4ed2-b9fe-d7319c5d56a6" (UID: "96c72d04-577b-4ed2-b9fe-d7319c5d56a6"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:31:30 crc kubenswrapper[4860]: I1211 08:31:30.866328 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-grdjp\" (UniqueName: \"kubernetes.io/projected/773668a5-7449-4ced-b8d0-509faab0eabb-kube-api-access-grdjp\") pod \"nova-cell1-conductor-0\" (UID: \"773668a5-7449-4ced-b8d0-509faab0eabb\") " pod="openstack/nova-cell1-conductor-0" Dec 11 08:31:30 crc kubenswrapper[4860]: I1211 08:31:30.866448 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/773668a5-7449-4ced-b8d0-509faab0eabb-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"773668a5-7449-4ced-b8d0-509faab0eabb\") " pod="openstack/nova-cell1-conductor-0" Dec 11 08:31:30 crc kubenswrapper[4860]: I1211 08:31:30.866574 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/773668a5-7449-4ced-b8d0-509faab0eabb-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"773668a5-7449-4ced-b8d0-509faab0eabb\") " pod="openstack/nova-cell1-conductor-0" Dec 11 08:31:30 crc kubenswrapper[4860]: I1211 08:31:30.866678 4860 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/96c72d04-577b-4ed2-b9fe-d7319c5d56a6-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 11 08:31:30 crc kubenswrapper[4860]: I1211 08:31:30.866696 4860 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/96c72d04-577b-4ed2-b9fe-d7319c5d56a6-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 08:31:30 crc kubenswrapper[4860]: I1211 08:31:30.866711 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xhvtj\" (UniqueName: \"kubernetes.io/projected/96c72d04-577b-4ed2-b9fe-d7319c5d56a6-kube-api-access-xhvtj\") on node \"crc\" DevicePath \"\"" Dec 11 08:31:30 crc kubenswrapper[4860]: I1211 08:31:30.866725 4860 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96c72d04-577b-4ed2-b9fe-d7319c5d56a6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 08:31:30 crc kubenswrapper[4860]: I1211 08:31:30.866735 4860 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/96c72d04-577b-4ed2-b9fe-d7319c5d56a6-logs\") on node \"crc\" DevicePath \"\"" Dec 11 08:31:30 crc kubenswrapper[4860]: I1211 08:31:30.871785 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/773668a5-7449-4ced-b8d0-509faab0eabb-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"773668a5-7449-4ced-b8d0-509faab0eabb\") " pod="openstack/nova-cell1-conductor-0" Dec 11 08:31:30 crc kubenswrapper[4860]: I1211 08:31:30.871915 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/773668a5-7449-4ced-b8d0-509faab0eabb-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"773668a5-7449-4ced-b8d0-509faab0eabb\") " pod="openstack/nova-cell1-conductor-0" Dec 11 08:31:30 crc kubenswrapper[4860]: I1211 08:31:30.891174 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-grdjp\" (UniqueName: \"kubernetes.io/projected/773668a5-7449-4ced-b8d0-509faab0eabb-kube-api-access-grdjp\") pod \"nova-cell1-conductor-0\" (UID: \"773668a5-7449-4ced-b8d0-509faab0eabb\") " pod="openstack/nova-cell1-conductor-0" Dec 11 08:31:31 crc kubenswrapper[4860]: I1211 08:31:31.011774 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 11 08:31:31 crc kubenswrapper[4860]: I1211 08:31:31.513941 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 11 08:31:31 crc kubenswrapper[4860]: I1211 08:31:31.599231 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8ad63b0c-4ee1-4a57-9116-a8fed2dae193" path="/var/lib/kubelet/pods/8ad63b0c-4ee1-4a57-9116-a8fed2dae193/volumes" Dec 11 08:31:31 crc kubenswrapper[4860]: I1211 08:31:31.624884 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 11 08:31:31 crc kubenswrapper[4860]: I1211 08:31:31.630511 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"773668a5-7449-4ced-b8d0-509faab0eabb","Type":"ContainerStarted","Data":"4395c5a8074bd4cc91ba931d5824a851ac23e2b63555ead71b70c695d3c80790"} Dec 11 08:31:31 crc kubenswrapper[4860]: I1211 08:31:31.701717 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 11 08:31:31 crc kubenswrapper[4860]: I1211 08:31:31.718750 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 11 08:31:31 crc kubenswrapper[4860]: I1211 08:31:31.734613 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 11 08:31:31 crc kubenswrapper[4860]: I1211 08:31:31.736300 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 11 08:31:31 crc kubenswrapper[4860]: I1211 08:31:31.739849 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 11 08:31:31 crc kubenswrapper[4860]: I1211 08:31:31.740079 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 11 08:31:31 crc kubenswrapper[4860]: I1211 08:31:31.746396 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 11 08:31:31 crc kubenswrapper[4860]: I1211 08:31:31.890675 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a82c6492-9c83-4860-b478-7172462f1be8-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"a82c6492-9c83-4860-b478-7172462f1be8\") " pod="openstack/nova-metadata-0" Dec 11 08:31:31 crc kubenswrapper[4860]: I1211 08:31:31.890771 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a82c6492-9c83-4860-b478-7172462f1be8-logs\") pod \"nova-metadata-0\" (UID: \"a82c6492-9c83-4860-b478-7172462f1be8\") " pod="openstack/nova-metadata-0" Dec 11 08:31:31 crc kubenswrapper[4860]: I1211 08:31:31.890818 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a82c6492-9c83-4860-b478-7172462f1be8-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"a82c6492-9c83-4860-b478-7172462f1be8\") " pod="openstack/nova-metadata-0" Dec 11 08:31:31 crc kubenswrapper[4860]: I1211 08:31:31.891017 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qlgk6\" (UniqueName: \"kubernetes.io/projected/a82c6492-9c83-4860-b478-7172462f1be8-kube-api-access-qlgk6\") pod \"nova-metadata-0\" (UID: \"a82c6492-9c83-4860-b478-7172462f1be8\") " pod="openstack/nova-metadata-0" Dec 11 08:31:31 crc kubenswrapper[4860]: I1211 08:31:31.891108 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a82c6492-9c83-4860-b478-7172462f1be8-config-data\") pod \"nova-metadata-0\" (UID: \"a82c6492-9c83-4860-b478-7172462f1be8\") " pod="openstack/nova-metadata-0" Dec 11 08:31:31 crc kubenswrapper[4860]: I1211 08:31:31.992790 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a82c6492-9c83-4860-b478-7172462f1be8-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"a82c6492-9c83-4860-b478-7172462f1be8\") " pod="openstack/nova-metadata-0" Dec 11 08:31:31 crc kubenswrapper[4860]: I1211 08:31:31.992860 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a82c6492-9c83-4860-b478-7172462f1be8-logs\") pod \"nova-metadata-0\" (UID: \"a82c6492-9c83-4860-b478-7172462f1be8\") " pod="openstack/nova-metadata-0" Dec 11 08:31:31 crc kubenswrapper[4860]: I1211 08:31:31.992885 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a82c6492-9c83-4860-b478-7172462f1be8-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"a82c6492-9c83-4860-b478-7172462f1be8\") " pod="openstack/nova-metadata-0" Dec 11 08:31:31 crc kubenswrapper[4860]: I1211 08:31:31.992910 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qlgk6\" (UniqueName: \"kubernetes.io/projected/a82c6492-9c83-4860-b478-7172462f1be8-kube-api-access-qlgk6\") pod \"nova-metadata-0\" (UID: \"a82c6492-9c83-4860-b478-7172462f1be8\") " pod="openstack/nova-metadata-0" Dec 11 08:31:31 crc kubenswrapper[4860]: I1211 08:31:31.992950 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a82c6492-9c83-4860-b478-7172462f1be8-config-data\") pod \"nova-metadata-0\" (UID: \"a82c6492-9c83-4860-b478-7172462f1be8\") " pod="openstack/nova-metadata-0" Dec 11 08:31:31 crc kubenswrapper[4860]: I1211 08:31:31.993429 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a82c6492-9c83-4860-b478-7172462f1be8-logs\") pod \"nova-metadata-0\" (UID: \"a82c6492-9c83-4860-b478-7172462f1be8\") " pod="openstack/nova-metadata-0" Dec 11 08:31:31 crc kubenswrapper[4860]: I1211 08:31:31.997760 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a82c6492-9c83-4860-b478-7172462f1be8-config-data\") pod \"nova-metadata-0\" (UID: \"a82c6492-9c83-4860-b478-7172462f1be8\") " pod="openstack/nova-metadata-0" Dec 11 08:31:31 crc kubenswrapper[4860]: I1211 08:31:31.997880 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a82c6492-9c83-4860-b478-7172462f1be8-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"a82c6492-9c83-4860-b478-7172462f1be8\") " pod="openstack/nova-metadata-0" Dec 11 08:31:32 crc kubenswrapper[4860]: I1211 08:31:32.006424 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a82c6492-9c83-4860-b478-7172462f1be8-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"a82c6492-9c83-4860-b478-7172462f1be8\") " pod="openstack/nova-metadata-0" Dec 11 08:31:32 crc kubenswrapper[4860]: I1211 08:31:32.023723 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qlgk6\" (UniqueName: \"kubernetes.io/projected/a82c6492-9c83-4860-b478-7172462f1be8-kube-api-access-qlgk6\") pod \"nova-metadata-0\" (UID: \"a82c6492-9c83-4860-b478-7172462f1be8\") " pod="openstack/nova-metadata-0" Dec 11 08:31:32 crc kubenswrapper[4860]: I1211 08:31:32.067511 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 11 08:31:32 crc kubenswrapper[4860]: I1211 08:31:32.557148 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 11 08:31:32 crc kubenswrapper[4860]: I1211 08:31:32.637382 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"773668a5-7449-4ced-b8d0-509faab0eabb","Type":"ContainerStarted","Data":"7624f84a8309b53e3e7fda3654f1714b9b4cccf57090c87b5553a0a508312fc1"} Dec 11 08:31:32 crc kubenswrapper[4860]: I1211 08:31:32.639288 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a82c6492-9c83-4860-b478-7172462f1be8","Type":"ContainerStarted","Data":"2f797655173402203be6f6db324ed212192a675354cd2505404e7cff0528dbe2"} Dec 11 08:31:32 crc kubenswrapper[4860]: I1211 08:31:32.642438 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Dec 11 08:31:32 crc kubenswrapper[4860]: I1211 08:31:32.665490 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.665453148 podStartE2EDuration="2.665453148s" podCreationTimestamp="2025-12-11 08:31:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:31:32.662304919 +0000 UTC m=+1225.390824004" watchObservedRunningTime="2025-12-11 08:31:32.665453148 +0000 UTC m=+1225.393972203" Dec 11 08:31:33 crc kubenswrapper[4860]: E1211 08:31:33.022863 4860 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7506728391e138c7cb3f179f605e565bb04063c28ba7fa01c16af44f73c687a2" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 11 08:31:33 crc kubenswrapper[4860]: E1211 08:31:33.025226 4860 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7506728391e138c7cb3f179f605e565bb04063c28ba7fa01c16af44f73c687a2" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 11 08:31:33 crc kubenswrapper[4860]: E1211 08:31:33.029194 4860 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7506728391e138c7cb3f179f605e565bb04063c28ba7fa01c16af44f73c687a2" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 11 08:31:33 crc kubenswrapper[4860]: E1211 08:31:33.029347 4860 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="2ee186b9-2dbe-4492-89ff-4962cd4c942a" containerName="nova-scheduler-scheduler" Dec 11 08:31:33 crc kubenswrapper[4860]: I1211 08:31:33.592518 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96c72d04-577b-4ed2-b9fe-d7319c5d56a6" path="/var/lib/kubelet/pods/96c72d04-577b-4ed2-b9fe-d7319c5d56a6/volumes" Dec 11 08:31:33 crc kubenswrapper[4860]: I1211 08:31:33.651688 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a82c6492-9c83-4860-b478-7172462f1be8","Type":"ContainerStarted","Data":"94618ba17096594576975a12f9f1e29c8a01bd0b94daedd85c747df101f11437"} Dec 11 08:31:33 crc kubenswrapper[4860]: I1211 08:31:33.651787 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a82c6492-9c83-4860-b478-7172462f1be8","Type":"ContainerStarted","Data":"642aa739b6ebe2ab3d86ad681b34f4b367dd2aa72e868bd68c334e4d3da125e2"} Dec 11 08:31:33 crc kubenswrapper[4860]: I1211 08:31:33.673805 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.67378186 podStartE2EDuration="2.67378186s" podCreationTimestamp="2025-12-11 08:31:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:31:33.667860262 +0000 UTC m=+1226.396379337" watchObservedRunningTime="2025-12-11 08:31:33.67378186 +0000 UTC m=+1226.402300915" Dec 11 08:31:34 crc kubenswrapper[4860]: I1211 08:31:34.662791 4860 generic.go:334] "Generic (PLEG): container finished" podID="2ee186b9-2dbe-4492-89ff-4962cd4c942a" containerID="7506728391e138c7cb3f179f605e565bb04063c28ba7fa01c16af44f73c687a2" exitCode=0 Dec 11 08:31:34 crc kubenswrapper[4860]: I1211 08:31:34.662958 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"2ee186b9-2dbe-4492-89ff-4962cd4c942a","Type":"ContainerDied","Data":"7506728391e138c7cb3f179f605e565bb04063c28ba7fa01c16af44f73c687a2"} Dec 11 08:31:34 crc kubenswrapper[4860]: I1211 08:31:34.919457 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 11 08:31:35 crc kubenswrapper[4860]: I1211 08:31:35.063746 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ggrt5\" (UniqueName: \"kubernetes.io/projected/2ee186b9-2dbe-4492-89ff-4962cd4c942a-kube-api-access-ggrt5\") pod \"2ee186b9-2dbe-4492-89ff-4962cd4c942a\" (UID: \"2ee186b9-2dbe-4492-89ff-4962cd4c942a\") " Dec 11 08:31:35 crc kubenswrapper[4860]: I1211 08:31:35.063784 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ee186b9-2dbe-4492-89ff-4962cd4c942a-combined-ca-bundle\") pod \"2ee186b9-2dbe-4492-89ff-4962cd4c942a\" (UID: \"2ee186b9-2dbe-4492-89ff-4962cd4c942a\") " Dec 11 08:31:35 crc kubenswrapper[4860]: I1211 08:31:35.063906 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ee186b9-2dbe-4492-89ff-4962cd4c942a-config-data\") pod \"2ee186b9-2dbe-4492-89ff-4962cd4c942a\" (UID: \"2ee186b9-2dbe-4492-89ff-4962cd4c942a\") " Dec 11 08:31:35 crc kubenswrapper[4860]: I1211 08:31:35.069243 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ee186b9-2dbe-4492-89ff-4962cd4c942a-kube-api-access-ggrt5" (OuterVolumeSpecName: "kube-api-access-ggrt5") pod "2ee186b9-2dbe-4492-89ff-4962cd4c942a" (UID: "2ee186b9-2dbe-4492-89ff-4962cd4c942a"). InnerVolumeSpecName "kube-api-access-ggrt5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:31:35 crc kubenswrapper[4860]: E1211 08:31:35.093916 4860 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2ee186b9-2dbe-4492-89ff-4962cd4c942a-combined-ca-bundle podName:2ee186b9-2dbe-4492-89ff-4962cd4c942a nodeName:}" failed. No retries permitted until 2025-12-11 08:31:35.593879808 +0000 UTC m=+1228.322398883 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "combined-ca-bundle" (UniqueName: "kubernetes.io/secret/2ee186b9-2dbe-4492-89ff-4962cd4c942a-combined-ca-bundle") pod "2ee186b9-2dbe-4492-89ff-4962cd4c942a" (UID: "2ee186b9-2dbe-4492-89ff-4962cd4c942a") : error deleting /var/lib/kubelet/pods/2ee186b9-2dbe-4492-89ff-4962cd4c942a/volume-subpaths: remove /var/lib/kubelet/pods/2ee186b9-2dbe-4492-89ff-4962cd4c942a/volume-subpaths: no such file or directory Dec 11 08:31:35 crc kubenswrapper[4860]: I1211 08:31:35.096524 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ee186b9-2dbe-4492-89ff-4962cd4c942a-config-data" (OuterVolumeSpecName: "config-data") pod "2ee186b9-2dbe-4492-89ff-4962cd4c942a" (UID: "2ee186b9-2dbe-4492-89ff-4962cd4c942a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:31:35 crc kubenswrapper[4860]: I1211 08:31:35.167203 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ggrt5\" (UniqueName: \"kubernetes.io/projected/2ee186b9-2dbe-4492-89ff-4962cd4c942a-kube-api-access-ggrt5\") on node \"crc\" DevicePath \"\"" Dec 11 08:31:35 crc kubenswrapper[4860]: I1211 08:31:35.167265 4860 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ee186b9-2dbe-4492-89ff-4962cd4c942a-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 08:31:35 crc kubenswrapper[4860]: I1211 08:31:35.664069 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 11 08:31:35 crc kubenswrapper[4860]: I1211 08:31:35.674511 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 11 08:31:35 crc kubenswrapper[4860]: I1211 08:31:35.674537 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"2ee186b9-2dbe-4492-89ff-4962cd4c942a","Type":"ContainerDied","Data":"0ae686847bf69d44e17110b43f7a37203a7e7c485760132db74d8a3daad8b9e3"} Dec 11 08:31:35 crc kubenswrapper[4860]: I1211 08:31:35.674599 4860 scope.go:117] "RemoveContainer" containerID="7506728391e138c7cb3f179f605e565bb04063c28ba7fa01c16af44f73c687a2" Dec 11 08:31:35 crc kubenswrapper[4860]: I1211 08:31:35.677061 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ee186b9-2dbe-4492-89ff-4962cd4c942a-combined-ca-bundle\") pod \"2ee186b9-2dbe-4492-89ff-4962cd4c942a\" (UID: \"2ee186b9-2dbe-4492-89ff-4962cd4c942a\") " Dec 11 08:31:35 crc kubenswrapper[4860]: I1211 08:31:35.677245 4860 generic.go:334] "Generic (PLEG): container finished" podID="59f08d16-b65e-4758-b17f-471dcb23781b" containerID="513c5f1935162d62bb01b78f733a9f9eb73e25f7cb0e6739aafc410d1b82df65" exitCode=0 Dec 11 08:31:35 crc kubenswrapper[4860]: I1211 08:31:35.677319 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 11 08:31:35 crc kubenswrapper[4860]: I1211 08:31:35.677349 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"59f08d16-b65e-4758-b17f-471dcb23781b","Type":"ContainerDied","Data":"513c5f1935162d62bb01b78f733a9f9eb73e25f7cb0e6739aafc410d1b82df65"} Dec 11 08:31:35 crc kubenswrapper[4860]: I1211 08:31:35.677535 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"59f08d16-b65e-4758-b17f-471dcb23781b","Type":"ContainerDied","Data":"9ff5149f18d31af8a19b8040918f69c098dd1646825505c9219be72eee0b7fee"} Dec 11 08:31:35 crc kubenswrapper[4860]: I1211 08:31:35.687932 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ee186b9-2dbe-4492-89ff-4962cd4c942a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2ee186b9-2dbe-4492-89ff-4962cd4c942a" (UID: "2ee186b9-2dbe-4492-89ff-4962cd4c942a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:31:35 crc kubenswrapper[4860]: I1211 08:31:35.720534 4860 scope.go:117] "RemoveContainer" containerID="513c5f1935162d62bb01b78f733a9f9eb73e25f7cb0e6739aafc410d1b82df65" Dec 11 08:31:35 crc kubenswrapper[4860]: I1211 08:31:35.774206 4860 scope.go:117] "RemoveContainer" containerID="b91d2fecd9207ff2bc6ae98d9326d1848747386c2259481d5a8b51b88bd731e1" Dec 11 08:31:35 crc kubenswrapper[4860]: I1211 08:31:35.779517 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-77fg9\" (UniqueName: \"kubernetes.io/projected/59f08d16-b65e-4758-b17f-471dcb23781b-kube-api-access-77fg9\") pod \"59f08d16-b65e-4758-b17f-471dcb23781b\" (UID: \"59f08d16-b65e-4758-b17f-471dcb23781b\") " Dec 11 08:31:35 crc kubenswrapper[4860]: I1211 08:31:35.779571 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59f08d16-b65e-4758-b17f-471dcb23781b-combined-ca-bundle\") pod \"59f08d16-b65e-4758-b17f-471dcb23781b\" (UID: \"59f08d16-b65e-4758-b17f-471dcb23781b\") " Dec 11 08:31:35 crc kubenswrapper[4860]: I1211 08:31:35.779761 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/59f08d16-b65e-4758-b17f-471dcb23781b-logs\") pod \"59f08d16-b65e-4758-b17f-471dcb23781b\" (UID: \"59f08d16-b65e-4758-b17f-471dcb23781b\") " Dec 11 08:31:35 crc kubenswrapper[4860]: I1211 08:31:35.779830 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59f08d16-b65e-4758-b17f-471dcb23781b-config-data\") pod \"59f08d16-b65e-4758-b17f-471dcb23781b\" (UID: \"59f08d16-b65e-4758-b17f-471dcb23781b\") " Dec 11 08:31:35 crc kubenswrapper[4860]: I1211 08:31:35.780282 4860 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ee186b9-2dbe-4492-89ff-4962cd4c942a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 08:31:35 crc kubenswrapper[4860]: I1211 08:31:35.780674 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/59f08d16-b65e-4758-b17f-471dcb23781b-logs" (OuterVolumeSpecName: "logs") pod "59f08d16-b65e-4758-b17f-471dcb23781b" (UID: "59f08d16-b65e-4758-b17f-471dcb23781b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:31:35 crc kubenswrapper[4860]: I1211 08:31:35.793819 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/59f08d16-b65e-4758-b17f-471dcb23781b-kube-api-access-77fg9" (OuterVolumeSpecName: "kube-api-access-77fg9") pod "59f08d16-b65e-4758-b17f-471dcb23781b" (UID: "59f08d16-b65e-4758-b17f-471dcb23781b"). InnerVolumeSpecName "kube-api-access-77fg9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:31:35 crc kubenswrapper[4860]: I1211 08:31:35.808888 4860 scope.go:117] "RemoveContainer" containerID="513c5f1935162d62bb01b78f733a9f9eb73e25f7cb0e6739aafc410d1b82df65" Dec 11 08:31:35 crc kubenswrapper[4860]: E1211 08:31:35.809835 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"513c5f1935162d62bb01b78f733a9f9eb73e25f7cb0e6739aafc410d1b82df65\": container with ID starting with 513c5f1935162d62bb01b78f733a9f9eb73e25f7cb0e6739aafc410d1b82df65 not found: ID does not exist" containerID="513c5f1935162d62bb01b78f733a9f9eb73e25f7cb0e6739aafc410d1b82df65" Dec 11 08:31:35 crc kubenswrapper[4860]: I1211 08:31:35.810315 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"513c5f1935162d62bb01b78f733a9f9eb73e25f7cb0e6739aafc410d1b82df65"} err="failed to get container status \"513c5f1935162d62bb01b78f733a9f9eb73e25f7cb0e6739aafc410d1b82df65\": rpc error: code = NotFound desc = could not find container \"513c5f1935162d62bb01b78f733a9f9eb73e25f7cb0e6739aafc410d1b82df65\": container with ID starting with 513c5f1935162d62bb01b78f733a9f9eb73e25f7cb0e6739aafc410d1b82df65 not found: ID does not exist" Dec 11 08:31:35 crc kubenswrapper[4860]: I1211 08:31:35.810480 4860 scope.go:117] "RemoveContainer" containerID="b91d2fecd9207ff2bc6ae98d9326d1848747386c2259481d5a8b51b88bd731e1" Dec 11 08:31:35 crc kubenswrapper[4860]: I1211 08:31:35.810818 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/59f08d16-b65e-4758-b17f-471dcb23781b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "59f08d16-b65e-4758-b17f-471dcb23781b" (UID: "59f08d16-b65e-4758-b17f-471dcb23781b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:31:35 crc kubenswrapper[4860]: E1211 08:31:35.814014 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b91d2fecd9207ff2bc6ae98d9326d1848747386c2259481d5a8b51b88bd731e1\": container with ID starting with b91d2fecd9207ff2bc6ae98d9326d1848747386c2259481d5a8b51b88bd731e1 not found: ID does not exist" containerID="b91d2fecd9207ff2bc6ae98d9326d1848747386c2259481d5a8b51b88bd731e1" Dec 11 08:31:35 crc kubenswrapper[4860]: I1211 08:31:35.814165 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b91d2fecd9207ff2bc6ae98d9326d1848747386c2259481d5a8b51b88bd731e1"} err="failed to get container status \"b91d2fecd9207ff2bc6ae98d9326d1848747386c2259481d5a8b51b88bd731e1\": rpc error: code = NotFound desc = could not find container \"b91d2fecd9207ff2bc6ae98d9326d1848747386c2259481d5a8b51b88bd731e1\": container with ID starting with b91d2fecd9207ff2bc6ae98d9326d1848747386c2259481d5a8b51b88bd731e1 not found: ID does not exist" Dec 11 08:31:35 crc kubenswrapper[4860]: I1211 08:31:35.816510 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/59f08d16-b65e-4758-b17f-471dcb23781b-config-data" (OuterVolumeSpecName: "config-data") pod "59f08d16-b65e-4758-b17f-471dcb23781b" (UID: "59f08d16-b65e-4758-b17f-471dcb23781b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:31:35 crc kubenswrapper[4860]: I1211 08:31:35.882261 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-77fg9\" (UniqueName: \"kubernetes.io/projected/59f08d16-b65e-4758-b17f-471dcb23781b-kube-api-access-77fg9\") on node \"crc\" DevicePath \"\"" Dec 11 08:31:35 crc kubenswrapper[4860]: I1211 08:31:35.882499 4860 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59f08d16-b65e-4758-b17f-471dcb23781b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 08:31:35 crc kubenswrapper[4860]: I1211 08:31:35.882559 4860 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/59f08d16-b65e-4758-b17f-471dcb23781b-logs\") on node \"crc\" DevicePath \"\"" Dec 11 08:31:35 crc kubenswrapper[4860]: I1211 08:31:35.882621 4860 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59f08d16-b65e-4758-b17f-471dcb23781b-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 08:31:36 crc kubenswrapper[4860]: I1211 08:31:36.015103 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 11 08:31:36 crc kubenswrapper[4860]: I1211 08:31:36.024220 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 11 08:31:36 crc kubenswrapper[4860]: I1211 08:31:36.066928 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Dec 11 08:31:36 crc kubenswrapper[4860]: I1211 08:31:36.074208 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 11 08:31:36 crc kubenswrapper[4860]: I1211 08:31:36.096074 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 11 08:31:36 crc kubenswrapper[4860]: I1211 08:31:36.111271 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 11 08:31:36 crc kubenswrapper[4860]: E1211 08:31:36.112525 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59f08d16-b65e-4758-b17f-471dcb23781b" containerName="nova-api-api" Dec 11 08:31:36 crc kubenswrapper[4860]: I1211 08:31:36.112546 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="59f08d16-b65e-4758-b17f-471dcb23781b" containerName="nova-api-api" Dec 11 08:31:36 crc kubenswrapper[4860]: E1211 08:31:36.112575 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ee186b9-2dbe-4492-89ff-4962cd4c942a" containerName="nova-scheduler-scheduler" Dec 11 08:31:36 crc kubenswrapper[4860]: I1211 08:31:36.112581 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ee186b9-2dbe-4492-89ff-4962cd4c942a" containerName="nova-scheduler-scheduler" Dec 11 08:31:36 crc kubenswrapper[4860]: E1211 08:31:36.112626 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59f08d16-b65e-4758-b17f-471dcb23781b" containerName="nova-api-log" Dec 11 08:31:36 crc kubenswrapper[4860]: I1211 08:31:36.112633 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="59f08d16-b65e-4758-b17f-471dcb23781b" containerName="nova-api-log" Dec 11 08:31:36 crc kubenswrapper[4860]: I1211 08:31:36.112901 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ee186b9-2dbe-4492-89ff-4962cd4c942a" containerName="nova-scheduler-scheduler" Dec 11 08:31:36 crc kubenswrapper[4860]: I1211 08:31:36.112920 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="59f08d16-b65e-4758-b17f-471dcb23781b" containerName="nova-api-log" Dec 11 08:31:36 crc kubenswrapper[4860]: I1211 08:31:36.112929 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="59f08d16-b65e-4758-b17f-471dcb23781b" containerName="nova-api-api" Dec 11 08:31:36 crc kubenswrapper[4860]: I1211 08:31:36.113696 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 11 08:31:36 crc kubenswrapper[4860]: I1211 08:31:36.117211 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 11 08:31:36 crc kubenswrapper[4860]: I1211 08:31:36.119177 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 11 08:31:36 crc kubenswrapper[4860]: I1211 08:31:36.138088 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 11 08:31:36 crc kubenswrapper[4860]: I1211 08:31:36.139940 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 11 08:31:36 crc kubenswrapper[4860]: I1211 08:31:36.142057 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 11 08:31:36 crc kubenswrapper[4860]: I1211 08:31:36.148776 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 11 08:31:36 crc kubenswrapper[4860]: I1211 08:31:36.187916 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jrhbt\" (UniqueName: \"kubernetes.io/projected/66451d3a-291f-494e-bbe7-f68e32fa2305-kube-api-access-jrhbt\") pod \"nova-scheduler-0\" (UID: \"66451d3a-291f-494e-bbe7-f68e32fa2305\") " pod="openstack/nova-scheduler-0" Dec 11 08:31:36 crc kubenswrapper[4860]: I1211 08:31:36.188051 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66451d3a-291f-494e-bbe7-f68e32fa2305-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"66451d3a-291f-494e-bbe7-f68e32fa2305\") " pod="openstack/nova-scheduler-0" Dec 11 08:31:36 crc kubenswrapper[4860]: I1211 08:31:36.188080 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/66451d3a-291f-494e-bbe7-f68e32fa2305-config-data\") pod \"nova-scheduler-0\" (UID: \"66451d3a-291f-494e-bbe7-f68e32fa2305\") " pod="openstack/nova-scheduler-0" Dec 11 08:31:36 crc kubenswrapper[4860]: I1211 08:31:36.290075 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jrhbt\" (UniqueName: \"kubernetes.io/projected/66451d3a-291f-494e-bbe7-f68e32fa2305-kube-api-access-jrhbt\") pod \"nova-scheduler-0\" (UID: \"66451d3a-291f-494e-bbe7-f68e32fa2305\") " pod="openstack/nova-scheduler-0" Dec 11 08:31:36 crc kubenswrapper[4860]: I1211 08:31:36.290217 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2714dc33-cc1b-4fed-af2c-9ab58397d1c0-logs\") pod \"nova-api-0\" (UID: \"2714dc33-cc1b-4fed-af2c-9ab58397d1c0\") " pod="openstack/nova-api-0" Dec 11 08:31:36 crc kubenswrapper[4860]: I1211 08:31:36.290248 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2714dc33-cc1b-4fed-af2c-9ab58397d1c0-config-data\") pod \"nova-api-0\" (UID: \"2714dc33-cc1b-4fed-af2c-9ab58397d1c0\") " pod="openstack/nova-api-0" Dec 11 08:31:36 crc kubenswrapper[4860]: I1211 08:31:36.290366 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66451d3a-291f-494e-bbe7-f68e32fa2305-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"66451d3a-291f-494e-bbe7-f68e32fa2305\") " pod="openstack/nova-scheduler-0" Dec 11 08:31:36 crc kubenswrapper[4860]: I1211 08:31:36.290400 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/66451d3a-291f-494e-bbe7-f68e32fa2305-config-data\") pod \"nova-scheduler-0\" (UID: \"66451d3a-291f-494e-bbe7-f68e32fa2305\") " pod="openstack/nova-scheduler-0" Dec 11 08:31:36 crc kubenswrapper[4860]: I1211 08:31:36.290435 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2714dc33-cc1b-4fed-af2c-9ab58397d1c0-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"2714dc33-cc1b-4fed-af2c-9ab58397d1c0\") " pod="openstack/nova-api-0" Dec 11 08:31:36 crc kubenswrapper[4860]: I1211 08:31:36.290489 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xjkdc\" (UniqueName: \"kubernetes.io/projected/2714dc33-cc1b-4fed-af2c-9ab58397d1c0-kube-api-access-xjkdc\") pod \"nova-api-0\" (UID: \"2714dc33-cc1b-4fed-af2c-9ab58397d1c0\") " pod="openstack/nova-api-0" Dec 11 08:31:36 crc kubenswrapper[4860]: I1211 08:31:36.295699 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/66451d3a-291f-494e-bbe7-f68e32fa2305-config-data\") pod \"nova-scheduler-0\" (UID: \"66451d3a-291f-494e-bbe7-f68e32fa2305\") " pod="openstack/nova-scheduler-0" Dec 11 08:31:36 crc kubenswrapper[4860]: I1211 08:31:36.296534 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66451d3a-291f-494e-bbe7-f68e32fa2305-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"66451d3a-291f-494e-bbe7-f68e32fa2305\") " pod="openstack/nova-scheduler-0" Dec 11 08:31:36 crc kubenswrapper[4860]: I1211 08:31:36.315760 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jrhbt\" (UniqueName: \"kubernetes.io/projected/66451d3a-291f-494e-bbe7-f68e32fa2305-kube-api-access-jrhbt\") pod \"nova-scheduler-0\" (UID: \"66451d3a-291f-494e-bbe7-f68e32fa2305\") " pod="openstack/nova-scheduler-0" Dec 11 08:31:36 crc kubenswrapper[4860]: I1211 08:31:36.392524 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2714dc33-cc1b-4fed-af2c-9ab58397d1c0-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"2714dc33-cc1b-4fed-af2c-9ab58397d1c0\") " pod="openstack/nova-api-0" Dec 11 08:31:36 crc kubenswrapper[4860]: I1211 08:31:36.392602 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xjkdc\" (UniqueName: \"kubernetes.io/projected/2714dc33-cc1b-4fed-af2c-9ab58397d1c0-kube-api-access-xjkdc\") pod \"nova-api-0\" (UID: \"2714dc33-cc1b-4fed-af2c-9ab58397d1c0\") " pod="openstack/nova-api-0" Dec 11 08:31:36 crc kubenswrapper[4860]: I1211 08:31:36.392771 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2714dc33-cc1b-4fed-af2c-9ab58397d1c0-logs\") pod \"nova-api-0\" (UID: \"2714dc33-cc1b-4fed-af2c-9ab58397d1c0\") " pod="openstack/nova-api-0" Dec 11 08:31:36 crc kubenswrapper[4860]: I1211 08:31:36.392796 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2714dc33-cc1b-4fed-af2c-9ab58397d1c0-config-data\") pod \"nova-api-0\" (UID: \"2714dc33-cc1b-4fed-af2c-9ab58397d1c0\") " pod="openstack/nova-api-0" Dec 11 08:31:36 crc kubenswrapper[4860]: I1211 08:31:36.393703 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2714dc33-cc1b-4fed-af2c-9ab58397d1c0-logs\") pod \"nova-api-0\" (UID: \"2714dc33-cc1b-4fed-af2c-9ab58397d1c0\") " pod="openstack/nova-api-0" Dec 11 08:31:36 crc kubenswrapper[4860]: I1211 08:31:36.396403 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2714dc33-cc1b-4fed-af2c-9ab58397d1c0-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"2714dc33-cc1b-4fed-af2c-9ab58397d1c0\") " pod="openstack/nova-api-0" Dec 11 08:31:36 crc kubenswrapper[4860]: I1211 08:31:36.396955 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2714dc33-cc1b-4fed-af2c-9ab58397d1c0-config-data\") pod \"nova-api-0\" (UID: \"2714dc33-cc1b-4fed-af2c-9ab58397d1c0\") " pod="openstack/nova-api-0" Dec 11 08:31:36 crc kubenswrapper[4860]: I1211 08:31:36.412562 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xjkdc\" (UniqueName: \"kubernetes.io/projected/2714dc33-cc1b-4fed-af2c-9ab58397d1c0-kube-api-access-xjkdc\") pod \"nova-api-0\" (UID: \"2714dc33-cc1b-4fed-af2c-9ab58397d1c0\") " pod="openstack/nova-api-0" Dec 11 08:31:36 crc kubenswrapper[4860]: I1211 08:31:36.436288 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 11 08:31:36 crc kubenswrapper[4860]: I1211 08:31:36.458140 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 11 08:31:36 crc kubenswrapper[4860]: I1211 08:31:36.896103 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 11 08:31:36 crc kubenswrapper[4860]: W1211 08:31:36.897384 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod66451d3a_291f_494e_bbe7_f68e32fa2305.slice/crio-e3424c913919879c2dc6fc9aa31e826369c3887b5e8ecc72d7aac58dc14366e0 WatchSource:0}: Error finding container e3424c913919879c2dc6fc9aa31e826369c3887b5e8ecc72d7aac58dc14366e0: Status 404 returned error can't find the container with id e3424c913919879c2dc6fc9aa31e826369c3887b5e8ecc72d7aac58dc14366e0 Dec 11 08:31:37 crc kubenswrapper[4860]: I1211 08:31:37.013164 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 11 08:31:37 crc kubenswrapper[4860]: W1211 08:31:37.014594 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2714dc33_cc1b_4fed_af2c_9ab58397d1c0.slice/crio-ccf35f6708555af04c9ee1ad842ee9584065fe2cc2491f5483920f9b0be404f0 WatchSource:0}: Error finding container ccf35f6708555af04c9ee1ad842ee9584065fe2cc2491f5483920f9b0be404f0: Status 404 returned error can't find the container with id ccf35f6708555af04c9ee1ad842ee9584065fe2cc2491f5483920f9b0be404f0 Dec 11 08:31:37 crc kubenswrapper[4860]: I1211 08:31:37.068155 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 11 08:31:37 crc kubenswrapper[4860]: I1211 08:31:37.068212 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 11 08:31:37 crc kubenswrapper[4860]: I1211 08:31:37.596194 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2ee186b9-2dbe-4492-89ff-4962cd4c942a" path="/var/lib/kubelet/pods/2ee186b9-2dbe-4492-89ff-4962cd4c942a/volumes" Dec 11 08:31:37 crc kubenswrapper[4860]: I1211 08:31:37.597805 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="59f08d16-b65e-4758-b17f-471dcb23781b" path="/var/lib/kubelet/pods/59f08d16-b65e-4758-b17f-471dcb23781b/volumes" Dec 11 08:31:37 crc kubenswrapper[4860]: I1211 08:31:37.604258 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 11 08:31:37 crc kubenswrapper[4860]: I1211 08:31:37.708932 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2714dc33-cc1b-4fed-af2c-9ab58397d1c0","Type":"ContainerStarted","Data":"ec7f65b71fd7e5c584de43621051867f16fc801d9f4efd6258a6b70b740910ca"} Dec 11 08:31:37 crc kubenswrapper[4860]: I1211 08:31:37.709253 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2714dc33-cc1b-4fed-af2c-9ab58397d1c0","Type":"ContainerStarted","Data":"7268c87e4750580d441f2217171d227fd7ee59a77faf3c385af1d37c4272934c"} Dec 11 08:31:37 crc kubenswrapper[4860]: I1211 08:31:37.709367 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2714dc33-cc1b-4fed-af2c-9ab58397d1c0","Type":"ContainerStarted","Data":"ccf35f6708555af04c9ee1ad842ee9584065fe2cc2491f5483920f9b0be404f0"} Dec 11 08:31:37 crc kubenswrapper[4860]: I1211 08:31:37.712186 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"66451d3a-291f-494e-bbe7-f68e32fa2305","Type":"ContainerStarted","Data":"520ae17c7ce420448d27850929ee5f21437fe7d22b66f4d7cc435bbe85127584"} Dec 11 08:31:37 crc kubenswrapper[4860]: I1211 08:31:37.712212 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"66451d3a-291f-494e-bbe7-f68e32fa2305","Type":"ContainerStarted","Data":"e3424c913919879c2dc6fc9aa31e826369c3887b5e8ecc72d7aac58dc14366e0"} Dec 11 08:31:37 crc kubenswrapper[4860]: I1211 08:31:37.756218 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=1.756190665 podStartE2EDuration="1.756190665s" podCreationTimestamp="2025-12-11 08:31:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:31:37.740256681 +0000 UTC m=+1230.468775736" watchObservedRunningTime="2025-12-11 08:31:37.756190665 +0000 UTC m=+1230.484709720" Dec 11 08:31:37 crc kubenswrapper[4860]: I1211 08:31:37.762656 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=1.7626146679999999 podStartE2EDuration="1.762614668s" podCreationTimestamp="2025-12-11 08:31:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:31:37.754984907 +0000 UTC m=+1230.483503962" watchObservedRunningTime="2025-12-11 08:31:37.762614668 +0000 UTC m=+1230.491133723" Dec 11 08:31:38 crc kubenswrapper[4860]: I1211 08:31:38.796748 4860 patch_prober.go:28] interesting pod/machine-config-daemon-99qgp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 08:31:38 crc kubenswrapper[4860]: I1211 08:31:38.796827 4860 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 08:31:41 crc kubenswrapper[4860]: I1211 08:31:41.437514 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 11 08:31:42 crc kubenswrapper[4860]: I1211 08:31:42.068759 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 11 08:31:42 crc kubenswrapper[4860]: I1211 08:31:42.069200 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 11 08:31:43 crc kubenswrapper[4860]: I1211 08:31:43.083930 4860 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="a82c6492-9c83-4860-b478-7172462f1be8" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.191:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 11 08:31:43 crc kubenswrapper[4860]: I1211 08:31:43.083930 4860 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="a82c6492-9c83-4860-b478-7172462f1be8" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.191:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 11 08:31:46 crc kubenswrapper[4860]: I1211 08:31:46.436944 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 11 08:31:46 crc kubenswrapper[4860]: I1211 08:31:46.459329 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 11 08:31:46 crc kubenswrapper[4860]: I1211 08:31:46.459385 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 11 08:31:46 crc kubenswrapper[4860]: I1211 08:31:46.472720 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 11 08:31:46 crc kubenswrapper[4860]: I1211 08:31:46.834206 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 11 08:31:47 crc kubenswrapper[4860]: I1211 08:31:47.543232 4860 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="2714dc33-cc1b-4fed-af2c-9ab58397d1c0" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.193:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 11 08:31:47 crc kubenswrapper[4860]: I1211 08:31:47.544061 4860 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="2714dc33-cc1b-4fed-af2c-9ab58397d1c0" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.193:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 11 08:31:52 crc kubenswrapper[4860]: I1211 08:31:52.074464 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 11 08:31:52 crc kubenswrapper[4860]: I1211 08:31:52.076413 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 11 08:31:52 crc kubenswrapper[4860]: I1211 08:31:52.083082 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 11 08:31:52 crc kubenswrapper[4860]: I1211 08:31:52.083418 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 11 08:31:53 crc kubenswrapper[4860]: I1211 08:31:53.861021 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 11 08:31:53 crc kubenswrapper[4860]: I1211 08:31:53.892434 4860 generic.go:334] "Generic (PLEG): container finished" podID="3954f800-6e02-4387-bd5f-c9dd575ce3c8" containerID="b861a25799833453fb1b060105a06855f4c458ede12f385f939e4a216002e9bd" exitCode=137 Dec 11 08:31:53 crc kubenswrapper[4860]: I1211 08:31:53.892495 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 11 08:31:53 crc kubenswrapper[4860]: I1211 08:31:53.892520 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"3954f800-6e02-4387-bd5f-c9dd575ce3c8","Type":"ContainerDied","Data":"b861a25799833453fb1b060105a06855f4c458ede12f385f939e4a216002e9bd"} Dec 11 08:31:53 crc kubenswrapper[4860]: I1211 08:31:53.892573 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"3954f800-6e02-4387-bd5f-c9dd575ce3c8","Type":"ContainerDied","Data":"03d1b78bda444d6d3b54c9908434f2a3d90042a8510c403e9827c13cd23fd807"} Dec 11 08:31:53 crc kubenswrapper[4860]: I1211 08:31:53.892592 4860 scope.go:117] "RemoveContainer" containerID="b861a25799833453fb1b060105a06855f4c458ede12f385f939e4a216002e9bd" Dec 11 08:31:53 crc kubenswrapper[4860]: I1211 08:31:53.918723 4860 scope.go:117] "RemoveContainer" containerID="b861a25799833453fb1b060105a06855f4c458ede12f385f939e4a216002e9bd" Dec 11 08:31:53 crc kubenswrapper[4860]: E1211 08:31:53.919401 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b861a25799833453fb1b060105a06855f4c458ede12f385f939e4a216002e9bd\": container with ID starting with b861a25799833453fb1b060105a06855f4c458ede12f385f939e4a216002e9bd not found: ID does not exist" containerID="b861a25799833453fb1b060105a06855f4c458ede12f385f939e4a216002e9bd" Dec 11 08:31:53 crc kubenswrapper[4860]: I1211 08:31:53.919553 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b861a25799833453fb1b060105a06855f4c458ede12f385f939e4a216002e9bd"} err="failed to get container status \"b861a25799833453fb1b060105a06855f4c458ede12f385f939e4a216002e9bd\": rpc error: code = NotFound desc = could not find container \"b861a25799833453fb1b060105a06855f4c458ede12f385f939e4a216002e9bd\": container with ID starting with b861a25799833453fb1b060105a06855f4c458ede12f385f939e4a216002e9bd not found: ID does not exist" Dec 11 08:31:53 crc kubenswrapper[4860]: I1211 08:31:53.983517 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4t22\" (UniqueName: \"kubernetes.io/projected/3954f800-6e02-4387-bd5f-c9dd575ce3c8-kube-api-access-d4t22\") pod \"3954f800-6e02-4387-bd5f-c9dd575ce3c8\" (UID: \"3954f800-6e02-4387-bd5f-c9dd575ce3c8\") " Dec 11 08:31:53 crc kubenswrapper[4860]: I1211 08:31:53.983617 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3954f800-6e02-4387-bd5f-c9dd575ce3c8-config-data\") pod \"3954f800-6e02-4387-bd5f-c9dd575ce3c8\" (UID: \"3954f800-6e02-4387-bd5f-c9dd575ce3c8\") " Dec 11 08:31:53 crc kubenswrapper[4860]: I1211 08:31:53.983839 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3954f800-6e02-4387-bd5f-c9dd575ce3c8-combined-ca-bundle\") pod \"3954f800-6e02-4387-bd5f-c9dd575ce3c8\" (UID: \"3954f800-6e02-4387-bd5f-c9dd575ce3c8\") " Dec 11 08:31:53 crc kubenswrapper[4860]: I1211 08:31:53.991955 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3954f800-6e02-4387-bd5f-c9dd575ce3c8-kube-api-access-d4t22" (OuterVolumeSpecName: "kube-api-access-d4t22") pod "3954f800-6e02-4387-bd5f-c9dd575ce3c8" (UID: "3954f800-6e02-4387-bd5f-c9dd575ce3c8"). InnerVolumeSpecName "kube-api-access-d4t22". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:31:54 crc kubenswrapper[4860]: I1211 08:31:54.014314 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3954f800-6e02-4387-bd5f-c9dd575ce3c8-config-data" (OuterVolumeSpecName: "config-data") pod "3954f800-6e02-4387-bd5f-c9dd575ce3c8" (UID: "3954f800-6e02-4387-bd5f-c9dd575ce3c8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:31:54 crc kubenswrapper[4860]: I1211 08:31:54.015843 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3954f800-6e02-4387-bd5f-c9dd575ce3c8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3954f800-6e02-4387-bd5f-c9dd575ce3c8" (UID: "3954f800-6e02-4387-bd5f-c9dd575ce3c8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:31:54 crc kubenswrapper[4860]: I1211 08:31:54.086420 4860 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3954f800-6e02-4387-bd5f-c9dd575ce3c8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 08:31:54 crc kubenswrapper[4860]: I1211 08:31:54.086459 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4t22\" (UniqueName: \"kubernetes.io/projected/3954f800-6e02-4387-bd5f-c9dd575ce3c8-kube-api-access-d4t22\") on node \"crc\" DevicePath \"\"" Dec 11 08:31:54 crc kubenswrapper[4860]: I1211 08:31:54.086473 4860 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3954f800-6e02-4387-bd5f-c9dd575ce3c8-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 08:31:54 crc kubenswrapper[4860]: I1211 08:31:54.258388 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 11 08:31:54 crc kubenswrapper[4860]: I1211 08:31:54.271999 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 11 08:31:54 crc kubenswrapper[4860]: I1211 08:31:54.287040 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 11 08:31:54 crc kubenswrapper[4860]: E1211 08:31:54.287636 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3954f800-6e02-4387-bd5f-c9dd575ce3c8" containerName="nova-cell1-novncproxy-novncproxy" Dec 11 08:31:54 crc kubenswrapper[4860]: I1211 08:31:54.287674 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="3954f800-6e02-4387-bd5f-c9dd575ce3c8" containerName="nova-cell1-novncproxy-novncproxy" Dec 11 08:31:54 crc kubenswrapper[4860]: I1211 08:31:54.287929 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="3954f800-6e02-4387-bd5f-c9dd575ce3c8" containerName="nova-cell1-novncproxy-novncproxy" Dec 11 08:31:54 crc kubenswrapper[4860]: I1211 08:31:54.290228 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 11 08:31:54 crc kubenswrapper[4860]: I1211 08:31:54.293099 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 11 08:31:54 crc kubenswrapper[4860]: I1211 08:31:54.293315 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Dec 11 08:31:54 crc kubenswrapper[4860]: I1211 08:31:54.293497 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Dec 11 08:31:54 crc kubenswrapper[4860]: I1211 08:31:54.299908 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 11 08:31:54 crc kubenswrapper[4860]: I1211 08:31:54.392303 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4725b54-31da-4b85-a4f7-200b1d6a3069-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"f4725b54-31da-4b85-a4f7-200b1d6a3069\") " pod="openstack/nova-cell1-novncproxy-0" Dec 11 08:31:54 crc kubenswrapper[4860]: I1211 08:31:54.392466 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/f4725b54-31da-4b85-a4f7-200b1d6a3069-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"f4725b54-31da-4b85-a4f7-200b1d6a3069\") " pod="openstack/nova-cell1-novncproxy-0" Dec 11 08:31:54 crc kubenswrapper[4860]: I1211 08:31:54.392516 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r2tst\" (UniqueName: \"kubernetes.io/projected/f4725b54-31da-4b85-a4f7-200b1d6a3069-kube-api-access-r2tst\") pod \"nova-cell1-novncproxy-0\" (UID: \"f4725b54-31da-4b85-a4f7-200b1d6a3069\") " pod="openstack/nova-cell1-novncproxy-0" Dec 11 08:31:54 crc kubenswrapper[4860]: I1211 08:31:54.392551 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4725b54-31da-4b85-a4f7-200b1d6a3069-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"f4725b54-31da-4b85-a4f7-200b1d6a3069\") " pod="openstack/nova-cell1-novncproxy-0" Dec 11 08:31:54 crc kubenswrapper[4860]: I1211 08:31:54.392634 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/f4725b54-31da-4b85-a4f7-200b1d6a3069-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"f4725b54-31da-4b85-a4f7-200b1d6a3069\") " pod="openstack/nova-cell1-novncproxy-0" Dec 11 08:31:54 crc kubenswrapper[4860]: I1211 08:31:54.494368 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/f4725b54-31da-4b85-a4f7-200b1d6a3069-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"f4725b54-31da-4b85-a4f7-200b1d6a3069\") " pod="openstack/nova-cell1-novncproxy-0" Dec 11 08:31:54 crc kubenswrapper[4860]: I1211 08:31:54.494452 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r2tst\" (UniqueName: \"kubernetes.io/projected/f4725b54-31da-4b85-a4f7-200b1d6a3069-kube-api-access-r2tst\") pod \"nova-cell1-novncproxy-0\" (UID: \"f4725b54-31da-4b85-a4f7-200b1d6a3069\") " pod="openstack/nova-cell1-novncproxy-0" Dec 11 08:31:54 crc kubenswrapper[4860]: I1211 08:31:54.494516 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4725b54-31da-4b85-a4f7-200b1d6a3069-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"f4725b54-31da-4b85-a4f7-200b1d6a3069\") " pod="openstack/nova-cell1-novncproxy-0" Dec 11 08:31:54 crc kubenswrapper[4860]: I1211 08:31:54.494576 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/f4725b54-31da-4b85-a4f7-200b1d6a3069-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"f4725b54-31da-4b85-a4f7-200b1d6a3069\") " pod="openstack/nova-cell1-novncproxy-0" Dec 11 08:31:54 crc kubenswrapper[4860]: I1211 08:31:54.494764 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4725b54-31da-4b85-a4f7-200b1d6a3069-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"f4725b54-31da-4b85-a4f7-200b1d6a3069\") " pod="openstack/nova-cell1-novncproxy-0" Dec 11 08:31:54 crc kubenswrapper[4860]: I1211 08:31:54.498859 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/f4725b54-31da-4b85-a4f7-200b1d6a3069-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"f4725b54-31da-4b85-a4f7-200b1d6a3069\") " pod="openstack/nova-cell1-novncproxy-0" Dec 11 08:31:54 crc kubenswrapper[4860]: I1211 08:31:54.498909 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/f4725b54-31da-4b85-a4f7-200b1d6a3069-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"f4725b54-31da-4b85-a4f7-200b1d6a3069\") " pod="openstack/nova-cell1-novncproxy-0" Dec 11 08:31:54 crc kubenswrapper[4860]: I1211 08:31:54.499298 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4725b54-31da-4b85-a4f7-200b1d6a3069-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"f4725b54-31da-4b85-a4f7-200b1d6a3069\") " pod="openstack/nova-cell1-novncproxy-0" Dec 11 08:31:54 crc kubenswrapper[4860]: I1211 08:31:54.499719 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4725b54-31da-4b85-a4f7-200b1d6a3069-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"f4725b54-31da-4b85-a4f7-200b1d6a3069\") " pod="openstack/nova-cell1-novncproxy-0" Dec 11 08:31:54 crc kubenswrapper[4860]: I1211 08:31:54.511800 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r2tst\" (UniqueName: \"kubernetes.io/projected/f4725b54-31da-4b85-a4f7-200b1d6a3069-kube-api-access-r2tst\") pod \"nova-cell1-novncproxy-0\" (UID: \"f4725b54-31da-4b85-a4f7-200b1d6a3069\") " pod="openstack/nova-cell1-novncproxy-0" Dec 11 08:31:54 crc kubenswrapper[4860]: I1211 08:31:54.622722 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 11 08:31:55 crc kubenswrapper[4860]: I1211 08:31:55.065420 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 11 08:31:55 crc kubenswrapper[4860]: I1211 08:31:55.590928 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3954f800-6e02-4387-bd5f-c9dd575ce3c8" path="/var/lib/kubelet/pods/3954f800-6e02-4387-bd5f-c9dd575ce3c8/volumes" Dec 11 08:31:55 crc kubenswrapper[4860]: I1211 08:31:55.915281 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"f4725b54-31da-4b85-a4f7-200b1d6a3069","Type":"ContainerStarted","Data":"f2e26529324d7af589492bc7dd915b110db2a4a1dcbadcd662aad3b8af7e1061"} Dec 11 08:31:55 crc kubenswrapper[4860]: I1211 08:31:55.915337 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"f4725b54-31da-4b85-a4f7-200b1d6a3069","Type":"ContainerStarted","Data":"96e86a60ad0454a39b6428160865208533e5f1d3c93991a9b4e081e39a06937c"} Dec 11 08:31:55 crc kubenswrapper[4860]: I1211 08:31:55.945958 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=1.9459308530000001 podStartE2EDuration="1.945930853s" podCreationTimestamp="2025-12-11 08:31:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:31:55.935816383 +0000 UTC m=+1248.664335458" watchObservedRunningTime="2025-12-11 08:31:55.945930853 +0000 UTC m=+1248.674449918" Dec 11 08:31:56 crc kubenswrapper[4860]: I1211 08:31:56.466081 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 11 08:31:56 crc kubenswrapper[4860]: I1211 08:31:56.467229 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 11 08:31:56 crc kubenswrapper[4860]: I1211 08:31:56.470784 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 11 08:31:56 crc kubenswrapper[4860]: I1211 08:31:56.479582 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 11 08:31:56 crc kubenswrapper[4860]: I1211 08:31:56.923596 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 11 08:31:56 crc kubenswrapper[4860]: I1211 08:31:56.933326 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 11 08:31:57 crc kubenswrapper[4860]: I1211 08:31:57.157721 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c7b6c5df9-7zcv8"] Dec 11 08:31:57 crc kubenswrapper[4860]: I1211 08:31:57.159518 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c7b6c5df9-7zcv8" Dec 11 08:31:57 crc kubenswrapper[4860]: I1211 08:31:57.175076 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c7b6c5df9-7zcv8"] Dec 11 08:31:57 crc kubenswrapper[4860]: I1211 08:31:57.252688 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/33f4ccd8-b2d7-43f7-a6f6-daf6ea2fa5b8-dns-swift-storage-0\") pod \"dnsmasq-dns-5c7b6c5df9-7zcv8\" (UID: \"33f4ccd8-b2d7-43f7-a6f6-daf6ea2fa5b8\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-7zcv8" Dec 11 08:31:57 crc kubenswrapper[4860]: I1211 08:31:57.252860 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/33f4ccd8-b2d7-43f7-a6f6-daf6ea2fa5b8-config\") pod \"dnsmasq-dns-5c7b6c5df9-7zcv8\" (UID: \"33f4ccd8-b2d7-43f7-a6f6-daf6ea2fa5b8\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-7zcv8" Dec 11 08:31:57 crc kubenswrapper[4860]: I1211 08:31:57.253013 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/33f4ccd8-b2d7-43f7-a6f6-daf6ea2fa5b8-ovsdbserver-nb\") pod \"dnsmasq-dns-5c7b6c5df9-7zcv8\" (UID: \"33f4ccd8-b2d7-43f7-a6f6-daf6ea2fa5b8\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-7zcv8" Dec 11 08:31:57 crc kubenswrapper[4860]: I1211 08:31:57.253083 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/33f4ccd8-b2d7-43f7-a6f6-daf6ea2fa5b8-ovsdbserver-sb\") pod \"dnsmasq-dns-5c7b6c5df9-7zcv8\" (UID: \"33f4ccd8-b2d7-43f7-a6f6-daf6ea2fa5b8\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-7zcv8" Dec 11 08:31:57 crc kubenswrapper[4860]: I1211 08:31:57.253258 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5n794\" (UniqueName: \"kubernetes.io/projected/33f4ccd8-b2d7-43f7-a6f6-daf6ea2fa5b8-kube-api-access-5n794\") pod \"dnsmasq-dns-5c7b6c5df9-7zcv8\" (UID: \"33f4ccd8-b2d7-43f7-a6f6-daf6ea2fa5b8\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-7zcv8" Dec 11 08:31:57 crc kubenswrapper[4860]: I1211 08:31:57.253315 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/33f4ccd8-b2d7-43f7-a6f6-daf6ea2fa5b8-dns-svc\") pod \"dnsmasq-dns-5c7b6c5df9-7zcv8\" (UID: \"33f4ccd8-b2d7-43f7-a6f6-daf6ea2fa5b8\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-7zcv8" Dec 11 08:31:57 crc kubenswrapper[4860]: I1211 08:31:57.355716 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/33f4ccd8-b2d7-43f7-a6f6-daf6ea2fa5b8-ovsdbserver-nb\") pod \"dnsmasq-dns-5c7b6c5df9-7zcv8\" (UID: \"33f4ccd8-b2d7-43f7-a6f6-daf6ea2fa5b8\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-7zcv8" Dec 11 08:31:57 crc kubenswrapper[4860]: I1211 08:31:57.355939 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/33f4ccd8-b2d7-43f7-a6f6-daf6ea2fa5b8-ovsdbserver-sb\") pod \"dnsmasq-dns-5c7b6c5df9-7zcv8\" (UID: \"33f4ccd8-b2d7-43f7-a6f6-daf6ea2fa5b8\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-7zcv8" Dec 11 08:31:57 crc kubenswrapper[4860]: I1211 08:31:57.355996 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5n794\" (UniqueName: \"kubernetes.io/projected/33f4ccd8-b2d7-43f7-a6f6-daf6ea2fa5b8-kube-api-access-5n794\") pod \"dnsmasq-dns-5c7b6c5df9-7zcv8\" (UID: \"33f4ccd8-b2d7-43f7-a6f6-daf6ea2fa5b8\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-7zcv8" Dec 11 08:31:57 crc kubenswrapper[4860]: I1211 08:31:57.356014 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/33f4ccd8-b2d7-43f7-a6f6-daf6ea2fa5b8-dns-svc\") pod \"dnsmasq-dns-5c7b6c5df9-7zcv8\" (UID: \"33f4ccd8-b2d7-43f7-a6f6-daf6ea2fa5b8\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-7zcv8" Dec 11 08:31:57 crc kubenswrapper[4860]: I1211 08:31:57.356062 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/33f4ccd8-b2d7-43f7-a6f6-daf6ea2fa5b8-dns-swift-storage-0\") pod \"dnsmasq-dns-5c7b6c5df9-7zcv8\" (UID: \"33f4ccd8-b2d7-43f7-a6f6-daf6ea2fa5b8\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-7zcv8" Dec 11 08:31:57 crc kubenswrapper[4860]: I1211 08:31:57.356108 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/33f4ccd8-b2d7-43f7-a6f6-daf6ea2fa5b8-config\") pod \"dnsmasq-dns-5c7b6c5df9-7zcv8\" (UID: \"33f4ccd8-b2d7-43f7-a6f6-daf6ea2fa5b8\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-7zcv8" Dec 11 08:31:57 crc kubenswrapper[4860]: I1211 08:31:57.356718 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/33f4ccd8-b2d7-43f7-a6f6-daf6ea2fa5b8-ovsdbserver-nb\") pod \"dnsmasq-dns-5c7b6c5df9-7zcv8\" (UID: \"33f4ccd8-b2d7-43f7-a6f6-daf6ea2fa5b8\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-7zcv8" Dec 11 08:31:57 crc kubenswrapper[4860]: I1211 08:31:57.357149 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/33f4ccd8-b2d7-43f7-a6f6-daf6ea2fa5b8-config\") pod \"dnsmasq-dns-5c7b6c5df9-7zcv8\" (UID: \"33f4ccd8-b2d7-43f7-a6f6-daf6ea2fa5b8\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-7zcv8" Dec 11 08:31:57 crc kubenswrapper[4860]: I1211 08:31:57.357231 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/33f4ccd8-b2d7-43f7-a6f6-daf6ea2fa5b8-dns-svc\") pod \"dnsmasq-dns-5c7b6c5df9-7zcv8\" (UID: \"33f4ccd8-b2d7-43f7-a6f6-daf6ea2fa5b8\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-7zcv8" Dec 11 08:31:57 crc kubenswrapper[4860]: I1211 08:31:57.357513 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/33f4ccd8-b2d7-43f7-a6f6-daf6ea2fa5b8-dns-swift-storage-0\") pod \"dnsmasq-dns-5c7b6c5df9-7zcv8\" (UID: \"33f4ccd8-b2d7-43f7-a6f6-daf6ea2fa5b8\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-7zcv8" Dec 11 08:31:57 crc kubenswrapper[4860]: I1211 08:31:57.358001 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/33f4ccd8-b2d7-43f7-a6f6-daf6ea2fa5b8-ovsdbserver-sb\") pod \"dnsmasq-dns-5c7b6c5df9-7zcv8\" (UID: \"33f4ccd8-b2d7-43f7-a6f6-daf6ea2fa5b8\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-7zcv8" Dec 11 08:31:57 crc kubenswrapper[4860]: I1211 08:31:57.386074 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5n794\" (UniqueName: \"kubernetes.io/projected/33f4ccd8-b2d7-43f7-a6f6-daf6ea2fa5b8-kube-api-access-5n794\") pod \"dnsmasq-dns-5c7b6c5df9-7zcv8\" (UID: \"33f4ccd8-b2d7-43f7-a6f6-daf6ea2fa5b8\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-7zcv8" Dec 11 08:31:57 crc kubenswrapper[4860]: I1211 08:31:57.494444 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c7b6c5df9-7zcv8" Dec 11 08:31:58 crc kubenswrapper[4860]: I1211 08:31:58.001140 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c7b6c5df9-7zcv8"] Dec 11 08:31:58 crc kubenswrapper[4860]: I1211 08:31:58.957351 4860 generic.go:334] "Generic (PLEG): container finished" podID="33f4ccd8-b2d7-43f7-a6f6-daf6ea2fa5b8" containerID="1aa442d3e55c72e0c256ef664cf020cd34cd97189a1619b6819b4ae1a5696ea7" exitCode=0 Dec 11 08:31:58 crc kubenswrapper[4860]: I1211 08:31:58.959192 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c7b6c5df9-7zcv8" event={"ID":"33f4ccd8-b2d7-43f7-a6f6-daf6ea2fa5b8","Type":"ContainerDied","Data":"1aa442d3e55c72e0c256ef664cf020cd34cd97189a1619b6819b4ae1a5696ea7"} Dec 11 08:31:58 crc kubenswrapper[4860]: I1211 08:31:58.959311 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c7b6c5df9-7zcv8" event={"ID":"33f4ccd8-b2d7-43f7-a6f6-daf6ea2fa5b8","Type":"ContainerStarted","Data":"1c83d81db7c29d50bca861c14fde14a3cc6920ebec192ef3bb0fb2f2e272b71e"} Dec 11 08:31:59 crc kubenswrapper[4860]: I1211 08:31:59.245300 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 11 08:31:59 crc kubenswrapper[4860]: I1211 08:31:59.245817 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="47b450dd-42c9-428c-aa47-a971d2dfa3c5" containerName="ceilometer-central-agent" containerID="cri-o://c2ea26488b5d7472bd8221a6fff096ef6aef5589626b5e5b3f29566e841cfb3a" gracePeriod=30 Dec 11 08:31:59 crc kubenswrapper[4860]: I1211 08:31:59.245880 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="47b450dd-42c9-428c-aa47-a971d2dfa3c5" containerName="proxy-httpd" containerID="cri-o://98f36f332db8260b7bb17975f0dd0b38806bc4f0217970243d7c9ea4b2839ede" gracePeriod=30 Dec 11 08:31:59 crc kubenswrapper[4860]: I1211 08:31:59.245978 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="47b450dd-42c9-428c-aa47-a971d2dfa3c5" containerName="sg-core" containerID="cri-o://9f5fcb630d5bb7b7347be31a08fc385034a9bc7d82ee521a19e39962890ea047" gracePeriod=30 Dec 11 08:31:59 crc kubenswrapper[4860]: I1211 08:31:59.246004 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="47b450dd-42c9-428c-aa47-a971d2dfa3c5" containerName="ceilometer-notification-agent" containerID="cri-o://be79c2475bfd46873eb93a3a1f2a092d8cff2b6213a2d7e5015c7d8fa6d222d9" gracePeriod=30 Dec 11 08:31:59 crc kubenswrapper[4860]: I1211 08:31:59.625195 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 11 08:31:59 crc kubenswrapper[4860]: I1211 08:31:59.645623 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 11 08:31:59 crc kubenswrapper[4860]: I1211 08:31:59.970308 4860 generic.go:334] "Generic (PLEG): container finished" podID="47b450dd-42c9-428c-aa47-a971d2dfa3c5" containerID="98f36f332db8260b7bb17975f0dd0b38806bc4f0217970243d7c9ea4b2839ede" exitCode=0 Dec 11 08:31:59 crc kubenswrapper[4860]: I1211 08:31:59.970346 4860 generic.go:334] "Generic (PLEG): container finished" podID="47b450dd-42c9-428c-aa47-a971d2dfa3c5" containerID="9f5fcb630d5bb7b7347be31a08fc385034a9bc7d82ee521a19e39962890ea047" exitCode=2 Dec 11 08:31:59 crc kubenswrapper[4860]: I1211 08:31:59.970356 4860 generic.go:334] "Generic (PLEG): container finished" podID="47b450dd-42c9-428c-aa47-a971d2dfa3c5" containerID="c2ea26488b5d7472bd8221a6fff096ef6aef5589626b5e5b3f29566e841cfb3a" exitCode=0 Dec 11 08:31:59 crc kubenswrapper[4860]: I1211 08:31:59.970400 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"47b450dd-42c9-428c-aa47-a971d2dfa3c5","Type":"ContainerDied","Data":"98f36f332db8260b7bb17975f0dd0b38806bc4f0217970243d7c9ea4b2839ede"} Dec 11 08:31:59 crc kubenswrapper[4860]: I1211 08:31:59.970447 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"47b450dd-42c9-428c-aa47-a971d2dfa3c5","Type":"ContainerDied","Data":"9f5fcb630d5bb7b7347be31a08fc385034a9bc7d82ee521a19e39962890ea047"} Dec 11 08:31:59 crc kubenswrapper[4860]: I1211 08:31:59.970462 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"47b450dd-42c9-428c-aa47-a971d2dfa3c5","Type":"ContainerDied","Data":"c2ea26488b5d7472bd8221a6fff096ef6aef5589626b5e5b3f29566e841cfb3a"} Dec 11 08:31:59 crc kubenswrapper[4860]: I1211 08:31:59.972298 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c7b6c5df9-7zcv8" event={"ID":"33f4ccd8-b2d7-43f7-a6f6-daf6ea2fa5b8","Type":"ContainerStarted","Data":"0f40eceff896f6453086ec0213e1ebfaa19e60a191b1173daa71b9121db12111"} Dec 11 08:31:59 crc kubenswrapper[4860]: I1211 08:31:59.972491 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="2714dc33-cc1b-4fed-af2c-9ab58397d1c0" containerName="nova-api-log" containerID="cri-o://7268c87e4750580d441f2217171d227fd7ee59a77faf3c385af1d37c4272934c" gracePeriod=30 Dec 11 08:31:59 crc kubenswrapper[4860]: I1211 08:31:59.972615 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="2714dc33-cc1b-4fed-af2c-9ab58397d1c0" containerName="nova-api-api" containerID="cri-o://ec7f65b71fd7e5c584de43621051867f16fc801d9f4efd6258a6b70b740910ca" gracePeriod=30 Dec 11 08:32:00 crc kubenswrapper[4860]: I1211 08:32:00.002962 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5c7b6c5df9-7zcv8" podStartSLOduration=3.002937836 podStartE2EDuration="3.002937836s" podCreationTimestamp="2025-12-11 08:31:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:31:59.996490593 +0000 UTC m=+1252.725009688" watchObservedRunningTime="2025-12-11 08:32:00.002937836 +0000 UTC m=+1252.731456891" Dec 11 08:32:00 crc kubenswrapper[4860]: I1211 08:32:00.667190 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 11 08:32:00 crc kubenswrapper[4860]: I1211 08:32:00.833130 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47b450dd-42c9-428c-aa47-a971d2dfa3c5-config-data\") pod \"47b450dd-42c9-428c-aa47-a971d2dfa3c5\" (UID: \"47b450dd-42c9-428c-aa47-a971d2dfa3c5\") " Dec 11 08:32:00 crc kubenswrapper[4860]: I1211 08:32:00.833231 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/47b450dd-42c9-428c-aa47-a971d2dfa3c5-sg-core-conf-yaml\") pod \"47b450dd-42c9-428c-aa47-a971d2dfa3c5\" (UID: \"47b450dd-42c9-428c-aa47-a971d2dfa3c5\") " Dec 11 08:32:00 crc kubenswrapper[4860]: I1211 08:32:00.833284 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/47b450dd-42c9-428c-aa47-a971d2dfa3c5-run-httpd\") pod \"47b450dd-42c9-428c-aa47-a971d2dfa3c5\" (UID: \"47b450dd-42c9-428c-aa47-a971d2dfa3c5\") " Dec 11 08:32:00 crc kubenswrapper[4860]: I1211 08:32:00.833305 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/47b450dd-42c9-428c-aa47-a971d2dfa3c5-log-httpd\") pod \"47b450dd-42c9-428c-aa47-a971d2dfa3c5\" (UID: \"47b450dd-42c9-428c-aa47-a971d2dfa3c5\") " Dec 11 08:32:00 crc kubenswrapper[4860]: I1211 08:32:00.833349 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9z89t\" (UniqueName: \"kubernetes.io/projected/47b450dd-42c9-428c-aa47-a971d2dfa3c5-kube-api-access-9z89t\") pod \"47b450dd-42c9-428c-aa47-a971d2dfa3c5\" (UID: \"47b450dd-42c9-428c-aa47-a971d2dfa3c5\") " Dec 11 08:32:00 crc kubenswrapper[4860]: I1211 08:32:00.833414 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/47b450dd-42c9-428c-aa47-a971d2dfa3c5-scripts\") pod \"47b450dd-42c9-428c-aa47-a971d2dfa3c5\" (UID: \"47b450dd-42c9-428c-aa47-a971d2dfa3c5\") " Dec 11 08:32:00 crc kubenswrapper[4860]: I1211 08:32:00.833562 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/47b450dd-42c9-428c-aa47-a971d2dfa3c5-ceilometer-tls-certs\") pod \"47b450dd-42c9-428c-aa47-a971d2dfa3c5\" (UID: \"47b450dd-42c9-428c-aa47-a971d2dfa3c5\") " Dec 11 08:32:00 crc kubenswrapper[4860]: I1211 08:32:00.833593 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47b450dd-42c9-428c-aa47-a971d2dfa3c5-combined-ca-bundle\") pod \"47b450dd-42c9-428c-aa47-a971d2dfa3c5\" (UID: \"47b450dd-42c9-428c-aa47-a971d2dfa3c5\") " Dec 11 08:32:00 crc kubenswrapper[4860]: I1211 08:32:00.834122 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/47b450dd-42c9-428c-aa47-a971d2dfa3c5-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "47b450dd-42c9-428c-aa47-a971d2dfa3c5" (UID: "47b450dd-42c9-428c-aa47-a971d2dfa3c5"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:32:00 crc kubenswrapper[4860]: I1211 08:32:00.835279 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/47b450dd-42c9-428c-aa47-a971d2dfa3c5-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "47b450dd-42c9-428c-aa47-a971d2dfa3c5" (UID: "47b450dd-42c9-428c-aa47-a971d2dfa3c5"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:32:00 crc kubenswrapper[4860]: I1211 08:32:00.839889 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47b450dd-42c9-428c-aa47-a971d2dfa3c5-scripts" (OuterVolumeSpecName: "scripts") pod "47b450dd-42c9-428c-aa47-a971d2dfa3c5" (UID: "47b450dd-42c9-428c-aa47-a971d2dfa3c5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:32:00 crc kubenswrapper[4860]: I1211 08:32:00.839942 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/47b450dd-42c9-428c-aa47-a971d2dfa3c5-kube-api-access-9z89t" (OuterVolumeSpecName: "kube-api-access-9z89t") pod "47b450dd-42c9-428c-aa47-a971d2dfa3c5" (UID: "47b450dd-42c9-428c-aa47-a971d2dfa3c5"). InnerVolumeSpecName "kube-api-access-9z89t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:32:00 crc kubenswrapper[4860]: I1211 08:32:00.869026 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47b450dd-42c9-428c-aa47-a971d2dfa3c5-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "47b450dd-42c9-428c-aa47-a971d2dfa3c5" (UID: "47b450dd-42c9-428c-aa47-a971d2dfa3c5"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:32:00 crc kubenswrapper[4860]: I1211 08:32:00.895695 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47b450dd-42c9-428c-aa47-a971d2dfa3c5-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "47b450dd-42c9-428c-aa47-a971d2dfa3c5" (UID: "47b450dd-42c9-428c-aa47-a971d2dfa3c5"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:32:00 crc kubenswrapper[4860]: I1211 08:32:00.924863 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47b450dd-42c9-428c-aa47-a971d2dfa3c5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "47b450dd-42c9-428c-aa47-a971d2dfa3c5" (UID: "47b450dd-42c9-428c-aa47-a971d2dfa3c5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:32:00 crc kubenswrapper[4860]: I1211 08:32:00.935623 4860 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/47b450dd-42c9-428c-aa47-a971d2dfa3c5-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 11 08:32:00 crc kubenswrapper[4860]: I1211 08:32:00.935697 4860 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/47b450dd-42c9-428c-aa47-a971d2dfa3c5-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 11 08:32:00 crc kubenswrapper[4860]: I1211 08:32:00.935710 4860 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/47b450dd-42c9-428c-aa47-a971d2dfa3c5-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 11 08:32:00 crc kubenswrapper[4860]: I1211 08:32:00.935724 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9z89t\" (UniqueName: \"kubernetes.io/projected/47b450dd-42c9-428c-aa47-a971d2dfa3c5-kube-api-access-9z89t\") on node \"crc\" DevicePath \"\"" Dec 11 08:32:00 crc kubenswrapper[4860]: I1211 08:32:00.935734 4860 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/47b450dd-42c9-428c-aa47-a971d2dfa3c5-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 08:32:00 crc kubenswrapper[4860]: I1211 08:32:00.935742 4860 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/47b450dd-42c9-428c-aa47-a971d2dfa3c5-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 11 08:32:00 crc kubenswrapper[4860]: I1211 08:32:00.935752 4860 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47b450dd-42c9-428c-aa47-a971d2dfa3c5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 08:32:00 crc kubenswrapper[4860]: I1211 08:32:00.961119 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47b450dd-42c9-428c-aa47-a971d2dfa3c5-config-data" (OuterVolumeSpecName: "config-data") pod "47b450dd-42c9-428c-aa47-a971d2dfa3c5" (UID: "47b450dd-42c9-428c-aa47-a971d2dfa3c5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:32:00 crc kubenswrapper[4860]: I1211 08:32:00.988684 4860 generic.go:334] "Generic (PLEG): container finished" podID="2714dc33-cc1b-4fed-af2c-9ab58397d1c0" containerID="7268c87e4750580d441f2217171d227fd7ee59a77faf3c385af1d37c4272934c" exitCode=143 Dec 11 08:32:00 crc kubenswrapper[4860]: I1211 08:32:00.988762 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2714dc33-cc1b-4fed-af2c-9ab58397d1c0","Type":"ContainerDied","Data":"7268c87e4750580d441f2217171d227fd7ee59a77faf3c385af1d37c4272934c"} Dec 11 08:32:00 crc kubenswrapper[4860]: I1211 08:32:00.991388 4860 generic.go:334] "Generic (PLEG): container finished" podID="47b450dd-42c9-428c-aa47-a971d2dfa3c5" containerID="be79c2475bfd46873eb93a3a1f2a092d8cff2b6213a2d7e5015c7d8fa6d222d9" exitCode=0 Dec 11 08:32:00 crc kubenswrapper[4860]: I1211 08:32:00.991446 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 11 08:32:00 crc kubenswrapper[4860]: I1211 08:32:00.991530 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"47b450dd-42c9-428c-aa47-a971d2dfa3c5","Type":"ContainerDied","Data":"be79c2475bfd46873eb93a3a1f2a092d8cff2b6213a2d7e5015c7d8fa6d222d9"} Dec 11 08:32:00 crc kubenswrapper[4860]: I1211 08:32:00.991570 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"47b450dd-42c9-428c-aa47-a971d2dfa3c5","Type":"ContainerDied","Data":"8b435cff8d1221c4dc29857a7afa255bf2214be17b34ae8ce7b90506c02b5094"} Dec 11 08:32:00 crc kubenswrapper[4860]: I1211 08:32:00.991589 4860 scope.go:117] "RemoveContainer" containerID="98f36f332db8260b7bb17975f0dd0b38806bc4f0217970243d7c9ea4b2839ede" Dec 11 08:32:00 crc kubenswrapper[4860]: I1211 08:32:00.993627 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5c7b6c5df9-7zcv8" Dec 11 08:32:01 crc kubenswrapper[4860]: I1211 08:32:01.041541 4860 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47b450dd-42c9-428c-aa47-a971d2dfa3c5-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 08:32:01 crc kubenswrapper[4860]: I1211 08:32:01.044335 4860 scope.go:117] "RemoveContainer" containerID="9f5fcb630d5bb7b7347be31a08fc385034a9bc7d82ee521a19e39962890ea047" Dec 11 08:32:01 crc kubenswrapper[4860]: I1211 08:32:01.055723 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 11 08:32:01 crc kubenswrapper[4860]: I1211 08:32:01.075968 4860 scope.go:117] "RemoveContainer" containerID="be79c2475bfd46873eb93a3a1f2a092d8cff2b6213a2d7e5015c7d8fa6d222d9" Dec 11 08:32:01 crc kubenswrapper[4860]: I1211 08:32:01.081703 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 11 08:32:01 crc kubenswrapper[4860]: I1211 08:32:01.100154 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 11 08:32:01 crc kubenswrapper[4860]: E1211 08:32:01.100669 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47b450dd-42c9-428c-aa47-a971d2dfa3c5" containerName="ceilometer-notification-agent" Dec 11 08:32:01 crc kubenswrapper[4860]: I1211 08:32:01.100688 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="47b450dd-42c9-428c-aa47-a971d2dfa3c5" containerName="ceilometer-notification-agent" Dec 11 08:32:01 crc kubenswrapper[4860]: E1211 08:32:01.100721 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47b450dd-42c9-428c-aa47-a971d2dfa3c5" containerName="ceilometer-central-agent" Dec 11 08:32:01 crc kubenswrapper[4860]: I1211 08:32:01.100730 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="47b450dd-42c9-428c-aa47-a971d2dfa3c5" containerName="ceilometer-central-agent" Dec 11 08:32:01 crc kubenswrapper[4860]: E1211 08:32:01.100754 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47b450dd-42c9-428c-aa47-a971d2dfa3c5" containerName="sg-core" Dec 11 08:32:01 crc kubenswrapper[4860]: I1211 08:32:01.100760 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="47b450dd-42c9-428c-aa47-a971d2dfa3c5" containerName="sg-core" Dec 11 08:32:01 crc kubenswrapper[4860]: E1211 08:32:01.100775 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47b450dd-42c9-428c-aa47-a971d2dfa3c5" containerName="proxy-httpd" Dec 11 08:32:01 crc kubenswrapper[4860]: I1211 08:32:01.100781 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="47b450dd-42c9-428c-aa47-a971d2dfa3c5" containerName="proxy-httpd" Dec 11 08:32:01 crc kubenswrapper[4860]: I1211 08:32:01.100952 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="47b450dd-42c9-428c-aa47-a971d2dfa3c5" containerName="sg-core" Dec 11 08:32:01 crc kubenswrapper[4860]: I1211 08:32:01.100968 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="47b450dd-42c9-428c-aa47-a971d2dfa3c5" containerName="ceilometer-notification-agent" Dec 11 08:32:01 crc kubenswrapper[4860]: I1211 08:32:01.100986 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="47b450dd-42c9-428c-aa47-a971d2dfa3c5" containerName="proxy-httpd" Dec 11 08:32:01 crc kubenswrapper[4860]: I1211 08:32:01.100998 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="47b450dd-42c9-428c-aa47-a971d2dfa3c5" containerName="ceilometer-central-agent" Dec 11 08:32:01 crc kubenswrapper[4860]: I1211 08:32:01.102945 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 11 08:32:01 crc kubenswrapper[4860]: I1211 08:32:01.107165 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 11 08:32:01 crc kubenswrapper[4860]: I1211 08:32:01.107174 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 11 08:32:01 crc kubenswrapper[4860]: I1211 08:32:01.109802 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 11 08:32:01 crc kubenswrapper[4860]: I1211 08:32:01.119913 4860 scope.go:117] "RemoveContainer" containerID="c2ea26488b5d7472bd8221a6fff096ef6aef5589626b5e5b3f29566e841cfb3a" Dec 11 08:32:01 crc kubenswrapper[4860]: I1211 08:32:01.121635 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 11 08:32:01 crc kubenswrapper[4860]: I1211 08:32:01.157102 4860 scope.go:117] "RemoveContainer" containerID="98f36f332db8260b7bb17975f0dd0b38806bc4f0217970243d7c9ea4b2839ede" Dec 11 08:32:01 crc kubenswrapper[4860]: E1211 08:32:01.162890 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"98f36f332db8260b7bb17975f0dd0b38806bc4f0217970243d7c9ea4b2839ede\": container with ID starting with 98f36f332db8260b7bb17975f0dd0b38806bc4f0217970243d7c9ea4b2839ede not found: ID does not exist" containerID="98f36f332db8260b7bb17975f0dd0b38806bc4f0217970243d7c9ea4b2839ede" Dec 11 08:32:01 crc kubenswrapper[4860]: I1211 08:32:01.162965 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"98f36f332db8260b7bb17975f0dd0b38806bc4f0217970243d7c9ea4b2839ede"} err="failed to get container status \"98f36f332db8260b7bb17975f0dd0b38806bc4f0217970243d7c9ea4b2839ede\": rpc error: code = NotFound desc = could not find container \"98f36f332db8260b7bb17975f0dd0b38806bc4f0217970243d7c9ea4b2839ede\": container with ID starting with 98f36f332db8260b7bb17975f0dd0b38806bc4f0217970243d7c9ea4b2839ede not found: ID does not exist" Dec 11 08:32:01 crc kubenswrapper[4860]: I1211 08:32:01.163017 4860 scope.go:117] "RemoveContainer" containerID="9f5fcb630d5bb7b7347be31a08fc385034a9bc7d82ee521a19e39962890ea047" Dec 11 08:32:01 crc kubenswrapper[4860]: E1211 08:32:01.164089 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9f5fcb630d5bb7b7347be31a08fc385034a9bc7d82ee521a19e39962890ea047\": container with ID starting with 9f5fcb630d5bb7b7347be31a08fc385034a9bc7d82ee521a19e39962890ea047 not found: ID does not exist" containerID="9f5fcb630d5bb7b7347be31a08fc385034a9bc7d82ee521a19e39962890ea047" Dec 11 08:32:01 crc kubenswrapper[4860]: I1211 08:32:01.164142 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f5fcb630d5bb7b7347be31a08fc385034a9bc7d82ee521a19e39962890ea047"} err="failed to get container status \"9f5fcb630d5bb7b7347be31a08fc385034a9bc7d82ee521a19e39962890ea047\": rpc error: code = NotFound desc = could not find container \"9f5fcb630d5bb7b7347be31a08fc385034a9bc7d82ee521a19e39962890ea047\": container with ID starting with 9f5fcb630d5bb7b7347be31a08fc385034a9bc7d82ee521a19e39962890ea047 not found: ID does not exist" Dec 11 08:32:01 crc kubenswrapper[4860]: I1211 08:32:01.164175 4860 scope.go:117] "RemoveContainer" containerID="be79c2475bfd46873eb93a3a1f2a092d8cff2b6213a2d7e5015c7d8fa6d222d9" Dec 11 08:32:01 crc kubenswrapper[4860]: E1211 08:32:01.164592 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"be79c2475bfd46873eb93a3a1f2a092d8cff2b6213a2d7e5015c7d8fa6d222d9\": container with ID starting with be79c2475bfd46873eb93a3a1f2a092d8cff2b6213a2d7e5015c7d8fa6d222d9 not found: ID does not exist" containerID="be79c2475bfd46873eb93a3a1f2a092d8cff2b6213a2d7e5015c7d8fa6d222d9" Dec 11 08:32:01 crc kubenswrapper[4860]: I1211 08:32:01.164663 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"be79c2475bfd46873eb93a3a1f2a092d8cff2b6213a2d7e5015c7d8fa6d222d9"} err="failed to get container status \"be79c2475bfd46873eb93a3a1f2a092d8cff2b6213a2d7e5015c7d8fa6d222d9\": rpc error: code = NotFound desc = could not find container \"be79c2475bfd46873eb93a3a1f2a092d8cff2b6213a2d7e5015c7d8fa6d222d9\": container with ID starting with be79c2475bfd46873eb93a3a1f2a092d8cff2b6213a2d7e5015c7d8fa6d222d9 not found: ID does not exist" Dec 11 08:32:01 crc kubenswrapper[4860]: I1211 08:32:01.164697 4860 scope.go:117] "RemoveContainer" containerID="c2ea26488b5d7472bd8221a6fff096ef6aef5589626b5e5b3f29566e841cfb3a" Dec 11 08:32:01 crc kubenswrapper[4860]: E1211 08:32:01.165047 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c2ea26488b5d7472bd8221a6fff096ef6aef5589626b5e5b3f29566e841cfb3a\": container with ID starting with c2ea26488b5d7472bd8221a6fff096ef6aef5589626b5e5b3f29566e841cfb3a not found: ID does not exist" containerID="c2ea26488b5d7472bd8221a6fff096ef6aef5589626b5e5b3f29566e841cfb3a" Dec 11 08:32:01 crc kubenswrapper[4860]: I1211 08:32:01.165094 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c2ea26488b5d7472bd8221a6fff096ef6aef5589626b5e5b3f29566e841cfb3a"} err="failed to get container status \"c2ea26488b5d7472bd8221a6fff096ef6aef5589626b5e5b3f29566e841cfb3a\": rpc error: code = NotFound desc = could not find container \"c2ea26488b5d7472bd8221a6fff096ef6aef5589626b5e5b3f29566e841cfb3a\": container with ID starting with c2ea26488b5d7472bd8221a6fff096ef6aef5589626b5e5b3f29566e841cfb3a not found: ID does not exist" Dec 11 08:32:01 crc kubenswrapper[4860]: I1211 08:32:01.245245 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/90181451-368d-48a7-87f3-542be3873484-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"90181451-368d-48a7-87f3-542be3873484\") " pod="openstack/ceilometer-0" Dec 11 08:32:01 crc kubenswrapper[4860]: I1211 08:32:01.245314 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/90181451-368d-48a7-87f3-542be3873484-run-httpd\") pod \"ceilometer-0\" (UID: \"90181451-368d-48a7-87f3-542be3873484\") " pod="openstack/ceilometer-0" Dec 11 08:32:01 crc kubenswrapper[4860]: I1211 08:32:01.245376 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/90181451-368d-48a7-87f3-542be3873484-log-httpd\") pod \"ceilometer-0\" (UID: \"90181451-368d-48a7-87f3-542be3873484\") " pod="openstack/ceilometer-0" Dec 11 08:32:01 crc kubenswrapper[4860]: I1211 08:32:01.245401 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/90181451-368d-48a7-87f3-542be3873484-scripts\") pod \"ceilometer-0\" (UID: \"90181451-368d-48a7-87f3-542be3873484\") " pod="openstack/ceilometer-0" Dec 11 08:32:01 crc kubenswrapper[4860]: I1211 08:32:01.245445 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90181451-368d-48a7-87f3-542be3873484-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"90181451-368d-48a7-87f3-542be3873484\") " pod="openstack/ceilometer-0" Dec 11 08:32:01 crc kubenswrapper[4860]: I1211 08:32:01.245485 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8qn58\" (UniqueName: \"kubernetes.io/projected/90181451-368d-48a7-87f3-542be3873484-kube-api-access-8qn58\") pod \"ceilometer-0\" (UID: \"90181451-368d-48a7-87f3-542be3873484\") " pod="openstack/ceilometer-0" Dec 11 08:32:01 crc kubenswrapper[4860]: I1211 08:32:01.245835 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/90181451-368d-48a7-87f3-542be3873484-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"90181451-368d-48a7-87f3-542be3873484\") " pod="openstack/ceilometer-0" Dec 11 08:32:01 crc kubenswrapper[4860]: I1211 08:32:01.251093 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/90181451-368d-48a7-87f3-542be3873484-config-data\") pod \"ceilometer-0\" (UID: \"90181451-368d-48a7-87f3-542be3873484\") " pod="openstack/ceilometer-0" Dec 11 08:32:01 crc kubenswrapper[4860]: I1211 08:32:01.284612 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 11 08:32:01 crc kubenswrapper[4860]: E1211 08:32:01.285527 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[ceilometer-tls-certs combined-ca-bundle config-data kube-api-access-8qn58 log-httpd run-httpd scripts sg-core-conf-yaml], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/ceilometer-0" podUID="90181451-368d-48a7-87f3-542be3873484" Dec 11 08:32:01 crc kubenswrapper[4860]: I1211 08:32:01.353338 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/90181451-368d-48a7-87f3-542be3873484-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"90181451-368d-48a7-87f3-542be3873484\") " pod="openstack/ceilometer-0" Dec 11 08:32:01 crc kubenswrapper[4860]: I1211 08:32:01.353409 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/90181451-368d-48a7-87f3-542be3873484-config-data\") pod \"ceilometer-0\" (UID: \"90181451-368d-48a7-87f3-542be3873484\") " pod="openstack/ceilometer-0" Dec 11 08:32:01 crc kubenswrapper[4860]: I1211 08:32:01.353435 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/90181451-368d-48a7-87f3-542be3873484-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"90181451-368d-48a7-87f3-542be3873484\") " pod="openstack/ceilometer-0" Dec 11 08:32:01 crc kubenswrapper[4860]: I1211 08:32:01.353465 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/90181451-368d-48a7-87f3-542be3873484-run-httpd\") pod \"ceilometer-0\" (UID: \"90181451-368d-48a7-87f3-542be3873484\") " pod="openstack/ceilometer-0" Dec 11 08:32:01 crc kubenswrapper[4860]: I1211 08:32:01.353504 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/90181451-368d-48a7-87f3-542be3873484-log-httpd\") pod \"ceilometer-0\" (UID: \"90181451-368d-48a7-87f3-542be3873484\") " pod="openstack/ceilometer-0" Dec 11 08:32:01 crc kubenswrapper[4860]: I1211 08:32:01.353525 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/90181451-368d-48a7-87f3-542be3873484-scripts\") pod \"ceilometer-0\" (UID: \"90181451-368d-48a7-87f3-542be3873484\") " pod="openstack/ceilometer-0" Dec 11 08:32:01 crc kubenswrapper[4860]: I1211 08:32:01.353559 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90181451-368d-48a7-87f3-542be3873484-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"90181451-368d-48a7-87f3-542be3873484\") " pod="openstack/ceilometer-0" Dec 11 08:32:01 crc kubenswrapper[4860]: I1211 08:32:01.353595 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8qn58\" (UniqueName: \"kubernetes.io/projected/90181451-368d-48a7-87f3-542be3873484-kube-api-access-8qn58\") pod \"ceilometer-0\" (UID: \"90181451-368d-48a7-87f3-542be3873484\") " pod="openstack/ceilometer-0" Dec 11 08:32:01 crc kubenswrapper[4860]: I1211 08:32:01.354296 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/90181451-368d-48a7-87f3-542be3873484-run-httpd\") pod \"ceilometer-0\" (UID: \"90181451-368d-48a7-87f3-542be3873484\") " pod="openstack/ceilometer-0" Dec 11 08:32:01 crc kubenswrapper[4860]: I1211 08:32:01.354565 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/90181451-368d-48a7-87f3-542be3873484-log-httpd\") pod \"ceilometer-0\" (UID: \"90181451-368d-48a7-87f3-542be3873484\") " pod="openstack/ceilometer-0" Dec 11 08:32:01 crc kubenswrapper[4860]: I1211 08:32:01.358457 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/90181451-368d-48a7-87f3-542be3873484-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"90181451-368d-48a7-87f3-542be3873484\") " pod="openstack/ceilometer-0" Dec 11 08:32:01 crc kubenswrapper[4860]: I1211 08:32:01.358892 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/90181451-368d-48a7-87f3-542be3873484-scripts\") pod \"ceilometer-0\" (UID: \"90181451-368d-48a7-87f3-542be3873484\") " pod="openstack/ceilometer-0" Dec 11 08:32:01 crc kubenswrapper[4860]: I1211 08:32:01.359153 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90181451-368d-48a7-87f3-542be3873484-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"90181451-368d-48a7-87f3-542be3873484\") " pod="openstack/ceilometer-0" Dec 11 08:32:01 crc kubenswrapper[4860]: I1211 08:32:01.359634 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/90181451-368d-48a7-87f3-542be3873484-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"90181451-368d-48a7-87f3-542be3873484\") " pod="openstack/ceilometer-0" Dec 11 08:32:01 crc kubenswrapper[4860]: I1211 08:32:01.360461 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/90181451-368d-48a7-87f3-542be3873484-config-data\") pod \"ceilometer-0\" (UID: \"90181451-368d-48a7-87f3-542be3873484\") " pod="openstack/ceilometer-0" Dec 11 08:32:01 crc kubenswrapper[4860]: I1211 08:32:01.372859 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8qn58\" (UniqueName: \"kubernetes.io/projected/90181451-368d-48a7-87f3-542be3873484-kube-api-access-8qn58\") pod \"ceilometer-0\" (UID: \"90181451-368d-48a7-87f3-542be3873484\") " pod="openstack/ceilometer-0" Dec 11 08:32:01 crc kubenswrapper[4860]: I1211 08:32:01.597777 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="47b450dd-42c9-428c-aa47-a971d2dfa3c5" path="/var/lib/kubelet/pods/47b450dd-42c9-428c-aa47-a971d2dfa3c5/volumes" Dec 11 08:32:02 crc kubenswrapper[4860]: I1211 08:32:02.002281 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 11 08:32:02 crc kubenswrapper[4860]: I1211 08:32:02.017021 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 11 08:32:02 crc kubenswrapper[4860]: I1211 08:32:02.169130 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/90181451-368d-48a7-87f3-542be3873484-config-data\") pod \"90181451-368d-48a7-87f3-542be3873484\" (UID: \"90181451-368d-48a7-87f3-542be3873484\") " Dec 11 08:32:02 crc kubenswrapper[4860]: I1211 08:32:02.169238 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/90181451-368d-48a7-87f3-542be3873484-run-httpd\") pod \"90181451-368d-48a7-87f3-542be3873484\" (UID: \"90181451-368d-48a7-87f3-542be3873484\") " Dec 11 08:32:02 crc kubenswrapper[4860]: I1211 08:32:02.169410 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/90181451-368d-48a7-87f3-542be3873484-ceilometer-tls-certs\") pod \"90181451-368d-48a7-87f3-542be3873484\" (UID: \"90181451-368d-48a7-87f3-542be3873484\") " Dec 11 08:32:02 crc kubenswrapper[4860]: I1211 08:32:02.169520 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8qn58\" (UniqueName: \"kubernetes.io/projected/90181451-368d-48a7-87f3-542be3873484-kube-api-access-8qn58\") pod \"90181451-368d-48a7-87f3-542be3873484\" (UID: \"90181451-368d-48a7-87f3-542be3873484\") " Dec 11 08:32:02 crc kubenswrapper[4860]: I1211 08:32:02.169829 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/90181451-368d-48a7-87f3-542be3873484-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "90181451-368d-48a7-87f3-542be3873484" (UID: "90181451-368d-48a7-87f3-542be3873484"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:32:02 crc kubenswrapper[4860]: I1211 08:32:02.169858 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/90181451-368d-48a7-87f3-542be3873484-log-httpd\") pod \"90181451-368d-48a7-87f3-542be3873484\" (UID: \"90181451-368d-48a7-87f3-542be3873484\") " Dec 11 08:32:02 crc kubenswrapper[4860]: I1211 08:32:02.169997 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/90181451-368d-48a7-87f3-542be3873484-sg-core-conf-yaml\") pod \"90181451-368d-48a7-87f3-542be3873484\" (UID: \"90181451-368d-48a7-87f3-542be3873484\") " Dec 11 08:32:02 crc kubenswrapper[4860]: I1211 08:32:02.170104 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90181451-368d-48a7-87f3-542be3873484-combined-ca-bundle\") pod \"90181451-368d-48a7-87f3-542be3873484\" (UID: \"90181451-368d-48a7-87f3-542be3873484\") " Dec 11 08:32:02 crc kubenswrapper[4860]: I1211 08:32:02.170179 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/90181451-368d-48a7-87f3-542be3873484-scripts\") pod \"90181451-368d-48a7-87f3-542be3873484\" (UID: \"90181451-368d-48a7-87f3-542be3873484\") " Dec 11 08:32:02 crc kubenswrapper[4860]: I1211 08:32:02.170236 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/90181451-368d-48a7-87f3-542be3873484-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "90181451-368d-48a7-87f3-542be3873484" (UID: "90181451-368d-48a7-87f3-542be3873484"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:32:02 crc kubenswrapper[4860]: I1211 08:32:02.171232 4860 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/90181451-368d-48a7-87f3-542be3873484-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 11 08:32:02 crc kubenswrapper[4860]: I1211 08:32:02.171282 4860 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/90181451-368d-48a7-87f3-542be3873484-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 11 08:32:02 crc kubenswrapper[4860]: I1211 08:32:02.176573 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/90181451-368d-48a7-87f3-542be3873484-kube-api-access-8qn58" (OuterVolumeSpecName: "kube-api-access-8qn58") pod "90181451-368d-48a7-87f3-542be3873484" (UID: "90181451-368d-48a7-87f3-542be3873484"). InnerVolumeSpecName "kube-api-access-8qn58". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:32:02 crc kubenswrapper[4860]: I1211 08:32:02.177560 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/90181451-368d-48a7-87f3-542be3873484-scripts" (OuterVolumeSpecName: "scripts") pod "90181451-368d-48a7-87f3-542be3873484" (UID: "90181451-368d-48a7-87f3-542be3873484"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:32:02 crc kubenswrapper[4860]: I1211 08:32:02.178722 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/90181451-368d-48a7-87f3-542be3873484-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "90181451-368d-48a7-87f3-542be3873484" (UID: "90181451-368d-48a7-87f3-542be3873484"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:32:02 crc kubenswrapper[4860]: I1211 08:32:02.179841 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/90181451-368d-48a7-87f3-542be3873484-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "90181451-368d-48a7-87f3-542be3873484" (UID: "90181451-368d-48a7-87f3-542be3873484"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:32:02 crc kubenswrapper[4860]: I1211 08:32:02.184148 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/90181451-368d-48a7-87f3-542be3873484-config-data" (OuterVolumeSpecName: "config-data") pod "90181451-368d-48a7-87f3-542be3873484" (UID: "90181451-368d-48a7-87f3-542be3873484"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:32:02 crc kubenswrapper[4860]: I1211 08:32:02.184888 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/90181451-368d-48a7-87f3-542be3873484-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "90181451-368d-48a7-87f3-542be3873484" (UID: "90181451-368d-48a7-87f3-542be3873484"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:32:02 crc kubenswrapper[4860]: I1211 08:32:02.273958 4860 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/90181451-368d-48a7-87f3-542be3873484-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 11 08:32:02 crc kubenswrapper[4860]: I1211 08:32:02.274616 4860 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90181451-368d-48a7-87f3-542be3873484-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 08:32:02 crc kubenswrapper[4860]: I1211 08:32:02.274730 4860 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/90181451-368d-48a7-87f3-542be3873484-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 08:32:02 crc kubenswrapper[4860]: I1211 08:32:02.274809 4860 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/90181451-368d-48a7-87f3-542be3873484-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 08:32:02 crc kubenswrapper[4860]: I1211 08:32:02.274920 4860 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/90181451-368d-48a7-87f3-542be3873484-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 11 08:32:02 crc kubenswrapper[4860]: I1211 08:32:02.275018 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8qn58\" (UniqueName: \"kubernetes.io/projected/90181451-368d-48a7-87f3-542be3873484-kube-api-access-8qn58\") on node \"crc\" DevicePath \"\"" Dec 11 08:32:03 crc kubenswrapper[4860]: I1211 08:32:03.010006 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 11 08:32:03 crc kubenswrapper[4860]: I1211 08:32:03.096951 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 11 08:32:03 crc kubenswrapper[4860]: I1211 08:32:03.116911 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 11 08:32:03 crc kubenswrapper[4860]: I1211 08:32:03.125505 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 11 08:32:03 crc kubenswrapper[4860]: I1211 08:32:03.128178 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 11 08:32:03 crc kubenswrapper[4860]: I1211 08:32:03.132519 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 11 08:32:03 crc kubenswrapper[4860]: I1211 08:32:03.150185 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 11 08:32:03 crc kubenswrapper[4860]: I1211 08:32:03.151922 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 11 08:32:03 crc kubenswrapper[4860]: I1211 08:32:03.152249 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 11 08:32:03 crc kubenswrapper[4860]: I1211 08:32:03.301923 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f0d70280-9b14-4d9d-9b7a-7818f388c4fc-log-httpd\") pod \"ceilometer-0\" (UID: \"f0d70280-9b14-4d9d-9b7a-7818f388c4fc\") " pod="openstack/ceilometer-0" Dec 11 08:32:03 crc kubenswrapper[4860]: I1211 08:32:03.302298 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f0d70280-9b14-4d9d-9b7a-7818f388c4fc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f0d70280-9b14-4d9d-9b7a-7818f388c4fc\") " pod="openstack/ceilometer-0" Dec 11 08:32:03 crc kubenswrapper[4860]: I1211 08:32:03.302344 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f0d70280-9b14-4d9d-9b7a-7818f388c4fc-scripts\") pod \"ceilometer-0\" (UID: \"f0d70280-9b14-4d9d-9b7a-7818f388c4fc\") " pod="openstack/ceilometer-0" Dec 11 08:32:03 crc kubenswrapper[4860]: I1211 08:32:03.302398 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f0d70280-9b14-4d9d-9b7a-7818f388c4fc-run-httpd\") pod \"ceilometer-0\" (UID: \"f0d70280-9b14-4d9d-9b7a-7818f388c4fc\") " pod="openstack/ceilometer-0" Dec 11 08:32:03 crc kubenswrapper[4860]: I1211 08:32:03.302443 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f0d70280-9b14-4d9d-9b7a-7818f388c4fc-config-data\") pod \"ceilometer-0\" (UID: \"f0d70280-9b14-4d9d-9b7a-7818f388c4fc\") " pod="openstack/ceilometer-0" Dec 11 08:32:03 crc kubenswrapper[4860]: I1211 08:32:03.302564 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0d70280-9b14-4d9d-9b7a-7818f388c4fc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f0d70280-9b14-4d9d-9b7a-7818f388c4fc\") " pod="openstack/ceilometer-0" Dec 11 08:32:03 crc kubenswrapper[4860]: I1211 08:32:03.302629 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f0d70280-9b14-4d9d-9b7a-7818f388c4fc-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"f0d70280-9b14-4d9d-9b7a-7818f388c4fc\") " pod="openstack/ceilometer-0" Dec 11 08:32:03 crc kubenswrapper[4860]: I1211 08:32:03.302766 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fqkn8\" (UniqueName: \"kubernetes.io/projected/f0d70280-9b14-4d9d-9b7a-7818f388c4fc-kube-api-access-fqkn8\") pod \"ceilometer-0\" (UID: \"f0d70280-9b14-4d9d-9b7a-7818f388c4fc\") " pod="openstack/ceilometer-0" Dec 11 08:32:03 crc kubenswrapper[4860]: I1211 08:32:03.404607 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0d70280-9b14-4d9d-9b7a-7818f388c4fc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f0d70280-9b14-4d9d-9b7a-7818f388c4fc\") " pod="openstack/ceilometer-0" Dec 11 08:32:03 crc kubenswrapper[4860]: I1211 08:32:03.404677 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f0d70280-9b14-4d9d-9b7a-7818f388c4fc-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"f0d70280-9b14-4d9d-9b7a-7818f388c4fc\") " pod="openstack/ceilometer-0" Dec 11 08:32:03 crc kubenswrapper[4860]: I1211 08:32:03.404740 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fqkn8\" (UniqueName: \"kubernetes.io/projected/f0d70280-9b14-4d9d-9b7a-7818f388c4fc-kube-api-access-fqkn8\") pod \"ceilometer-0\" (UID: \"f0d70280-9b14-4d9d-9b7a-7818f388c4fc\") " pod="openstack/ceilometer-0" Dec 11 08:32:03 crc kubenswrapper[4860]: I1211 08:32:03.404820 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f0d70280-9b14-4d9d-9b7a-7818f388c4fc-log-httpd\") pod \"ceilometer-0\" (UID: \"f0d70280-9b14-4d9d-9b7a-7818f388c4fc\") " pod="openstack/ceilometer-0" Dec 11 08:32:03 crc kubenswrapper[4860]: I1211 08:32:03.404886 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f0d70280-9b14-4d9d-9b7a-7818f388c4fc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f0d70280-9b14-4d9d-9b7a-7818f388c4fc\") " pod="openstack/ceilometer-0" Dec 11 08:32:03 crc kubenswrapper[4860]: I1211 08:32:03.404920 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f0d70280-9b14-4d9d-9b7a-7818f388c4fc-scripts\") pod \"ceilometer-0\" (UID: \"f0d70280-9b14-4d9d-9b7a-7818f388c4fc\") " pod="openstack/ceilometer-0" Dec 11 08:32:03 crc kubenswrapper[4860]: I1211 08:32:03.404966 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f0d70280-9b14-4d9d-9b7a-7818f388c4fc-run-httpd\") pod \"ceilometer-0\" (UID: \"f0d70280-9b14-4d9d-9b7a-7818f388c4fc\") " pod="openstack/ceilometer-0" Dec 11 08:32:03 crc kubenswrapper[4860]: I1211 08:32:03.405020 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f0d70280-9b14-4d9d-9b7a-7818f388c4fc-config-data\") pod \"ceilometer-0\" (UID: \"f0d70280-9b14-4d9d-9b7a-7818f388c4fc\") " pod="openstack/ceilometer-0" Dec 11 08:32:03 crc kubenswrapper[4860]: I1211 08:32:03.405515 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f0d70280-9b14-4d9d-9b7a-7818f388c4fc-run-httpd\") pod \"ceilometer-0\" (UID: \"f0d70280-9b14-4d9d-9b7a-7818f388c4fc\") " pod="openstack/ceilometer-0" Dec 11 08:32:03 crc kubenswrapper[4860]: I1211 08:32:03.405518 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f0d70280-9b14-4d9d-9b7a-7818f388c4fc-log-httpd\") pod \"ceilometer-0\" (UID: \"f0d70280-9b14-4d9d-9b7a-7818f388c4fc\") " pod="openstack/ceilometer-0" Dec 11 08:32:03 crc kubenswrapper[4860]: I1211 08:32:03.413219 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f0d70280-9b14-4d9d-9b7a-7818f388c4fc-scripts\") pod \"ceilometer-0\" (UID: \"f0d70280-9b14-4d9d-9b7a-7818f388c4fc\") " pod="openstack/ceilometer-0" Dec 11 08:32:03 crc kubenswrapper[4860]: I1211 08:32:03.414044 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f0d70280-9b14-4d9d-9b7a-7818f388c4fc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f0d70280-9b14-4d9d-9b7a-7818f388c4fc\") " pod="openstack/ceilometer-0" Dec 11 08:32:03 crc kubenswrapper[4860]: I1211 08:32:03.414929 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f0d70280-9b14-4d9d-9b7a-7818f388c4fc-config-data\") pod \"ceilometer-0\" (UID: \"f0d70280-9b14-4d9d-9b7a-7818f388c4fc\") " pod="openstack/ceilometer-0" Dec 11 08:32:03 crc kubenswrapper[4860]: I1211 08:32:03.415665 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0d70280-9b14-4d9d-9b7a-7818f388c4fc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f0d70280-9b14-4d9d-9b7a-7818f388c4fc\") " pod="openstack/ceilometer-0" Dec 11 08:32:03 crc kubenswrapper[4860]: I1211 08:32:03.416257 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f0d70280-9b14-4d9d-9b7a-7818f388c4fc-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"f0d70280-9b14-4d9d-9b7a-7818f388c4fc\") " pod="openstack/ceilometer-0" Dec 11 08:32:03 crc kubenswrapper[4860]: I1211 08:32:03.434898 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fqkn8\" (UniqueName: \"kubernetes.io/projected/f0d70280-9b14-4d9d-9b7a-7818f388c4fc-kube-api-access-fqkn8\") pod \"ceilometer-0\" (UID: \"f0d70280-9b14-4d9d-9b7a-7818f388c4fc\") " pod="openstack/ceilometer-0" Dec 11 08:32:03 crc kubenswrapper[4860]: I1211 08:32:03.465663 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 11 08:32:03 crc kubenswrapper[4860]: I1211 08:32:03.608590 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="90181451-368d-48a7-87f3-542be3873484" path="/var/lib/kubelet/pods/90181451-368d-48a7-87f3-542be3873484/volumes" Dec 11 08:32:03 crc kubenswrapper[4860]: I1211 08:32:03.639879 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 11 08:32:03 crc kubenswrapper[4860]: I1211 08:32:03.811036 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2714dc33-cc1b-4fed-af2c-9ab58397d1c0-combined-ca-bundle\") pod \"2714dc33-cc1b-4fed-af2c-9ab58397d1c0\" (UID: \"2714dc33-cc1b-4fed-af2c-9ab58397d1c0\") " Dec 11 08:32:03 crc kubenswrapper[4860]: I1211 08:32:03.811463 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2714dc33-cc1b-4fed-af2c-9ab58397d1c0-logs\") pod \"2714dc33-cc1b-4fed-af2c-9ab58397d1c0\" (UID: \"2714dc33-cc1b-4fed-af2c-9ab58397d1c0\") " Dec 11 08:32:03 crc kubenswrapper[4860]: I1211 08:32:03.811580 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xjkdc\" (UniqueName: \"kubernetes.io/projected/2714dc33-cc1b-4fed-af2c-9ab58397d1c0-kube-api-access-xjkdc\") pod \"2714dc33-cc1b-4fed-af2c-9ab58397d1c0\" (UID: \"2714dc33-cc1b-4fed-af2c-9ab58397d1c0\") " Dec 11 08:32:03 crc kubenswrapper[4860]: I1211 08:32:03.811770 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2714dc33-cc1b-4fed-af2c-9ab58397d1c0-config-data\") pod \"2714dc33-cc1b-4fed-af2c-9ab58397d1c0\" (UID: \"2714dc33-cc1b-4fed-af2c-9ab58397d1c0\") " Dec 11 08:32:03 crc kubenswrapper[4860]: I1211 08:32:03.812322 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2714dc33-cc1b-4fed-af2c-9ab58397d1c0-logs" (OuterVolumeSpecName: "logs") pod "2714dc33-cc1b-4fed-af2c-9ab58397d1c0" (UID: "2714dc33-cc1b-4fed-af2c-9ab58397d1c0"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:32:03 crc kubenswrapper[4860]: I1211 08:32:03.817791 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2714dc33-cc1b-4fed-af2c-9ab58397d1c0-kube-api-access-xjkdc" (OuterVolumeSpecName: "kube-api-access-xjkdc") pod "2714dc33-cc1b-4fed-af2c-9ab58397d1c0" (UID: "2714dc33-cc1b-4fed-af2c-9ab58397d1c0"). InnerVolumeSpecName "kube-api-access-xjkdc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:32:03 crc kubenswrapper[4860]: I1211 08:32:03.852867 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2714dc33-cc1b-4fed-af2c-9ab58397d1c0-config-data" (OuterVolumeSpecName: "config-data") pod "2714dc33-cc1b-4fed-af2c-9ab58397d1c0" (UID: "2714dc33-cc1b-4fed-af2c-9ab58397d1c0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:32:03 crc kubenswrapper[4860]: I1211 08:32:03.853388 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2714dc33-cc1b-4fed-af2c-9ab58397d1c0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2714dc33-cc1b-4fed-af2c-9ab58397d1c0" (UID: "2714dc33-cc1b-4fed-af2c-9ab58397d1c0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:32:03 crc kubenswrapper[4860]: I1211 08:32:03.914755 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xjkdc\" (UniqueName: \"kubernetes.io/projected/2714dc33-cc1b-4fed-af2c-9ab58397d1c0-kube-api-access-xjkdc\") on node \"crc\" DevicePath \"\"" Dec 11 08:32:03 crc kubenswrapper[4860]: I1211 08:32:03.914802 4860 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2714dc33-cc1b-4fed-af2c-9ab58397d1c0-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 08:32:03 crc kubenswrapper[4860]: I1211 08:32:03.914832 4860 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2714dc33-cc1b-4fed-af2c-9ab58397d1c0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 08:32:03 crc kubenswrapper[4860]: I1211 08:32:03.914842 4860 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2714dc33-cc1b-4fed-af2c-9ab58397d1c0-logs\") on node \"crc\" DevicePath \"\"" Dec 11 08:32:03 crc kubenswrapper[4860]: I1211 08:32:03.971237 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 11 08:32:04 crc kubenswrapper[4860]: I1211 08:32:04.032681 4860 generic.go:334] "Generic (PLEG): container finished" podID="2714dc33-cc1b-4fed-af2c-9ab58397d1c0" containerID="ec7f65b71fd7e5c584de43621051867f16fc801d9f4efd6258a6b70b740910ca" exitCode=0 Dec 11 08:32:04 crc kubenswrapper[4860]: I1211 08:32:04.032747 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2714dc33-cc1b-4fed-af2c-9ab58397d1c0","Type":"ContainerDied","Data":"ec7f65b71fd7e5c584de43621051867f16fc801d9f4efd6258a6b70b740910ca"} Dec 11 08:32:04 crc kubenswrapper[4860]: I1211 08:32:04.032774 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2714dc33-cc1b-4fed-af2c-9ab58397d1c0","Type":"ContainerDied","Data":"ccf35f6708555af04c9ee1ad842ee9584065fe2cc2491f5483920f9b0be404f0"} Dec 11 08:32:04 crc kubenswrapper[4860]: I1211 08:32:04.032793 4860 scope.go:117] "RemoveContainer" containerID="ec7f65b71fd7e5c584de43621051867f16fc801d9f4efd6258a6b70b740910ca" Dec 11 08:32:04 crc kubenswrapper[4860]: I1211 08:32:04.032920 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 11 08:32:04 crc kubenswrapper[4860]: I1211 08:32:04.038600 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f0d70280-9b14-4d9d-9b7a-7818f388c4fc","Type":"ContainerStarted","Data":"7da0c4d67668b5af5f8b8754541b79883d95589a6034aebfde556a66a9dea47d"} Dec 11 08:32:04 crc kubenswrapper[4860]: I1211 08:32:04.068701 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 11 08:32:04 crc kubenswrapper[4860]: I1211 08:32:04.073717 4860 scope.go:117] "RemoveContainer" containerID="7268c87e4750580d441f2217171d227fd7ee59a77faf3c385af1d37c4272934c" Dec 11 08:32:04 crc kubenswrapper[4860]: I1211 08:32:04.082708 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 11 08:32:04 crc kubenswrapper[4860]: I1211 08:32:04.100269 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 11 08:32:04 crc kubenswrapper[4860]: E1211 08:32:04.100863 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2714dc33-cc1b-4fed-af2c-9ab58397d1c0" containerName="nova-api-log" Dec 11 08:32:04 crc kubenswrapper[4860]: I1211 08:32:04.100890 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="2714dc33-cc1b-4fed-af2c-9ab58397d1c0" containerName="nova-api-log" Dec 11 08:32:04 crc kubenswrapper[4860]: E1211 08:32:04.100930 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2714dc33-cc1b-4fed-af2c-9ab58397d1c0" containerName="nova-api-api" Dec 11 08:32:04 crc kubenswrapper[4860]: I1211 08:32:04.100939 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="2714dc33-cc1b-4fed-af2c-9ab58397d1c0" containerName="nova-api-api" Dec 11 08:32:04 crc kubenswrapper[4860]: I1211 08:32:04.101170 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="2714dc33-cc1b-4fed-af2c-9ab58397d1c0" containerName="nova-api-api" Dec 11 08:32:04 crc kubenswrapper[4860]: I1211 08:32:04.101193 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="2714dc33-cc1b-4fed-af2c-9ab58397d1c0" containerName="nova-api-log" Dec 11 08:32:04 crc kubenswrapper[4860]: I1211 08:32:04.102541 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 11 08:32:04 crc kubenswrapper[4860]: I1211 08:32:04.106749 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 11 08:32:04 crc kubenswrapper[4860]: I1211 08:32:04.107001 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 11 08:32:04 crc kubenswrapper[4860]: I1211 08:32:04.107328 4860 scope.go:117] "RemoveContainer" containerID="ec7f65b71fd7e5c584de43621051867f16fc801d9f4efd6258a6b70b740910ca" Dec 11 08:32:04 crc kubenswrapper[4860]: I1211 08:32:04.109160 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 11 08:32:04 crc kubenswrapper[4860]: E1211 08:32:04.109473 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec7f65b71fd7e5c584de43621051867f16fc801d9f4efd6258a6b70b740910ca\": container with ID starting with ec7f65b71fd7e5c584de43621051867f16fc801d9f4efd6258a6b70b740910ca not found: ID does not exist" containerID="ec7f65b71fd7e5c584de43621051867f16fc801d9f4efd6258a6b70b740910ca" Dec 11 08:32:04 crc kubenswrapper[4860]: I1211 08:32:04.109505 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec7f65b71fd7e5c584de43621051867f16fc801d9f4efd6258a6b70b740910ca"} err="failed to get container status \"ec7f65b71fd7e5c584de43621051867f16fc801d9f4efd6258a6b70b740910ca\": rpc error: code = NotFound desc = could not find container \"ec7f65b71fd7e5c584de43621051867f16fc801d9f4efd6258a6b70b740910ca\": container with ID starting with ec7f65b71fd7e5c584de43621051867f16fc801d9f4efd6258a6b70b740910ca not found: ID does not exist" Dec 11 08:32:04 crc kubenswrapper[4860]: I1211 08:32:04.109534 4860 scope.go:117] "RemoveContainer" containerID="7268c87e4750580d441f2217171d227fd7ee59a77faf3c385af1d37c4272934c" Dec 11 08:32:04 crc kubenswrapper[4860]: I1211 08:32:04.109838 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 11 08:32:04 crc kubenswrapper[4860]: E1211 08:32:04.109992 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7268c87e4750580d441f2217171d227fd7ee59a77faf3c385af1d37c4272934c\": container with ID starting with 7268c87e4750580d441f2217171d227fd7ee59a77faf3c385af1d37c4272934c not found: ID does not exist" containerID="7268c87e4750580d441f2217171d227fd7ee59a77faf3c385af1d37c4272934c" Dec 11 08:32:04 crc kubenswrapper[4860]: I1211 08:32:04.110018 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7268c87e4750580d441f2217171d227fd7ee59a77faf3c385af1d37c4272934c"} err="failed to get container status \"7268c87e4750580d441f2217171d227fd7ee59a77faf3c385af1d37c4272934c\": rpc error: code = NotFound desc = could not find container \"7268c87e4750580d441f2217171d227fd7ee59a77faf3c385af1d37c4272934c\": container with ID starting with 7268c87e4750580d441f2217171d227fd7ee59a77faf3c385af1d37c4272934c not found: ID does not exist" Dec 11 08:32:04 crc kubenswrapper[4860]: I1211 08:32:04.223956 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/18c09963-6bc7-4dfc-8dfb-a89618d2466d-logs\") pod \"nova-api-0\" (UID: \"18c09963-6bc7-4dfc-8dfb-a89618d2466d\") " pod="openstack/nova-api-0" Dec 11 08:32:04 crc kubenswrapper[4860]: I1211 08:32:04.224319 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b5dfk\" (UniqueName: \"kubernetes.io/projected/18c09963-6bc7-4dfc-8dfb-a89618d2466d-kube-api-access-b5dfk\") pod \"nova-api-0\" (UID: \"18c09963-6bc7-4dfc-8dfb-a89618d2466d\") " pod="openstack/nova-api-0" Dec 11 08:32:04 crc kubenswrapper[4860]: I1211 08:32:04.224451 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/18c09963-6bc7-4dfc-8dfb-a89618d2466d-internal-tls-certs\") pod \"nova-api-0\" (UID: \"18c09963-6bc7-4dfc-8dfb-a89618d2466d\") " pod="openstack/nova-api-0" Dec 11 08:32:04 crc kubenswrapper[4860]: I1211 08:32:04.224522 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18c09963-6bc7-4dfc-8dfb-a89618d2466d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"18c09963-6bc7-4dfc-8dfb-a89618d2466d\") " pod="openstack/nova-api-0" Dec 11 08:32:04 crc kubenswrapper[4860]: I1211 08:32:04.224601 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/18c09963-6bc7-4dfc-8dfb-a89618d2466d-public-tls-certs\") pod \"nova-api-0\" (UID: \"18c09963-6bc7-4dfc-8dfb-a89618d2466d\") " pod="openstack/nova-api-0" Dec 11 08:32:04 crc kubenswrapper[4860]: I1211 08:32:04.224723 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18c09963-6bc7-4dfc-8dfb-a89618d2466d-config-data\") pod \"nova-api-0\" (UID: \"18c09963-6bc7-4dfc-8dfb-a89618d2466d\") " pod="openstack/nova-api-0" Dec 11 08:32:04 crc kubenswrapper[4860]: I1211 08:32:04.326909 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/18c09963-6bc7-4dfc-8dfb-a89618d2466d-internal-tls-certs\") pod \"nova-api-0\" (UID: \"18c09963-6bc7-4dfc-8dfb-a89618d2466d\") " pod="openstack/nova-api-0" Dec 11 08:32:04 crc kubenswrapper[4860]: I1211 08:32:04.326967 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18c09963-6bc7-4dfc-8dfb-a89618d2466d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"18c09963-6bc7-4dfc-8dfb-a89618d2466d\") " pod="openstack/nova-api-0" Dec 11 08:32:04 crc kubenswrapper[4860]: I1211 08:32:04.326999 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/18c09963-6bc7-4dfc-8dfb-a89618d2466d-public-tls-certs\") pod \"nova-api-0\" (UID: \"18c09963-6bc7-4dfc-8dfb-a89618d2466d\") " pod="openstack/nova-api-0" Dec 11 08:32:04 crc kubenswrapper[4860]: I1211 08:32:04.327075 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18c09963-6bc7-4dfc-8dfb-a89618d2466d-config-data\") pod \"nova-api-0\" (UID: \"18c09963-6bc7-4dfc-8dfb-a89618d2466d\") " pod="openstack/nova-api-0" Dec 11 08:32:04 crc kubenswrapper[4860]: I1211 08:32:04.327135 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/18c09963-6bc7-4dfc-8dfb-a89618d2466d-logs\") pod \"nova-api-0\" (UID: \"18c09963-6bc7-4dfc-8dfb-a89618d2466d\") " pod="openstack/nova-api-0" Dec 11 08:32:04 crc kubenswrapper[4860]: I1211 08:32:04.327216 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b5dfk\" (UniqueName: \"kubernetes.io/projected/18c09963-6bc7-4dfc-8dfb-a89618d2466d-kube-api-access-b5dfk\") pod \"nova-api-0\" (UID: \"18c09963-6bc7-4dfc-8dfb-a89618d2466d\") " pod="openstack/nova-api-0" Dec 11 08:32:04 crc kubenswrapper[4860]: I1211 08:32:04.327987 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/18c09963-6bc7-4dfc-8dfb-a89618d2466d-logs\") pod \"nova-api-0\" (UID: \"18c09963-6bc7-4dfc-8dfb-a89618d2466d\") " pod="openstack/nova-api-0" Dec 11 08:32:04 crc kubenswrapper[4860]: I1211 08:32:04.333717 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/18c09963-6bc7-4dfc-8dfb-a89618d2466d-public-tls-certs\") pod \"nova-api-0\" (UID: \"18c09963-6bc7-4dfc-8dfb-a89618d2466d\") " pod="openstack/nova-api-0" Dec 11 08:32:04 crc kubenswrapper[4860]: I1211 08:32:04.334383 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18c09963-6bc7-4dfc-8dfb-a89618d2466d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"18c09963-6bc7-4dfc-8dfb-a89618d2466d\") " pod="openstack/nova-api-0" Dec 11 08:32:04 crc kubenswrapper[4860]: I1211 08:32:04.335098 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18c09963-6bc7-4dfc-8dfb-a89618d2466d-config-data\") pod \"nova-api-0\" (UID: \"18c09963-6bc7-4dfc-8dfb-a89618d2466d\") " pod="openstack/nova-api-0" Dec 11 08:32:04 crc kubenswrapper[4860]: I1211 08:32:04.340390 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/18c09963-6bc7-4dfc-8dfb-a89618d2466d-internal-tls-certs\") pod \"nova-api-0\" (UID: \"18c09963-6bc7-4dfc-8dfb-a89618d2466d\") " pod="openstack/nova-api-0" Dec 11 08:32:04 crc kubenswrapper[4860]: I1211 08:32:04.370279 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b5dfk\" (UniqueName: \"kubernetes.io/projected/18c09963-6bc7-4dfc-8dfb-a89618d2466d-kube-api-access-b5dfk\") pod \"nova-api-0\" (UID: \"18c09963-6bc7-4dfc-8dfb-a89618d2466d\") " pod="openstack/nova-api-0" Dec 11 08:32:04 crc kubenswrapper[4860]: I1211 08:32:04.444276 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 11 08:32:04 crc kubenswrapper[4860]: I1211 08:32:04.623695 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Dec 11 08:32:04 crc kubenswrapper[4860]: I1211 08:32:04.662467 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Dec 11 08:32:04 crc kubenswrapper[4860]: I1211 08:32:04.941314 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 11 08:32:04 crc kubenswrapper[4860]: W1211 08:32:04.960090 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod18c09963_6bc7_4dfc_8dfb_a89618d2466d.slice/crio-3ccee3116d0fb682a575b9030084970f24dadb839a1a2703928fe2f69cb846d9 WatchSource:0}: Error finding container 3ccee3116d0fb682a575b9030084970f24dadb839a1a2703928fe2f69cb846d9: Status 404 returned error can't find the container with id 3ccee3116d0fb682a575b9030084970f24dadb839a1a2703928fe2f69cb846d9 Dec 11 08:32:05 crc kubenswrapper[4860]: I1211 08:32:05.063519 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"18c09963-6bc7-4dfc-8dfb-a89618d2466d","Type":"ContainerStarted","Data":"3ccee3116d0fb682a575b9030084970f24dadb839a1a2703928fe2f69cb846d9"} Dec 11 08:32:05 crc kubenswrapper[4860]: I1211 08:32:05.087948 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Dec 11 08:32:05 crc kubenswrapper[4860]: I1211 08:32:05.252263 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-7nvzj"] Dec 11 08:32:05 crc kubenswrapper[4860]: I1211 08:32:05.254100 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-7nvzj" Dec 11 08:32:05 crc kubenswrapper[4860]: I1211 08:32:05.256512 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Dec 11 08:32:05 crc kubenswrapper[4860]: I1211 08:32:05.256989 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Dec 11 08:32:05 crc kubenswrapper[4860]: I1211 08:32:05.283397 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-7nvzj"] Dec 11 08:32:05 crc kubenswrapper[4860]: I1211 08:32:05.349422 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z579q\" (UniqueName: \"kubernetes.io/projected/d07f5ac6-54a5-4490-a6b4-09fe7325e58a-kube-api-access-z579q\") pod \"nova-cell1-cell-mapping-7nvzj\" (UID: \"d07f5ac6-54a5-4490-a6b4-09fe7325e58a\") " pod="openstack/nova-cell1-cell-mapping-7nvzj" Dec 11 08:32:05 crc kubenswrapper[4860]: I1211 08:32:05.349502 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d07f5ac6-54a5-4490-a6b4-09fe7325e58a-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-7nvzj\" (UID: \"d07f5ac6-54a5-4490-a6b4-09fe7325e58a\") " pod="openstack/nova-cell1-cell-mapping-7nvzj" Dec 11 08:32:05 crc kubenswrapper[4860]: I1211 08:32:05.349530 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d07f5ac6-54a5-4490-a6b4-09fe7325e58a-scripts\") pod \"nova-cell1-cell-mapping-7nvzj\" (UID: \"d07f5ac6-54a5-4490-a6b4-09fe7325e58a\") " pod="openstack/nova-cell1-cell-mapping-7nvzj" Dec 11 08:32:05 crc kubenswrapper[4860]: I1211 08:32:05.349576 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d07f5ac6-54a5-4490-a6b4-09fe7325e58a-config-data\") pod \"nova-cell1-cell-mapping-7nvzj\" (UID: \"d07f5ac6-54a5-4490-a6b4-09fe7325e58a\") " pod="openstack/nova-cell1-cell-mapping-7nvzj" Dec 11 08:32:05 crc kubenswrapper[4860]: I1211 08:32:05.452027 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z579q\" (UniqueName: \"kubernetes.io/projected/d07f5ac6-54a5-4490-a6b4-09fe7325e58a-kube-api-access-z579q\") pod \"nova-cell1-cell-mapping-7nvzj\" (UID: \"d07f5ac6-54a5-4490-a6b4-09fe7325e58a\") " pod="openstack/nova-cell1-cell-mapping-7nvzj" Dec 11 08:32:05 crc kubenswrapper[4860]: I1211 08:32:05.452135 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d07f5ac6-54a5-4490-a6b4-09fe7325e58a-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-7nvzj\" (UID: \"d07f5ac6-54a5-4490-a6b4-09fe7325e58a\") " pod="openstack/nova-cell1-cell-mapping-7nvzj" Dec 11 08:32:05 crc kubenswrapper[4860]: I1211 08:32:05.452169 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d07f5ac6-54a5-4490-a6b4-09fe7325e58a-scripts\") pod \"nova-cell1-cell-mapping-7nvzj\" (UID: \"d07f5ac6-54a5-4490-a6b4-09fe7325e58a\") " pod="openstack/nova-cell1-cell-mapping-7nvzj" Dec 11 08:32:05 crc kubenswrapper[4860]: I1211 08:32:05.452208 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d07f5ac6-54a5-4490-a6b4-09fe7325e58a-config-data\") pod \"nova-cell1-cell-mapping-7nvzj\" (UID: \"d07f5ac6-54a5-4490-a6b4-09fe7325e58a\") " pod="openstack/nova-cell1-cell-mapping-7nvzj" Dec 11 08:32:05 crc kubenswrapper[4860]: I1211 08:32:05.456374 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d07f5ac6-54a5-4490-a6b4-09fe7325e58a-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-7nvzj\" (UID: \"d07f5ac6-54a5-4490-a6b4-09fe7325e58a\") " pod="openstack/nova-cell1-cell-mapping-7nvzj" Dec 11 08:32:05 crc kubenswrapper[4860]: I1211 08:32:05.456440 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d07f5ac6-54a5-4490-a6b4-09fe7325e58a-config-data\") pod \"nova-cell1-cell-mapping-7nvzj\" (UID: \"d07f5ac6-54a5-4490-a6b4-09fe7325e58a\") " pod="openstack/nova-cell1-cell-mapping-7nvzj" Dec 11 08:32:05 crc kubenswrapper[4860]: I1211 08:32:05.458865 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d07f5ac6-54a5-4490-a6b4-09fe7325e58a-scripts\") pod \"nova-cell1-cell-mapping-7nvzj\" (UID: \"d07f5ac6-54a5-4490-a6b4-09fe7325e58a\") " pod="openstack/nova-cell1-cell-mapping-7nvzj" Dec 11 08:32:05 crc kubenswrapper[4860]: I1211 08:32:05.475389 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z579q\" (UniqueName: \"kubernetes.io/projected/d07f5ac6-54a5-4490-a6b4-09fe7325e58a-kube-api-access-z579q\") pod \"nova-cell1-cell-mapping-7nvzj\" (UID: \"d07f5ac6-54a5-4490-a6b4-09fe7325e58a\") " pod="openstack/nova-cell1-cell-mapping-7nvzj" Dec 11 08:32:05 crc kubenswrapper[4860]: I1211 08:32:05.594770 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2714dc33-cc1b-4fed-af2c-9ab58397d1c0" path="/var/lib/kubelet/pods/2714dc33-cc1b-4fed-af2c-9ab58397d1c0/volumes" Dec 11 08:32:05 crc kubenswrapper[4860]: I1211 08:32:05.659520 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-7nvzj" Dec 11 08:32:06 crc kubenswrapper[4860]: I1211 08:32:06.076937 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f0d70280-9b14-4d9d-9b7a-7818f388c4fc","Type":"ContainerStarted","Data":"bdc6c0a636d25726841bdc45f0160b5b18829ed32392c502cc1b6e155a391fb9"} Dec 11 08:32:06 crc kubenswrapper[4860]: I1211 08:32:06.077562 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f0d70280-9b14-4d9d-9b7a-7818f388c4fc","Type":"ContainerStarted","Data":"59d1df1d58b8b1e40e04c0c3d5b852f6a0c9ce416372c4bdc674cf7993ab9c14"} Dec 11 08:32:06 crc kubenswrapper[4860]: I1211 08:32:06.079415 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"18c09963-6bc7-4dfc-8dfb-a89618d2466d","Type":"ContainerStarted","Data":"51704a83c489f20401c73b4ba7fcd751bdfba6a32adc3d18ff41ec3553e4816a"} Dec 11 08:32:06 crc kubenswrapper[4860]: I1211 08:32:06.079471 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"18c09963-6bc7-4dfc-8dfb-a89618d2466d","Type":"ContainerStarted","Data":"352c4fa0ec3a1568584aa36cd6e6fb79dc5b916be01b6a2ef65dc7b6f1662dea"} Dec 11 08:32:06 crc kubenswrapper[4860]: I1211 08:32:06.115897 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.115789919 podStartE2EDuration="2.115789919s" podCreationTimestamp="2025-12-11 08:32:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:32:06.106415572 +0000 UTC m=+1258.834934627" watchObservedRunningTime="2025-12-11 08:32:06.115789919 +0000 UTC m=+1258.844308974" Dec 11 08:32:06 crc kubenswrapper[4860]: I1211 08:32:06.182953 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-7nvzj"] Dec 11 08:32:06 crc kubenswrapper[4860]: W1211 08:32:06.183767 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd07f5ac6_54a5_4490_a6b4_09fe7325e58a.slice/crio-9b632d13e89d5bf8fdd1dc9aa6ae8c02ff10d25234a3bde2635b2b84ec74a692 WatchSource:0}: Error finding container 9b632d13e89d5bf8fdd1dc9aa6ae8c02ff10d25234a3bde2635b2b84ec74a692: Status 404 returned error can't find the container with id 9b632d13e89d5bf8fdd1dc9aa6ae8c02ff10d25234a3bde2635b2b84ec74a692 Dec 11 08:32:07 crc kubenswrapper[4860]: I1211 08:32:07.099304 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f0d70280-9b14-4d9d-9b7a-7818f388c4fc","Type":"ContainerStarted","Data":"958ced138595925d27815f9ecd68bb13113d35b2f0c24c4369b0d400f43b8013"} Dec 11 08:32:07 crc kubenswrapper[4860]: I1211 08:32:07.100968 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-7nvzj" event={"ID":"d07f5ac6-54a5-4490-a6b4-09fe7325e58a","Type":"ContainerStarted","Data":"33f54983234fbd4f8d8f95d07373fa5a12a595642d29f4c19fb280a7e225d5ab"} Dec 11 08:32:07 crc kubenswrapper[4860]: I1211 08:32:07.101018 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-7nvzj" event={"ID":"d07f5ac6-54a5-4490-a6b4-09fe7325e58a","Type":"ContainerStarted","Data":"9b632d13e89d5bf8fdd1dc9aa6ae8c02ff10d25234a3bde2635b2b84ec74a692"} Dec 11 08:32:07 crc kubenswrapper[4860]: I1211 08:32:07.124222 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-7nvzj" podStartSLOduration=2.124188362 podStartE2EDuration="2.124188362s" podCreationTimestamp="2025-12-11 08:32:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:32:07.118580625 +0000 UTC m=+1259.847099680" watchObservedRunningTime="2025-12-11 08:32:07.124188362 +0000 UTC m=+1259.852707417" Dec 11 08:32:07 crc kubenswrapper[4860]: I1211 08:32:07.499395 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5c7b6c5df9-7zcv8" Dec 11 08:32:07 crc kubenswrapper[4860]: I1211 08:32:07.635169 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-865f5d856f-wz6cz"] Dec 11 08:32:07 crc kubenswrapper[4860]: I1211 08:32:07.635757 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-865f5d856f-wz6cz" podUID="7e338f34-d693-481b-a8ed-eb8ecbdd5a3b" containerName="dnsmasq-dns" containerID="cri-o://b291a77f964230ad3604e0f8070421f9d71a6daf17b5e1350871337ab4bd9f19" gracePeriod=10 Dec 11 08:32:08 crc kubenswrapper[4860]: I1211 08:32:08.113535 4860 generic.go:334] "Generic (PLEG): container finished" podID="7e338f34-d693-481b-a8ed-eb8ecbdd5a3b" containerID="b291a77f964230ad3604e0f8070421f9d71a6daf17b5e1350871337ab4bd9f19" exitCode=0 Dec 11 08:32:08 crc kubenswrapper[4860]: I1211 08:32:08.113585 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-865f5d856f-wz6cz" event={"ID":"7e338f34-d693-481b-a8ed-eb8ecbdd5a3b","Type":"ContainerDied","Data":"b291a77f964230ad3604e0f8070421f9d71a6daf17b5e1350871337ab4bd9f19"} Dec 11 08:32:08 crc kubenswrapper[4860]: I1211 08:32:08.113975 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-865f5d856f-wz6cz" event={"ID":"7e338f34-d693-481b-a8ed-eb8ecbdd5a3b","Type":"ContainerDied","Data":"33f05c5474bbc667f0539ad9f823dea04ac0f1a4e75a4e55f7e0c4f5c172a2ed"} Dec 11 08:32:08 crc kubenswrapper[4860]: I1211 08:32:08.114008 4860 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="33f05c5474bbc667f0539ad9f823dea04ac0f1a4e75a4e55f7e0c4f5c172a2ed" Dec 11 08:32:08 crc kubenswrapper[4860]: I1211 08:32:08.320072 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-865f5d856f-wz6cz" Dec 11 08:32:08 crc kubenswrapper[4860]: I1211 08:32:08.421691 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4c9vd\" (UniqueName: \"kubernetes.io/projected/7e338f34-d693-481b-a8ed-eb8ecbdd5a3b-kube-api-access-4c9vd\") pod \"7e338f34-d693-481b-a8ed-eb8ecbdd5a3b\" (UID: \"7e338f34-d693-481b-a8ed-eb8ecbdd5a3b\") " Dec 11 08:32:08 crc kubenswrapper[4860]: I1211 08:32:08.422104 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7e338f34-d693-481b-a8ed-eb8ecbdd5a3b-ovsdbserver-sb\") pod \"7e338f34-d693-481b-a8ed-eb8ecbdd5a3b\" (UID: \"7e338f34-d693-481b-a8ed-eb8ecbdd5a3b\") " Dec 11 08:32:08 crc kubenswrapper[4860]: I1211 08:32:08.422206 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7e338f34-d693-481b-a8ed-eb8ecbdd5a3b-ovsdbserver-nb\") pod \"7e338f34-d693-481b-a8ed-eb8ecbdd5a3b\" (UID: \"7e338f34-d693-481b-a8ed-eb8ecbdd5a3b\") " Dec 11 08:32:08 crc kubenswrapper[4860]: I1211 08:32:08.422278 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7e338f34-d693-481b-a8ed-eb8ecbdd5a3b-dns-svc\") pod \"7e338f34-d693-481b-a8ed-eb8ecbdd5a3b\" (UID: \"7e338f34-d693-481b-a8ed-eb8ecbdd5a3b\") " Dec 11 08:32:08 crc kubenswrapper[4860]: I1211 08:32:08.422328 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e338f34-d693-481b-a8ed-eb8ecbdd5a3b-config\") pod \"7e338f34-d693-481b-a8ed-eb8ecbdd5a3b\" (UID: \"7e338f34-d693-481b-a8ed-eb8ecbdd5a3b\") " Dec 11 08:32:08 crc kubenswrapper[4860]: I1211 08:32:08.422428 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7e338f34-d693-481b-a8ed-eb8ecbdd5a3b-dns-swift-storage-0\") pod \"7e338f34-d693-481b-a8ed-eb8ecbdd5a3b\" (UID: \"7e338f34-d693-481b-a8ed-eb8ecbdd5a3b\") " Dec 11 08:32:08 crc kubenswrapper[4860]: I1211 08:32:08.428524 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e338f34-d693-481b-a8ed-eb8ecbdd5a3b-kube-api-access-4c9vd" (OuterVolumeSpecName: "kube-api-access-4c9vd") pod "7e338f34-d693-481b-a8ed-eb8ecbdd5a3b" (UID: "7e338f34-d693-481b-a8ed-eb8ecbdd5a3b"). InnerVolumeSpecName "kube-api-access-4c9vd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:32:08 crc kubenswrapper[4860]: I1211 08:32:08.481991 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7e338f34-d693-481b-a8ed-eb8ecbdd5a3b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "7e338f34-d693-481b-a8ed-eb8ecbdd5a3b" (UID: "7e338f34-d693-481b-a8ed-eb8ecbdd5a3b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:32:08 crc kubenswrapper[4860]: I1211 08:32:08.489144 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7e338f34-d693-481b-a8ed-eb8ecbdd5a3b-config" (OuterVolumeSpecName: "config") pod "7e338f34-d693-481b-a8ed-eb8ecbdd5a3b" (UID: "7e338f34-d693-481b-a8ed-eb8ecbdd5a3b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:32:08 crc kubenswrapper[4860]: I1211 08:32:08.504122 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7e338f34-d693-481b-a8ed-eb8ecbdd5a3b-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "7e338f34-d693-481b-a8ed-eb8ecbdd5a3b" (UID: "7e338f34-d693-481b-a8ed-eb8ecbdd5a3b"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:32:08 crc kubenswrapper[4860]: I1211 08:32:08.513257 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7e338f34-d693-481b-a8ed-eb8ecbdd5a3b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7e338f34-d693-481b-a8ed-eb8ecbdd5a3b" (UID: "7e338f34-d693-481b-a8ed-eb8ecbdd5a3b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:32:08 crc kubenswrapper[4860]: I1211 08:32:08.523630 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7e338f34-d693-481b-a8ed-eb8ecbdd5a3b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7e338f34-d693-481b-a8ed-eb8ecbdd5a3b" (UID: "7e338f34-d693-481b-a8ed-eb8ecbdd5a3b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:32:08 crc kubenswrapper[4860]: I1211 08:32:08.525607 4860 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7e338f34-d693-481b-a8ed-eb8ecbdd5a3b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 11 08:32:08 crc kubenswrapper[4860]: I1211 08:32:08.525653 4860 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7e338f34-d693-481b-a8ed-eb8ecbdd5a3b-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 11 08:32:08 crc kubenswrapper[4860]: I1211 08:32:08.525668 4860 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e338f34-d693-481b-a8ed-eb8ecbdd5a3b-config\") on node \"crc\" DevicePath \"\"" Dec 11 08:32:08 crc kubenswrapper[4860]: I1211 08:32:08.525681 4860 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7e338f34-d693-481b-a8ed-eb8ecbdd5a3b-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 11 08:32:08 crc kubenswrapper[4860]: I1211 08:32:08.525694 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4c9vd\" (UniqueName: \"kubernetes.io/projected/7e338f34-d693-481b-a8ed-eb8ecbdd5a3b-kube-api-access-4c9vd\") on node \"crc\" DevicePath \"\"" Dec 11 08:32:08 crc kubenswrapper[4860]: I1211 08:32:08.525708 4860 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7e338f34-d693-481b-a8ed-eb8ecbdd5a3b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 11 08:32:08 crc kubenswrapper[4860]: I1211 08:32:08.795013 4860 patch_prober.go:28] interesting pod/machine-config-daemon-99qgp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 08:32:08 crc kubenswrapper[4860]: I1211 08:32:08.795085 4860 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 08:32:08 crc kubenswrapper[4860]: I1211 08:32:08.795402 4860 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" Dec 11 08:32:08 crc kubenswrapper[4860]: I1211 08:32:08.796286 4860 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6be7e2592a74fa92891c9f53dda9c3fdad0d78e647aef9f0b578adb34ea14caa"} pod="openshift-machine-config-operator/machine-config-daemon-99qgp" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 11 08:32:08 crc kubenswrapper[4860]: I1211 08:32:08.796370 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" containerName="machine-config-daemon" containerID="cri-o://6be7e2592a74fa92891c9f53dda9c3fdad0d78e647aef9f0b578adb34ea14caa" gracePeriod=600 Dec 11 08:32:09 crc kubenswrapper[4860]: I1211 08:32:09.127199 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f0d70280-9b14-4d9d-9b7a-7818f388c4fc","Type":"ContainerStarted","Data":"f67ac7ec73d48dd3d144002da0467253f2490a7110ae1a6f6a75dae5ef28b534"} Dec 11 08:32:09 crc kubenswrapper[4860]: I1211 08:32:09.127670 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 11 08:32:09 crc kubenswrapper[4860]: I1211 08:32:09.130898 4860 generic.go:334] "Generic (PLEG): container finished" podID="31c30642-6e60-41b4-a477-0d802424e0aa" containerID="6be7e2592a74fa92891c9f53dda9c3fdad0d78e647aef9f0b578adb34ea14caa" exitCode=0 Dec 11 08:32:09 crc kubenswrapper[4860]: I1211 08:32:09.130987 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-865f5d856f-wz6cz" Dec 11 08:32:09 crc kubenswrapper[4860]: I1211 08:32:09.130982 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" event={"ID":"31c30642-6e60-41b4-a477-0d802424e0aa","Type":"ContainerDied","Data":"6be7e2592a74fa92891c9f53dda9c3fdad0d78e647aef9f0b578adb34ea14caa"} Dec 11 08:32:09 crc kubenswrapper[4860]: I1211 08:32:09.131038 4860 scope.go:117] "RemoveContainer" containerID="565fbddca3ff6a011767df936f9e699e4560197af6e486d467234b4599b2d2f6" Dec 11 08:32:09 crc kubenswrapper[4860]: I1211 08:32:09.163677 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.997524495 podStartE2EDuration="6.163628444s" podCreationTimestamp="2025-12-11 08:32:03 +0000 UTC" firstStartedPulling="2025-12-11 08:32:03.976113801 +0000 UTC m=+1256.704632856" lastFinishedPulling="2025-12-11 08:32:08.14221775 +0000 UTC m=+1260.870736805" observedRunningTime="2025-12-11 08:32:09.150022414 +0000 UTC m=+1261.878541469" watchObservedRunningTime="2025-12-11 08:32:09.163628444 +0000 UTC m=+1261.892147499" Dec 11 08:32:09 crc kubenswrapper[4860]: I1211 08:32:09.181320 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-865f5d856f-wz6cz"] Dec 11 08:32:09 crc kubenswrapper[4860]: I1211 08:32:09.190683 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-865f5d856f-wz6cz"] Dec 11 08:32:09 crc kubenswrapper[4860]: I1211 08:32:09.623526 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7e338f34-d693-481b-a8ed-eb8ecbdd5a3b" path="/var/lib/kubelet/pods/7e338f34-d693-481b-a8ed-eb8ecbdd5a3b/volumes" Dec 11 08:32:10 crc kubenswrapper[4860]: I1211 08:32:10.146810 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" event={"ID":"31c30642-6e60-41b4-a477-0d802424e0aa","Type":"ContainerStarted","Data":"2c5567d203277d4236f1f4cd1d73721ecd8809240167b07eac4086db3323fa9f"} Dec 11 08:32:12 crc kubenswrapper[4860]: I1211 08:32:12.174497 4860 generic.go:334] "Generic (PLEG): container finished" podID="d07f5ac6-54a5-4490-a6b4-09fe7325e58a" containerID="33f54983234fbd4f8d8f95d07373fa5a12a595642d29f4c19fb280a7e225d5ab" exitCode=0 Dec 11 08:32:12 crc kubenswrapper[4860]: I1211 08:32:12.174562 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-7nvzj" event={"ID":"d07f5ac6-54a5-4490-a6b4-09fe7325e58a","Type":"ContainerDied","Data":"33f54983234fbd4f8d8f95d07373fa5a12a595642d29f4c19fb280a7e225d5ab"} Dec 11 08:32:13 crc kubenswrapper[4860]: I1211 08:32:13.218830 4860 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-865f5d856f-wz6cz" podUID="7e338f34-d693-481b-a8ed-eb8ecbdd5a3b" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.187:5353: i/o timeout" Dec 11 08:32:13 crc kubenswrapper[4860]: I1211 08:32:13.573750 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-7nvzj" Dec 11 08:32:13 crc kubenswrapper[4860]: I1211 08:32:13.649454 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d07f5ac6-54a5-4490-a6b4-09fe7325e58a-scripts\") pod \"d07f5ac6-54a5-4490-a6b4-09fe7325e58a\" (UID: \"d07f5ac6-54a5-4490-a6b4-09fe7325e58a\") " Dec 11 08:32:13 crc kubenswrapper[4860]: I1211 08:32:13.649536 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d07f5ac6-54a5-4490-a6b4-09fe7325e58a-combined-ca-bundle\") pod \"d07f5ac6-54a5-4490-a6b4-09fe7325e58a\" (UID: \"d07f5ac6-54a5-4490-a6b4-09fe7325e58a\") " Dec 11 08:32:13 crc kubenswrapper[4860]: I1211 08:32:13.649878 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d07f5ac6-54a5-4490-a6b4-09fe7325e58a-config-data\") pod \"d07f5ac6-54a5-4490-a6b4-09fe7325e58a\" (UID: \"d07f5ac6-54a5-4490-a6b4-09fe7325e58a\") " Dec 11 08:32:13 crc kubenswrapper[4860]: I1211 08:32:13.649936 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z579q\" (UniqueName: \"kubernetes.io/projected/d07f5ac6-54a5-4490-a6b4-09fe7325e58a-kube-api-access-z579q\") pod \"d07f5ac6-54a5-4490-a6b4-09fe7325e58a\" (UID: \"d07f5ac6-54a5-4490-a6b4-09fe7325e58a\") " Dec 11 08:32:13 crc kubenswrapper[4860]: I1211 08:32:13.655921 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d07f5ac6-54a5-4490-a6b4-09fe7325e58a-kube-api-access-z579q" (OuterVolumeSpecName: "kube-api-access-z579q") pod "d07f5ac6-54a5-4490-a6b4-09fe7325e58a" (UID: "d07f5ac6-54a5-4490-a6b4-09fe7325e58a"). InnerVolumeSpecName "kube-api-access-z579q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:32:13 crc kubenswrapper[4860]: I1211 08:32:13.656005 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d07f5ac6-54a5-4490-a6b4-09fe7325e58a-scripts" (OuterVolumeSpecName: "scripts") pod "d07f5ac6-54a5-4490-a6b4-09fe7325e58a" (UID: "d07f5ac6-54a5-4490-a6b4-09fe7325e58a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:32:13 crc kubenswrapper[4860]: I1211 08:32:13.677577 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d07f5ac6-54a5-4490-a6b4-09fe7325e58a-config-data" (OuterVolumeSpecName: "config-data") pod "d07f5ac6-54a5-4490-a6b4-09fe7325e58a" (UID: "d07f5ac6-54a5-4490-a6b4-09fe7325e58a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:32:13 crc kubenswrapper[4860]: I1211 08:32:13.680319 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d07f5ac6-54a5-4490-a6b4-09fe7325e58a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d07f5ac6-54a5-4490-a6b4-09fe7325e58a" (UID: "d07f5ac6-54a5-4490-a6b4-09fe7325e58a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:32:13 crc kubenswrapper[4860]: I1211 08:32:13.752474 4860 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d07f5ac6-54a5-4490-a6b4-09fe7325e58a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 08:32:13 crc kubenswrapper[4860]: I1211 08:32:13.752828 4860 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d07f5ac6-54a5-4490-a6b4-09fe7325e58a-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 08:32:13 crc kubenswrapper[4860]: I1211 08:32:13.752842 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z579q\" (UniqueName: \"kubernetes.io/projected/d07f5ac6-54a5-4490-a6b4-09fe7325e58a-kube-api-access-z579q\") on node \"crc\" DevicePath \"\"" Dec 11 08:32:13 crc kubenswrapper[4860]: I1211 08:32:13.752857 4860 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d07f5ac6-54a5-4490-a6b4-09fe7325e58a-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 08:32:14 crc kubenswrapper[4860]: I1211 08:32:14.197319 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-7nvzj" event={"ID":"d07f5ac6-54a5-4490-a6b4-09fe7325e58a","Type":"ContainerDied","Data":"9b632d13e89d5bf8fdd1dc9aa6ae8c02ff10d25234a3bde2635b2b84ec74a692"} Dec 11 08:32:14 crc kubenswrapper[4860]: I1211 08:32:14.197369 4860 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9b632d13e89d5bf8fdd1dc9aa6ae8c02ff10d25234a3bde2635b2b84ec74a692" Dec 11 08:32:14 crc kubenswrapper[4860]: I1211 08:32:14.197430 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-7nvzj" Dec 11 08:32:14 crc kubenswrapper[4860]: I1211 08:32:14.384915 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 11 08:32:14 crc kubenswrapper[4860]: I1211 08:32:14.385295 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="18c09963-6bc7-4dfc-8dfb-a89618d2466d" containerName="nova-api-log" containerID="cri-o://352c4fa0ec3a1568584aa36cd6e6fb79dc5b916be01b6a2ef65dc7b6f1662dea" gracePeriod=30 Dec 11 08:32:14 crc kubenswrapper[4860]: I1211 08:32:14.385463 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="18c09963-6bc7-4dfc-8dfb-a89618d2466d" containerName="nova-api-api" containerID="cri-o://51704a83c489f20401c73b4ba7fcd751bdfba6a32adc3d18ff41ec3553e4816a" gracePeriod=30 Dec 11 08:32:14 crc kubenswrapper[4860]: I1211 08:32:14.399766 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 11 08:32:14 crc kubenswrapper[4860]: I1211 08:32:14.400039 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="66451d3a-291f-494e-bbe7-f68e32fa2305" containerName="nova-scheduler-scheduler" containerID="cri-o://520ae17c7ce420448d27850929ee5f21437fe7d22b66f4d7cc435bbe85127584" gracePeriod=30 Dec 11 08:32:14 crc kubenswrapper[4860]: I1211 08:32:14.432629 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 11 08:32:14 crc kubenswrapper[4860]: I1211 08:32:14.432931 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="a82c6492-9c83-4860-b478-7172462f1be8" containerName="nova-metadata-log" containerID="cri-o://642aa739b6ebe2ab3d86ad681b34f4b367dd2aa72e868bd68c334e4d3da125e2" gracePeriod=30 Dec 11 08:32:14 crc kubenswrapper[4860]: I1211 08:32:14.433524 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="a82c6492-9c83-4860-b478-7172462f1be8" containerName="nova-metadata-metadata" containerID="cri-o://94618ba17096594576975a12f9f1e29c8a01bd0b94daedd85c747df101f11437" gracePeriod=30 Dec 11 08:32:14 crc kubenswrapper[4860]: I1211 08:32:14.983851 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 11 08:32:15 crc kubenswrapper[4860]: I1211 08:32:15.078961 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/18c09963-6bc7-4dfc-8dfb-a89618d2466d-logs\") pod \"18c09963-6bc7-4dfc-8dfb-a89618d2466d\" (UID: \"18c09963-6bc7-4dfc-8dfb-a89618d2466d\") " Dec 11 08:32:15 crc kubenswrapper[4860]: I1211 08:32:15.079003 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b5dfk\" (UniqueName: \"kubernetes.io/projected/18c09963-6bc7-4dfc-8dfb-a89618d2466d-kube-api-access-b5dfk\") pod \"18c09963-6bc7-4dfc-8dfb-a89618d2466d\" (UID: \"18c09963-6bc7-4dfc-8dfb-a89618d2466d\") " Dec 11 08:32:15 crc kubenswrapper[4860]: I1211 08:32:15.079077 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18c09963-6bc7-4dfc-8dfb-a89618d2466d-combined-ca-bundle\") pod \"18c09963-6bc7-4dfc-8dfb-a89618d2466d\" (UID: \"18c09963-6bc7-4dfc-8dfb-a89618d2466d\") " Dec 11 08:32:15 crc kubenswrapper[4860]: I1211 08:32:15.079111 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18c09963-6bc7-4dfc-8dfb-a89618d2466d-config-data\") pod \"18c09963-6bc7-4dfc-8dfb-a89618d2466d\" (UID: \"18c09963-6bc7-4dfc-8dfb-a89618d2466d\") " Dec 11 08:32:15 crc kubenswrapper[4860]: I1211 08:32:15.079196 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/18c09963-6bc7-4dfc-8dfb-a89618d2466d-public-tls-certs\") pod \"18c09963-6bc7-4dfc-8dfb-a89618d2466d\" (UID: \"18c09963-6bc7-4dfc-8dfb-a89618d2466d\") " Dec 11 08:32:15 crc kubenswrapper[4860]: I1211 08:32:15.079358 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/18c09963-6bc7-4dfc-8dfb-a89618d2466d-internal-tls-certs\") pod \"18c09963-6bc7-4dfc-8dfb-a89618d2466d\" (UID: \"18c09963-6bc7-4dfc-8dfb-a89618d2466d\") " Dec 11 08:32:15 crc kubenswrapper[4860]: I1211 08:32:15.079930 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/18c09963-6bc7-4dfc-8dfb-a89618d2466d-logs" (OuterVolumeSpecName: "logs") pod "18c09963-6bc7-4dfc-8dfb-a89618d2466d" (UID: "18c09963-6bc7-4dfc-8dfb-a89618d2466d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:32:15 crc kubenswrapper[4860]: I1211 08:32:15.086770 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/18c09963-6bc7-4dfc-8dfb-a89618d2466d-kube-api-access-b5dfk" (OuterVolumeSpecName: "kube-api-access-b5dfk") pod "18c09963-6bc7-4dfc-8dfb-a89618d2466d" (UID: "18c09963-6bc7-4dfc-8dfb-a89618d2466d"). InnerVolumeSpecName "kube-api-access-b5dfk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:32:15 crc kubenswrapper[4860]: I1211 08:32:15.112678 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18c09963-6bc7-4dfc-8dfb-a89618d2466d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "18c09963-6bc7-4dfc-8dfb-a89618d2466d" (UID: "18c09963-6bc7-4dfc-8dfb-a89618d2466d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:32:15 crc kubenswrapper[4860]: I1211 08:32:15.114622 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18c09963-6bc7-4dfc-8dfb-a89618d2466d-config-data" (OuterVolumeSpecName: "config-data") pod "18c09963-6bc7-4dfc-8dfb-a89618d2466d" (UID: "18c09963-6bc7-4dfc-8dfb-a89618d2466d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:32:15 crc kubenswrapper[4860]: I1211 08:32:15.134414 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18c09963-6bc7-4dfc-8dfb-a89618d2466d-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "18c09963-6bc7-4dfc-8dfb-a89618d2466d" (UID: "18c09963-6bc7-4dfc-8dfb-a89618d2466d"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:32:15 crc kubenswrapper[4860]: I1211 08:32:15.150611 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18c09963-6bc7-4dfc-8dfb-a89618d2466d-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "18c09963-6bc7-4dfc-8dfb-a89618d2466d" (UID: "18c09963-6bc7-4dfc-8dfb-a89618d2466d"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:32:15 crc kubenswrapper[4860]: I1211 08:32:15.182321 4860 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/18c09963-6bc7-4dfc-8dfb-a89618d2466d-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 11 08:32:15 crc kubenswrapper[4860]: I1211 08:32:15.182365 4860 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/18c09963-6bc7-4dfc-8dfb-a89618d2466d-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 11 08:32:15 crc kubenswrapper[4860]: I1211 08:32:15.182380 4860 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/18c09963-6bc7-4dfc-8dfb-a89618d2466d-logs\") on node \"crc\" DevicePath \"\"" Dec 11 08:32:15 crc kubenswrapper[4860]: I1211 08:32:15.182395 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b5dfk\" (UniqueName: \"kubernetes.io/projected/18c09963-6bc7-4dfc-8dfb-a89618d2466d-kube-api-access-b5dfk\") on node \"crc\" DevicePath \"\"" Dec 11 08:32:15 crc kubenswrapper[4860]: I1211 08:32:15.182410 4860 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18c09963-6bc7-4dfc-8dfb-a89618d2466d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 08:32:15 crc kubenswrapper[4860]: I1211 08:32:15.182421 4860 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18c09963-6bc7-4dfc-8dfb-a89618d2466d-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 08:32:15 crc kubenswrapper[4860]: I1211 08:32:15.210375 4860 generic.go:334] "Generic (PLEG): container finished" podID="18c09963-6bc7-4dfc-8dfb-a89618d2466d" containerID="51704a83c489f20401c73b4ba7fcd751bdfba6a32adc3d18ff41ec3553e4816a" exitCode=0 Dec 11 08:32:15 crc kubenswrapper[4860]: I1211 08:32:15.210422 4860 generic.go:334] "Generic (PLEG): container finished" podID="18c09963-6bc7-4dfc-8dfb-a89618d2466d" containerID="352c4fa0ec3a1568584aa36cd6e6fb79dc5b916be01b6a2ef65dc7b6f1662dea" exitCode=143 Dec 11 08:32:15 crc kubenswrapper[4860]: I1211 08:32:15.210438 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 11 08:32:15 crc kubenswrapper[4860]: I1211 08:32:15.210447 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"18c09963-6bc7-4dfc-8dfb-a89618d2466d","Type":"ContainerDied","Data":"51704a83c489f20401c73b4ba7fcd751bdfba6a32adc3d18ff41ec3553e4816a"} Dec 11 08:32:15 crc kubenswrapper[4860]: I1211 08:32:15.210531 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"18c09963-6bc7-4dfc-8dfb-a89618d2466d","Type":"ContainerDied","Data":"352c4fa0ec3a1568584aa36cd6e6fb79dc5b916be01b6a2ef65dc7b6f1662dea"} Dec 11 08:32:15 crc kubenswrapper[4860]: I1211 08:32:15.210547 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"18c09963-6bc7-4dfc-8dfb-a89618d2466d","Type":"ContainerDied","Data":"3ccee3116d0fb682a575b9030084970f24dadb839a1a2703928fe2f69cb846d9"} Dec 11 08:32:15 crc kubenswrapper[4860]: I1211 08:32:15.210569 4860 scope.go:117] "RemoveContainer" containerID="51704a83c489f20401c73b4ba7fcd751bdfba6a32adc3d18ff41ec3553e4816a" Dec 11 08:32:15 crc kubenswrapper[4860]: I1211 08:32:15.213575 4860 generic.go:334] "Generic (PLEG): container finished" podID="a82c6492-9c83-4860-b478-7172462f1be8" containerID="642aa739b6ebe2ab3d86ad681b34f4b367dd2aa72e868bd68c334e4d3da125e2" exitCode=143 Dec 11 08:32:15 crc kubenswrapper[4860]: I1211 08:32:15.213614 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a82c6492-9c83-4860-b478-7172462f1be8","Type":"ContainerDied","Data":"642aa739b6ebe2ab3d86ad681b34f4b367dd2aa72e868bd68c334e4d3da125e2"} Dec 11 08:32:15 crc kubenswrapper[4860]: I1211 08:32:15.241839 4860 scope.go:117] "RemoveContainer" containerID="352c4fa0ec3a1568584aa36cd6e6fb79dc5b916be01b6a2ef65dc7b6f1662dea" Dec 11 08:32:15 crc kubenswrapper[4860]: I1211 08:32:15.266002 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 11 08:32:15 crc kubenswrapper[4860]: I1211 08:32:15.266089 4860 scope.go:117] "RemoveContainer" containerID="51704a83c489f20401c73b4ba7fcd751bdfba6a32adc3d18ff41ec3553e4816a" Dec 11 08:32:15 crc kubenswrapper[4860]: E1211 08:32:15.267006 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"51704a83c489f20401c73b4ba7fcd751bdfba6a32adc3d18ff41ec3553e4816a\": container with ID starting with 51704a83c489f20401c73b4ba7fcd751bdfba6a32adc3d18ff41ec3553e4816a not found: ID does not exist" containerID="51704a83c489f20401c73b4ba7fcd751bdfba6a32adc3d18ff41ec3553e4816a" Dec 11 08:32:15 crc kubenswrapper[4860]: I1211 08:32:15.267061 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"51704a83c489f20401c73b4ba7fcd751bdfba6a32adc3d18ff41ec3553e4816a"} err="failed to get container status \"51704a83c489f20401c73b4ba7fcd751bdfba6a32adc3d18ff41ec3553e4816a\": rpc error: code = NotFound desc = could not find container \"51704a83c489f20401c73b4ba7fcd751bdfba6a32adc3d18ff41ec3553e4816a\": container with ID starting with 51704a83c489f20401c73b4ba7fcd751bdfba6a32adc3d18ff41ec3553e4816a not found: ID does not exist" Dec 11 08:32:15 crc kubenswrapper[4860]: I1211 08:32:15.267097 4860 scope.go:117] "RemoveContainer" containerID="352c4fa0ec3a1568584aa36cd6e6fb79dc5b916be01b6a2ef65dc7b6f1662dea" Dec 11 08:32:15 crc kubenswrapper[4860]: E1211 08:32:15.267388 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"352c4fa0ec3a1568584aa36cd6e6fb79dc5b916be01b6a2ef65dc7b6f1662dea\": container with ID starting with 352c4fa0ec3a1568584aa36cd6e6fb79dc5b916be01b6a2ef65dc7b6f1662dea not found: ID does not exist" containerID="352c4fa0ec3a1568584aa36cd6e6fb79dc5b916be01b6a2ef65dc7b6f1662dea" Dec 11 08:32:15 crc kubenswrapper[4860]: I1211 08:32:15.267429 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"352c4fa0ec3a1568584aa36cd6e6fb79dc5b916be01b6a2ef65dc7b6f1662dea"} err="failed to get container status \"352c4fa0ec3a1568584aa36cd6e6fb79dc5b916be01b6a2ef65dc7b6f1662dea\": rpc error: code = NotFound desc = could not find container \"352c4fa0ec3a1568584aa36cd6e6fb79dc5b916be01b6a2ef65dc7b6f1662dea\": container with ID starting with 352c4fa0ec3a1568584aa36cd6e6fb79dc5b916be01b6a2ef65dc7b6f1662dea not found: ID does not exist" Dec 11 08:32:15 crc kubenswrapper[4860]: I1211 08:32:15.267448 4860 scope.go:117] "RemoveContainer" containerID="51704a83c489f20401c73b4ba7fcd751bdfba6a32adc3d18ff41ec3553e4816a" Dec 11 08:32:15 crc kubenswrapper[4860]: I1211 08:32:15.270140 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"51704a83c489f20401c73b4ba7fcd751bdfba6a32adc3d18ff41ec3553e4816a"} err="failed to get container status \"51704a83c489f20401c73b4ba7fcd751bdfba6a32adc3d18ff41ec3553e4816a\": rpc error: code = NotFound desc = could not find container \"51704a83c489f20401c73b4ba7fcd751bdfba6a32adc3d18ff41ec3553e4816a\": container with ID starting with 51704a83c489f20401c73b4ba7fcd751bdfba6a32adc3d18ff41ec3553e4816a not found: ID does not exist" Dec 11 08:32:15 crc kubenswrapper[4860]: I1211 08:32:15.270182 4860 scope.go:117] "RemoveContainer" containerID="352c4fa0ec3a1568584aa36cd6e6fb79dc5b916be01b6a2ef65dc7b6f1662dea" Dec 11 08:32:15 crc kubenswrapper[4860]: I1211 08:32:15.270554 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"352c4fa0ec3a1568584aa36cd6e6fb79dc5b916be01b6a2ef65dc7b6f1662dea"} err="failed to get container status \"352c4fa0ec3a1568584aa36cd6e6fb79dc5b916be01b6a2ef65dc7b6f1662dea\": rpc error: code = NotFound desc = could not find container \"352c4fa0ec3a1568584aa36cd6e6fb79dc5b916be01b6a2ef65dc7b6f1662dea\": container with ID starting with 352c4fa0ec3a1568584aa36cd6e6fb79dc5b916be01b6a2ef65dc7b6f1662dea not found: ID does not exist" Dec 11 08:32:15 crc kubenswrapper[4860]: I1211 08:32:15.286180 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 11 08:32:15 crc kubenswrapper[4860]: I1211 08:32:15.300234 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 11 08:32:15 crc kubenswrapper[4860]: E1211 08:32:15.300775 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e338f34-d693-481b-a8ed-eb8ecbdd5a3b" containerName="init" Dec 11 08:32:15 crc kubenswrapper[4860]: I1211 08:32:15.300795 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e338f34-d693-481b-a8ed-eb8ecbdd5a3b" containerName="init" Dec 11 08:32:15 crc kubenswrapper[4860]: E1211 08:32:15.300805 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e338f34-d693-481b-a8ed-eb8ecbdd5a3b" containerName="dnsmasq-dns" Dec 11 08:32:15 crc kubenswrapper[4860]: I1211 08:32:15.300811 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e338f34-d693-481b-a8ed-eb8ecbdd5a3b" containerName="dnsmasq-dns" Dec 11 08:32:15 crc kubenswrapper[4860]: E1211 08:32:15.300831 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18c09963-6bc7-4dfc-8dfb-a89618d2466d" containerName="nova-api-log" Dec 11 08:32:15 crc kubenswrapper[4860]: I1211 08:32:15.300837 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="18c09963-6bc7-4dfc-8dfb-a89618d2466d" containerName="nova-api-log" Dec 11 08:32:15 crc kubenswrapper[4860]: E1211 08:32:15.300845 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d07f5ac6-54a5-4490-a6b4-09fe7325e58a" containerName="nova-manage" Dec 11 08:32:15 crc kubenswrapper[4860]: I1211 08:32:15.300851 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="d07f5ac6-54a5-4490-a6b4-09fe7325e58a" containerName="nova-manage" Dec 11 08:32:15 crc kubenswrapper[4860]: E1211 08:32:15.300863 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18c09963-6bc7-4dfc-8dfb-a89618d2466d" containerName="nova-api-api" Dec 11 08:32:15 crc kubenswrapper[4860]: I1211 08:32:15.300868 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="18c09963-6bc7-4dfc-8dfb-a89618d2466d" containerName="nova-api-api" Dec 11 08:32:15 crc kubenswrapper[4860]: I1211 08:32:15.301120 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="18c09963-6bc7-4dfc-8dfb-a89618d2466d" containerName="nova-api-log" Dec 11 08:32:15 crc kubenswrapper[4860]: I1211 08:32:15.301139 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="18c09963-6bc7-4dfc-8dfb-a89618d2466d" containerName="nova-api-api" Dec 11 08:32:15 crc kubenswrapper[4860]: I1211 08:32:15.301153 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e338f34-d693-481b-a8ed-eb8ecbdd5a3b" containerName="dnsmasq-dns" Dec 11 08:32:15 crc kubenswrapper[4860]: I1211 08:32:15.301166 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="d07f5ac6-54a5-4490-a6b4-09fe7325e58a" containerName="nova-manage" Dec 11 08:32:15 crc kubenswrapper[4860]: I1211 08:32:15.311910 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 11 08:32:15 crc kubenswrapper[4860]: I1211 08:32:15.312022 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 11 08:32:15 crc kubenswrapper[4860]: I1211 08:32:15.318728 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 11 08:32:15 crc kubenswrapper[4860]: I1211 08:32:15.319068 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 11 08:32:15 crc kubenswrapper[4860]: I1211 08:32:15.319416 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 11 08:32:15 crc kubenswrapper[4860]: I1211 08:32:15.386047 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fcbd0284-56fd-4cb0-895e-78e6a571868a-internal-tls-certs\") pod \"nova-api-0\" (UID: \"fcbd0284-56fd-4cb0-895e-78e6a571868a\") " pod="openstack/nova-api-0" Dec 11 08:32:15 crc kubenswrapper[4860]: I1211 08:32:15.386101 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fcbd0284-56fd-4cb0-895e-78e6a571868a-config-data\") pod \"nova-api-0\" (UID: \"fcbd0284-56fd-4cb0-895e-78e6a571868a\") " pod="openstack/nova-api-0" Dec 11 08:32:15 crc kubenswrapper[4860]: I1211 08:32:15.386127 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fcbd0284-56fd-4cb0-895e-78e6a571868a-logs\") pod \"nova-api-0\" (UID: \"fcbd0284-56fd-4cb0-895e-78e6a571868a\") " pod="openstack/nova-api-0" Dec 11 08:32:15 crc kubenswrapper[4860]: I1211 08:32:15.386180 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fcbd0284-56fd-4cb0-895e-78e6a571868a-public-tls-certs\") pod \"nova-api-0\" (UID: \"fcbd0284-56fd-4cb0-895e-78e6a571868a\") " pod="openstack/nova-api-0" Dec 11 08:32:15 crc kubenswrapper[4860]: I1211 08:32:15.386241 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fcbd0284-56fd-4cb0-895e-78e6a571868a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"fcbd0284-56fd-4cb0-895e-78e6a571868a\") " pod="openstack/nova-api-0" Dec 11 08:32:15 crc kubenswrapper[4860]: I1211 08:32:15.386287 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ghmtw\" (UniqueName: \"kubernetes.io/projected/fcbd0284-56fd-4cb0-895e-78e6a571868a-kube-api-access-ghmtw\") pod \"nova-api-0\" (UID: \"fcbd0284-56fd-4cb0-895e-78e6a571868a\") " pod="openstack/nova-api-0" Dec 11 08:32:15 crc kubenswrapper[4860]: I1211 08:32:15.488064 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ghmtw\" (UniqueName: \"kubernetes.io/projected/fcbd0284-56fd-4cb0-895e-78e6a571868a-kube-api-access-ghmtw\") pod \"nova-api-0\" (UID: \"fcbd0284-56fd-4cb0-895e-78e6a571868a\") " pod="openstack/nova-api-0" Dec 11 08:32:15 crc kubenswrapper[4860]: I1211 08:32:15.488135 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fcbd0284-56fd-4cb0-895e-78e6a571868a-internal-tls-certs\") pod \"nova-api-0\" (UID: \"fcbd0284-56fd-4cb0-895e-78e6a571868a\") " pod="openstack/nova-api-0" Dec 11 08:32:15 crc kubenswrapper[4860]: I1211 08:32:15.488161 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fcbd0284-56fd-4cb0-895e-78e6a571868a-config-data\") pod \"nova-api-0\" (UID: \"fcbd0284-56fd-4cb0-895e-78e6a571868a\") " pod="openstack/nova-api-0" Dec 11 08:32:15 crc kubenswrapper[4860]: I1211 08:32:15.488177 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fcbd0284-56fd-4cb0-895e-78e6a571868a-logs\") pod \"nova-api-0\" (UID: \"fcbd0284-56fd-4cb0-895e-78e6a571868a\") " pod="openstack/nova-api-0" Dec 11 08:32:15 crc kubenswrapper[4860]: I1211 08:32:15.488225 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fcbd0284-56fd-4cb0-895e-78e6a571868a-public-tls-certs\") pod \"nova-api-0\" (UID: \"fcbd0284-56fd-4cb0-895e-78e6a571868a\") " pod="openstack/nova-api-0" Dec 11 08:32:15 crc kubenswrapper[4860]: I1211 08:32:15.488280 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fcbd0284-56fd-4cb0-895e-78e6a571868a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"fcbd0284-56fd-4cb0-895e-78e6a571868a\") " pod="openstack/nova-api-0" Dec 11 08:32:15 crc kubenswrapper[4860]: I1211 08:32:15.488920 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fcbd0284-56fd-4cb0-895e-78e6a571868a-logs\") pod \"nova-api-0\" (UID: \"fcbd0284-56fd-4cb0-895e-78e6a571868a\") " pod="openstack/nova-api-0" Dec 11 08:32:15 crc kubenswrapper[4860]: I1211 08:32:15.492030 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fcbd0284-56fd-4cb0-895e-78e6a571868a-public-tls-certs\") pod \"nova-api-0\" (UID: \"fcbd0284-56fd-4cb0-895e-78e6a571868a\") " pod="openstack/nova-api-0" Dec 11 08:32:15 crc kubenswrapper[4860]: I1211 08:32:15.492030 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fcbd0284-56fd-4cb0-895e-78e6a571868a-config-data\") pod \"nova-api-0\" (UID: \"fcbd0284-56fd-4cb0-895e-78e6a571868a\") " pod="openstack/nova-api-0" Dec 11 08:32:15 crc kubenswrapper[4860]: I1211 08:32:15.492925 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fcbd0284-56fd-4cb0-895e-78e6a571868a-internal-tls-certs\") pod \"nova-api-0\" (UID: \"fcbd0284-56fd-4cb0-895e-78e6a571868a\") " pod="openstack/nova-api-0" Dec 11 08:32:15 crc kubenswrapper[4860]: I1211 08:32:15.501593 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fcbd0284-56fd-4cb0-895e-78e6a571868a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"fcbd0284-56fd-4cb0-895e-78e6a571868a\") " pod="openstack/nova-api-0" Dec 11 08:32:15 crc kubenswrapper[4860]: I1211 08:32:15.516328 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ghmtw\" (UniqueName: \"kubernetes.io/projected/fcbd0284-56fd-4cb0-895e-78e6a571868a-kube-api-access-ghmtw\") pod \"nova-api-0\" (UID: \"fcbd0284-56fd-4cb0-895e-78e6a571868a\") " pod="openstack/nova-api-0" Dec 11 08:32:15 crc kubenswrapper[4860]: I1211 08:32:15.593524 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="18c09963-6bc7-4dfc-8dfb-a89618d2466d" path="/var/lib/kubelet/pods/18c09963-6bc7-4dfc-8dfb-a89618d2466d/volumes" Dec 11 08:32:15 crc kubenswrapper[4860]: I1211 08:32:15.638321 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 11 08:32:15 crc kubenswrapper[4860]: I1211 08:32:15.799969 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 11 08:32:15 crc kubenswrapper[4860]: I1211 08:32:15.899184 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/66451d3a-291f-494e-bbe7-f68e32fa2305-config-data\") pod \"66451d3a-291f-494e-bbe7-f68e32fa2305\" (UID: \"66451d3a-291f-494e-bbe7-f68e32fa2305\") " Dec 11 08:32:15 crc kubenswrapper[4860]: I1211 08:32:15.899450 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66451d3a-291f-494e-bbe7-f68e32fa2305-combined-ca-bundle\") pod \"66451d3a-291f-494e-bbe7-f68e32fa2305\" (UID: \"66451d3a-291f-494e-bbe7-f68e32fa2305\") " Dec 11 08:32:15 crc kubenswrapper[4860]: I1211 08:32:15.899503 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jrhbt\" (UniqueName: \"kubernetes.io/projected/66451d3a-291f-494e-bbe7-f68e32fa2305-kube-api-access-jrhbt\") pod \"66451d3a-291f-494e-bbe7-f68e32fa2305\" (UID: \"66451d3a-291f-494e-bbe7-f68e32fa2305\") " Dec 11 08:32:15 crc kubenswrapper[4860]: I1211 08:32:15.906810 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66451d3a-291f-494e-bbe7-f68e32fa2305-kube-api-access-jrhbt" (OuterVolumeSpecName: "kube-api-access-jrhbt") pod "66451d3a-291f-494e-bbe7-f68e32fa2305" (UID: "66451d3a-291f-494e-bbe7-f68e32fa2305"). InnerVolumeSpecName "kube-api-access-jrhbt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:32:15 crc kubenswrapper[4860]: I1211 08:32:15.930707 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66451d3a-291f-494e-bbe7-f68e32fa2305-config-data" (OuterVolumeSpecName: "config-data") pod "66451d3a-291f-494e-bbe7-f68e32fa2305" (UID: "66451d3a-291f-494e-bbe7-f68e32fa2305"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:32:15 crc kubenswrapper[4860]: I1211 08:32:15.946606 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66451d3a-291f-494e-bbe7-f68e32fa2305-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "66451d3a-291f-494e-bbe7-f68e32fa2305" (UID: "66451d3a-291f-494e-bbe7-f68e32fa2305"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:32:16 crc kubenswrapper[4860]: I1211 08:32:16.002586 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jrhbt\" (UniqueName: \"kubernetes.io/projected/66451d3a-291f-494e-bbe7-f68e32fa2305-kube-api-access-jrhbt\") on node \"crc\" DevicePath \"\"" Dec 11 08:32:16 crc kubenswrapper[4860]: I1211 08:32:16.002634 4860 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/66451d3a-291f-494e-bbe7-f68e32fa2305-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 08:32:16 crc kubenswrapper[4860]: I1211 08:32:16.002670 4860 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66451d3a-291f-494e-bbe7-f68e32fa2305-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 08:32:16 crc kubenswrapper[4860]: I1211 08:32:16.143902 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 11 08:32:16 crc kubenswrapper[4860]: W1211 08:32:16.150028 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfcbd0284_56fd_4cb0_895e_78e6a571868a.slice/crio-06e4b2938c96d08f310fb6a353e95911a3a369ff69b3d3bacce44ba025a6002a WatchSource:0}: Error finding container 06e4b2938c96d08f310fb6a353e95911a3a369ff69b3d3bacce44ba025a6002a: Status 404 returned error can't find the container with id 06e4b2938c96d08f310fb6a353e95911a3a369ff69b3d3bacce44ba025a6002a Dec 11 08:32:16 crc kubenswrapper[4860]: I1211 08:32:16.227534 4860 generic.go:334] "Generic (PLEG): container finished" podID="66451d3a-291f-494e-bbe7-f68e32fa2305" containerID="520ae17c7ce420448d27850929ee5f21437fe7d22b66f4d7cc435bbe85127584" exitCode=0 Dec 11 08:32:16 crc kubenswrapper[4860]: I1211 08:32:16.227585 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 11 08:32:16 crc kubenswrapper[4860]: I1211 08:32:16.227630 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"66451d3a-291f-494e-bbe7-f68e32fa2305","Type":"ContainerDied","Data":"520ae17c7ce420448d27850929ee5f21437fe7d22b66f4d7cc435bbe85127584"} Dec 11 08:32:16 crc kubenswrapper[4860]: I1211 08:32:16.227687 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"66451d3a-291f-494e-bbe7-f68e32fa2305","Type":"ContainerDied","Data":"e3424c913919879c2dc6fc9aa31e826369c3887b5e8ecc72d7aac58dc14366e0"} Dec 11 08:32:16 crc kubenswrapper[4860]: I1211 08:32:16.227714 4860 scope.go:117] "RemoveContainer" containerID="520ae17c7ce420448d27850929ee5f21437fe7d22b66f4d7cc435bbe85127584" Dec 11 08:32:16 crc kubenswrapper[4860]: I1211 08:32:16.231709 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"fcbd0284-56fd-4cb0-895e-78e6a571868a","Type":"ContainerStarted","Data":"06e4b2938c96d08f310fb6a353e95911a3a369ff69b3d3bacce44ba025a6002a"} Dec 11 08:32:16 crc kubenswrapper[4860]: I1211 08:32:16.265506 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 11 08:32:16 crc kubenswrapper[4860]: I1211 08:32:16.266776 4860 scope.go:117] "RemoveContainer" containerID="520ae17c7ce420448d27850929ee5f21437fe7d22b66f4d7cc435bbe85127584" Dec 11 08:32:16 crc kubenswrapper[4860]: E1211 08:32:16.267285 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"520ae17c7ce420448d27850929ee5f21437fe7d22b66f4d7cc435bbe85127584\": container with ID starting with 520ae17c7ce420448d27850929ee5f21437fe7d22b66f4d7cc435bbe85127584 not found: ID does not exist" containerID="520ae17c7ce420448d27850929ee5f21437fe7d22b66f4d7cc435bbe85127584" Dec 11 08:32:16 crc kubenswrapper[4860]: I1211 08:32:16.267325 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"520ae17c7ce420448d27850929ee5f21437fe7d22b66f4d7cc435bbe85127584"} err="failed to get container status \"520ae17c7ce420448d27850929ee5f21437fe7d22b66f4d7cc435bbe85127584\": rpc error: code = NotFound desc = could not find container \"520ae17c7ce420448d27850929ee5f21437fe7d22b66f4d7cc435bbe85127584\": container with ID starting with 520ae17c7ce420448d27850929ee5f21437fe7d22b66f4d7cc435bbe85127584 not found: ID does not exist" Dec 11 08:32:16 crc kubenswrapper[4860]: I1211 08:32:16.282707 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 11 08:32:16 crc kubenswrapper[4860]: I1211 08:32:16.299061 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 11 08:32:16 crc kubenswrapper[4860]: E1211 08:32:16.299494 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66451d3a-291f-494e-bbe7-f68e32fa2305" containerName="nova-scheduler-scheduler" Dec 11 08:32:16 crc kubenswrapper[4860]: I1211 08:32:16.299516 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="66451d3a-291f-494e-bbe7-f68e32fa2305" containerName="nova-scheduler-scheduler" Dec 11 08:32:16 crc kubenswrapper[4860]: I1211 08:32:16.299744 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="66451d3a-291f-494e-bbe7-f68e32fa2305" containerName="nova-scheduler-scheduler" Dec 11 08:32:16 crc kubenswrapper[4860]: I1211 08:32:16.300370 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 11 08:32:16 crc kubenswrapper[4860]: I1211 08:32:16.303100 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 11 08:32:16 crc kubenswrapper[4860]: I1211 08:32:16.310733 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 11 08:32:16 crc kubenswrapper[4860]: I1211 08:32:16.417880 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08a3b710-84c1-4bbc-9547-a350c0d76318-config-data\") pod \"nova-scheduler-0\" (UID: \"08a3b710-84c1-4bbc-9547-a350c0d76318\") " pod="openstack/nova-scheduler-0" Dec 11 08:32:16 crc kubenswrapper[4860]: I1211 08:32:16.417925 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08a3b710-84c1-4bbc-9547-a350c0d76318-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"08a3b710-84c1-4bbc-9547-a350c0d76318\") " pod="openstack/nova-scheduler-0" Dec 11 08:32:16 crc kubenswrapper[4860]: I1211 08:32:16.418012 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cbhmd\" (UniqueName: \"kubernetes.io/projected/08a3b710-84c1-4bbc-9547-a350c0d76318-kube-api-access-cbhmd\") pod \"nova-scheduler-0\" (UID: \"08a3b710-84c1-4bbc-9547-a350c0d76318\") " pod="openstack/nova-scheduler-0" Dec 11 08:32:16 crc kubenswrapper[4860]: I1211 08:32:16.519700 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08a3b710-84c1-4bbc-9547-a350c0d76318-config-data\") pod \"nova-scheduler-0\" (UID: \"08a3b710-84c1-4bbc-9547-a350c0d76318\") " pod="openstack/nova-scheduler-0" Dec 11 08:32:16 crc kubenswrapper[4860]: I1211 08:32:16.520241 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08a3b710-84c1-4bbc-9547-a350c0d76318-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"08a3b710-84c1-4bbc-9547-a350c0d76318\") " pod="openstack/nova-scheduler-0" Dec 11 08:32:16 crc kubenswrapper[4860]: I1211 08:32:16.520410 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cbhmd\" (UniqueName: \"kubernetes.io/projected/08a3b710-84c1-4bbc-9547-a350c0d76318-kube-api-access-cbhmd\") pod \"nova-scheduler-0\" (UID: \"08a3b710-84c1-4bbc-9547-a350c0d76318\") " pod="openstack/nova-scheduler-0" Dec 11 08:32:16 crc kubenswrapper[4860]: I1211 08:32:16.524002 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08a3b710-84c1-4bbc-9547-a350c0d76318-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"08a3b710-84c1-4bbc-9547-a350c0d76318\") " pod="openstack/nova-scheduler-0" Dec 11 08:32:16 crc kubenswrapper[4860]: I1211 08:32:16.525638 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08a3b710-84c1-4bbc-9547-a350c0d76318-config-data\") pod \"nova-scheduler-0\" (UID: \"08a3b710-84c1-4bbc-9547-a350c0d76318\") " pod="openstack/nova-scheduler-0" Dec 11 08:32:16 crc kubenswrapper[4860]: I1211 08:32:16.542296 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cbhmd\" (UniqueName: \"kubernetes.io/projected/08a3b710-84c1-4bbc-9547-a350c0d76318-kube-api-access-cbhmd\") pod \"nova-scheduler-0\" (UID: \"08a3b710-84c1-4bbc-9547-a350c0d76318\") " pod="openstack/nova-scheduler-0" Dec 11 08:32:16 crc kubenswrapper[4860]: I1211 08:32:16.629723 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 11 08:32:17 crc kubenswrapper[4860]: I1211 08:32:17.075869 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 11 08:32:17 crc kubenswrapper[4860]: W1211 08:32:17.088047 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod08a3b710_84c1_4bbc_9547_a350c0d76318.slice/crio-c48d8bea5e0afb0478b235b135eeb4f02b81cdf20f8d226d3c8b778ef6f7681c WatchSource:0}: Error finding container c48d8bea5e0afb0478b235b135eeb4f02b81cdf20f8d226d3c8b778ef6f7681c: Status 404 returned error can't find the container with id c48d8bea5e0afb0478b235b135eeb4f02b81cdf20f8d226d3c8b778ef6f7681c Dec 11 08:32:17 crc kubenswrapper[4860]: I1211 08:32:17.246214 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"08a3b710-84c1-4bbc-9547-a350c0d76318","Type":"ContainerStarted","Data":"c48d8bea5e0afb0478b235b135eeb4f02b81cdf20f8d226d3c8b778ef6f7681c"} Dec 11 08:32:17 crc kubenswrapper[4860]: I1211 08:32:17.248515 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"fcbd0284-56fd-4cb0-895e-78e6a571868a","Type":"ContainerStarted","Data":"cbf3d71cb8e30a7a2142342000159a3d335d8a21f8a28a95167227d1e0fc76c9"} Dec 11 08:32:17 crc kubenswrapper[4860]: I1211 08:32:17.248548 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"fcbd0284-56fd-4cb0-895e-78e6a571868a","Type":"ContainerStarted","Data":"99ea3d6e0d38d51e954ee25820817c196fd0eaa9424d92ac5efd1138dfc7bff2"} Dec 11 08:32:17 crc kubenswrapper[4860]: I1211 08:32:17.279228 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.279201439 podStartE2EDuration="2.279201439s" podCreationTimestamp="2025-12-11 08:32:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:32:17.270853635 +0000 UTC m=+1269.999372710" watchObservedRunningTime="2025-12-11 08:32:17.279201439 +0000 UTC m=+1270.007720504" Dec 11 08:32:17 crc kubenswrapper[4860]: I1211 08:32:17.569489 4860 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="a82c6492-9c83-4860-b478-7172462f1be8" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.191:8775/\": read tcp 10.217.0.2:55658->10.217.0.191:8775: read: connection reset by peer" Dec 11 08:32:17 crc kubenswrapper[4860]: I1211 08:32:17.570221 4860 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="a82c6492-9c83-4860-b478-7172462f1be8" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.191:8775/\": read tcp 10.217.0.2:55674->10.217.0.191:8775: read: connection reset by peer" Dec 11 08:32:17 crc kubenswrapper[4860]: I1211 08:32:17.607197 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="66451d3a-291f-494e-bbe7-f68e32fa2305" path="/var/lib/kubelet/pods/66451d3a-291f-494e-bbe7-f68e32fa2305/volumes" Dec 11 08:32:18 crc kubenswrapper[4860]: I1211 08:32:18.060254 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 11 08:32:18 crc kubenswrapper[4860]: I1211 08:32:18.157360 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qlgk6\" (UniqueName: \"kubernetes.io/projected/a82c6492-9c83-4860-b478-7172462f1be8-kube-api-access-qlgk6\") pod \"a82c6492-9c83-4860-b478-7172462f1be8\" (UID: \"a82c6492-9c83-4860-b478-7172462f1be8\") " Dec 11 08:32:18 crc kubenswrapper[4860]: I1211 08:32:18.157560 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a82c6492-9c83-4860-b478-7172462f1be8-nova-metadata-tls-certs\") pod \"a82c6492-9c83-4860-b478-7172462f1be8\" (UID: \"a82c6492-9c83-4860-b478-7172462f1be8\") " Dec 11 08:32:18 crc kubenswrapper[4860]: I1211 08:32:18.157632 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a82c6492-9c83-4860-b478-7172462f1be8-combined-ca-bundle\") pod \"a82c6492-9c83-4860-b478-7172462f1be8\" (UID: \"a82c6492-9c83-4860-b478-7172462f1be8\") " Dec 11 08:32:18 crc kubenswrapper[4860]: I1211 08:32:18.157687 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a82c6492-9c83-4860-b478-7172462f1be8-config-data\") pod \"a82c6492-9c83-4860-b478-7172462f1be8\" (UID: \"a82c6492-9c83-4860-b478-7172462f1be8\") " Dec 11 08:32:18 crc kubenswrapper[4860]: I1211 08:32:18.157802 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a82c6492-9c83-4860-b478-7172462f1be8-logs\") pod \"a82c6492-9c83-4860-b478-7172462f1be8\" (UID: \"a82c6492-9c83-4860-b478-7172462f1be8\") " Dec 11 08:32:18 crc kubenswrapper[4860]: I1211 08:32:18.158803 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a82c6492-9c83-4860-b478-7172462f1be8-logs" (OuterVolumeSpecName: "logs") pod "a82c6492-9c83-4860-b478-7172462f1be8" (UID: "a82c6492-9c83-4860-b478-7172462f1be8"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:32:18 crc kubenswrapper[4860]: I1211 08:32:18.195360 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a82c6492-9c83-4860-b478-7172462f1be8-kube-api-access-qlgk6" (OuterVolumeSpecName: "kube-api-access-qlgk6") pod "a82c6492-9c83-4860-b478-7172462f1be8" (UID: "a82c6492-9c83-4860-b478-7172462f1be8"). InnerVolumeSpecName "kube-api-access-qlgk6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:32:18 crc kubenswrapper[4860]: I1211 08:32:18.205779 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a82c6492-9c83-4860-b478-7172462f1be8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a82c6492-9c83-4860-b478-7172462f1be8" (UID: "a82c6492-9c83-4860-b478-7172462f1be8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:32:18 crc kubenswrapper[4860]: I1211 08:32:18.255260 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a82c6492-9c83-4860-b478-7172462f1be8-config-data" (OuterVolumeSpecName: "config-data") pod "a82c6492-9c83-4860-b478-7172462f1be8" (UID: "a82c6492-9c83-4860-b478-7172462f1be8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:32:18 crc kubenswrapper[4860]: I1211 08:32:18.261316 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qlgk6\" (UniqueName: \"kubernetes.io/projected/a82c6492-9c83-4860-b478-7172462f1be8-kube-api-access-qlgk6\") on node \"crc\" DevicePath \"\"" Dec 11 08:32:18 crc kubenswrapper[4860]: I1211 08:32:18.261346 4860 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a82c6492-9c83-4860-b478-7172462f1be8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 08:32:18 crc kubenswrapper[4860]: I1211 08:32:18.261357 4860 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a82c6492-9c83-4860-b478-7172462f1be8-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 08:32:18 crc kubenswrapper[4860]: I1211 08:32:18.261367 4860 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a82c6492-9c83-4860-b478-7172462f1be8-logs\") on node \"crc\" DevicePath \"\"" Dec 11 08:32:18 crc kubenswrapper[4860]: I1211 08:32:18.266388 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"08a3b710-84c1-4bbc-9547-a350c0d76318","Type":"ContainerStarted","Data":"c41194b8598b238d039e01aee608c65f29b6edcc9bd1318b19d149718c656e55"} Dec 11 08:32:18 crc kubenswrapper[4860]: I1211 08:32:18.297683 4860 generic.go:334] "Generic (PLEG): container finished" podID="a82c6492-9c83-4860-b478-7172462f1be8" containerID="94618ba17096594576975a12f9f1e29c8a01bd0b94daedd85c747df101f11437" exitCode=0 Dec 11 08:32:18 crc kubenswrapper[4860]: I1211 08:32:18.297879 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a82c6492-9c83-4860-b478-7172462f1be8-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "a82c6492-9c83-4860-b478-7172462f1be8" (UID: "a82c6492-9c83-4860-b478-7172462f1be8"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:32:18 crc kubenswrapper[4860]: I1211 08:32:18.298013 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 11 08:32:18 crc kubenswrapper[4860]: I1211 08:32:18.298624 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a82c6492-9c83-4860-b478-7172462f1be8","Type":"ContainerDied","Data":"94618ba17096594576975a12f9f1e29c8a01bd0b94daedd85c747df101f11437"} Dec 11 08:32:18 crc kubenswrapper[4860]: I1211 08:32:18.298671 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a82c6492-9c83-4860-b478-7172462f1be8","Type":"ContainerDied","Data":"2f797655173402203be6f6db324ed212192a675354cd2505404e7cff0528dbe2"} Dec 11 08:32:18 crc kubenswrapper[4860]: I1211 08:32:18.298690 4860 scope.go:117] "RemoveContainer" containerID="94618ba17096594576975a12f9f1e29c8a01bd0b94daedd85c747df101f11437" Dec 11 08:32:18 crc kubenswrapper[4860]: I1211 08:32:18.303481 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.303450623 podStartE2EDuration="2.303450623s" podCreationTimestamp="2025-12-11 08:32:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:32:18.297465374 +0000 UTC m=+1271.025984439" watchObservedRunningTime="2025-12-11 08:32:18.303450623 +0000 UTC m=+1271.031969678" Dec 11 08:32:18 crc kubenswrapper[4860]: I1211 08:32:18.357695 4860 scope.go:117] "RemoveContainer" containerID="642aa739b6ebe2ab3d86ad681b34f4b367dd2aa72e868bd68c334e4d3da125e2" Dec 11 08:32:18 crc kubenswrapper[4860]: I1211 08:32:18.366913 4860 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a82c6492-9c83-4860-b478-7172462f1be8-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 11 08:32:18 crc kubenswrapper[4860]: I1211 08:32:18.373516 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 11 08:32:18 crc kubenswrapper[4860]: I1211 08:32:18.386876 4860 scope.go:117] "RemoveContainer" containerID="94618ba17096594576975a12f9f1e29c8a01bd0b94daedd85c747df101f11437" Dec 11 08:32:18 crc kubenswrapper[4860]: E1211 08:32:18.391339 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"94618ba17096594576975a12f9f1e29c8a01bd0b94daedd85c747df101f11437\": container with ID starting with 94618ba17096594576975a12f9f1e29c8a01bd0b94daedd85c747df101f11437 not found: ID does not exist" containerID="94618ba17096594576975a12f9f1e29c8a01bd0b94daedd85c747df101f11437" Dec 11 08:32:18 crc kubenswrapper[4860]: I1211 08:32:18.391388 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"94618ba17096594576975a12f9f1e29c8a01bd0b94daedd85c747df101f11437"} err="failed to get container status \"94618ba17096594576975a12f9f1e29c8a01bd0b94daedd85c747df101f11437\": rpc error: code = NotFound desc = could not find container \"94618ba17096594576975a12f9f1e29c8a01bd0b94daedd85c747df101f11437\": container with ID starting with 94618ba17096594576975a12f9f1e29c8a01bd0b94daedd85c747df101f11437 not found: ID does not exist" Dec 11 08:32:18 crc kubenswrapper[4860]: I1211 08:32:18.391421 4860 scope.go:117] "RemoveContainer" containerID="642aa739b6ebe2ab3d86ad681b34f4b367dd2aa72e868bd68c334e4d3da125e2" Dec 11 08:32:18 crc kubenswrapper[4860]: I1211 08:32:18.395331 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 11 08:32:18 crc kubenswrapper[4860]: E1211 08:32:18.395713 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"642aa739b6ebe2ab3d86ad681b34f4b367dd2aa72e868bd68c334e4d3da125e2\": container with ID starting with 642aa739b6ebe2ab3d86ad681b34f4b367dd2aa72e868bd68c334e4d3da125e2 not found: ID does not exist" containerID="642aa739b6ebe2ab3d86ad681b34f4b367dd2aa72e868bd68c334e4d3da125e2" Dec 11 08:32:18 crc kubenswrapper[4860]: I1211 08:32:18.395739 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"642aa739b6ebe2ab3d86ad681b34f4b367dd2aa72e868bd68c334e4d3da125e2"} err="failed to get container status \"642aa739b6ebe2ab3d86ad681b34f4b367dd2aa72e868bd68c334e4d3da125e2\": rpc error: code = NotFound desc = could not find container \"642aa739b6ebe2ab3d86ad681b34f4b367dd2aa72e868bd68c334e4d3da125e2\": container with ID starting with 642aa739b6ebe2ab3d86ad681b34f4b367dd2aa72e868bd68c334e4d3da125e2 not found: ID does not exist" Dec 11 08:32:18 crc kubenswrapper[4860]: I1211 08:32:18.425874 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 11 08:32:18 crc kubenswrapper[4860]: E1211 08:32:18.426315 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a82c6492-9c83-4860-b478-7172462f1be8" containerName="nova-metadata-log" Dec 11 08:32:18 crc kubenswrapper[4860]: I1211 08:32:18.426339 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="a82c6492-9c83-4860-b478-7172462f1be8" containerName="nova-metadata-log" Dec 11 08:32:18 crc kubenswrapper[4860]: E1211 08:32:18.426364 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a82c6492-9c83-4860-b478-7172462f1be8" containerName="nova-metadata-metadata" Dec 11 08:32:18 crc kubenswrapper[4860]: I1211 08:32:18.426371 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="a82c6492-9c83-4860-b478-7172462f1be8" containerName="nova-metadata-metadata" Dec 11 08:32:18 crc kubenswrapper[4860]: I1211 08:32:18.426698 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="a82c6492-9c83-4860-b478-7172462f1be8" containerName="nova-metadata-metadata" Dec 11 08:32:18 crc kubenswrapper[4860]: I1211 08:32:18.426714 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="a82c6492-9c83-4860-b478-7172462f1be8" containerName="nova-metadata-log" Dec 11 08:32:18 crc kubenswrapper[4860]: I1211 08:32:18.428013 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 11 08:32:18 crc kubenswrapper[4860]: I1211 08:32:18.430137 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 11 08:32:18 crc kubenswrapper[4860]: I1211 08:32:18.430262 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 11 08:32:18 crc kubenswrapper[4860]: I1211 08:32:18.436622 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 11 08:32:18 crc kubenswrapper[4860]: I1211 08:32:18.571062 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/4b75f4be-3810-4273-a18e-381ab6c6ef2b-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"4b75f4be-3810-4273-a18e-381ab6c6ef2b\") " pod="openstack/nova-metadata-0" Dec 11 08:32:18 crc kubenswrapper[4860]: I1211 08:32:18.571157 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-94xdr\" (UniqueName: \"kubernetes.io/projected/4b75f4be-3810-4273-a18e-381ab6c6ef2b-kube-api-access-94xdr\") pod \"nova-metadata-0\" (UID: \"4b75f4be-3810-4273-a18e-381ab6c6ef2b\") " pod="openstack/nova-metadata-0" Dec 11 08:32:18 crc kubenswrapper[4860]: I1211 08:32:18.571211 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b75f4be-3810-4273-a18e-381ab6c6ef2b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"4b75f4be-3810-4273-a18e-381ab6c6ef2b\") " pod="openstack/nova-metadata-0" Dec 11 08:32:18 crc kubenswrapper[4860]: I1211 08:32:18.571385 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b75f4be-3810-4273-a18e-381ab6c6ef2b-config-data\") pod \"nova-metadata-0\" (UID: \"4b75f4be-3810-4273-a18e-381ab6c6ef2b\") " pod="openstack/nova-metadata-0" Dec 11 08:32:18 crc kubenswrapper[4860]: I1211 08:32:18.571438 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4b75f4be-3810-4273-a18e-381ab6c6ef2b-logs\") pod \"nova-metadata-0\" (UID: \"4b75f4be-3810-4273-a18e-381ab6c6ef2b\") " pod="openstack/nova-metadata-0" Dec 11 08:32:18 crc kubenswrapper[4860]: I1211 08:32:18.673903 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4b75f4be-3810-4273-a18e-381ab6c6ef2b-logs\") pod \"nova-metadata-0\" (UID: \"4b75f4be-3810-4273-a18e-381ab6c6ef2b\") " pod="openstack/nova-metadata-0" Dec 11 08:32:18 crc kubenswrapper[4860]: I1211 08:32:18.674128 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/4b75f4be-3810-4273-a18e-381ab6c6ef2b-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"4b75f4be-3810-4273-a18e-381ab6c6ef2b\") " pod="openstack/nova-metadata-0" Dec 11 08:32:18 crc kubenswrapper[4860]: I1211 08:32:18.674291 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-94xdr\" (UniqueName: \"kubernetes.io/projected/4b75f4be-3810-4273-a18e-381ab6c6ef2b-kube-api-access-94xdr\") pod \"nova-metadata-0\" (UID: \"4b75f4be-3810-4273-a18e-381ab6c6ef2b\") " pod="openstack/nova-metadata-0" Dec 11 08:32:18 crc kubenswrapper[4860]: I1211 08:32:18.674329 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b75f4be-3810-4273-a18e-381ab6c6ef2b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"4b75f4be-3810-4273-a18e-381ab6c6ef2b\") " pod="openstack/nova-metadata-0" Dec 11 08:32:18 crc kubenswrapper[4860]: I1211 08:32:18.674399 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b75f4be-3810-4273-a18e-381ab6c6ef2b-config-data\") pod \"nova-metadata-0\" (UID: \"4b75f4be-3810-4273-a18e-381ab6c6ef2b\") " pod="openstack/nova-metadata-0" Dec 11 08:32:18 crc kubenswrapper[4860]: I1211 08:32:18.674603 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4b75f4be-3810-4273-a18e-381ab6c6ef2b-logs\") pod \"nova-metadata-0\" (UID: \"4b75f4be-3810-4273-a18e-381ab6c6ef2b\") " pod="openstack/nova-metadata-0" Dec 11 08:32:18 crc kubenswrapper[4860]: I1211 08:32:18.679239 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b75f4be-3810-4273-a18e-381ab6c6ef2b-config-data\") pod \"nova-metadata-0\" (UID: \"4b75f4be-3810-4273-a18e-381ab6c6ef2b\") " pod="openstack/nova-metadata-0" Dec 11 08:32:18 crc kubenswrapper[4860]: I1211 08:32:18.679875 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/4b75f4be-3810-4273-a18e-381ab6c6ef2b-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"4b75f4be-3810-4273-a18e-381ab6c6ef2b\") " pod="openstack/nova-metadata-0" Dec 11 08:32:18 crc kubenswrapper[4860]: I1211 08:32:18.684502 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b75f4be-3810-4273-a18e-381ab6c6ef2b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"4b75f4be-3810-4273-a18e-381ab6c6ef2b\") " pod="openstack/nova-metadata-0" Dec 11 08:32:18 crc kubenswrapper[4860]: I1211 08:32:18.691921 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-94xdr\" (UniqueName: \"kubernetes.io/projected/4b75f4be-3810-4273-a18e-381ab6c6ef2b-kube-api-access-94xdr\") pod \"nova-metadata-0\" (UID: \"4b75f4be-3810-4273-a18e-381ab6c6ef2b\") " pod="openstack/nova-metadata-0" Dec 11 08:32:18 crc kubenswrapper[4860]: I1211 08:32:18.767258 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 11 08:32:19 crc kubenswrapper[4860]: I1211 08:32:19.214283 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 11 08:32:19 crc kubenswrapper[4860]: W1211 08:32:19.215859 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4b75f4be_3810_4273_a18e_381ab6c6ef2b.slice/crio-b97b281ef5d88d74e6354d218fc814c53ea06a92e1039b6701ad50bd00e49c69 WatchSource:0}: Error finding container b97b281ef5d88d74e6354d218fc814c53ea06a92e1039b6701ad50bd00e49c69: Status 404 returned error can't find the container with id b97b281ef5d88d74e6354d218fc814c53ea06a92e1039b6701ad50bd00e49c69 Dec 11 08:32:19 crc kubenswrapper[4860]: I1211 08:32:19.318393 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4b75f4be-3810-4273-a18e-381ab6c6ef2b","Type":"ContainerStarted","Data":"b97b281ef5d88d74e6354d218fc814c53ea06a92e1039b6701ad50bd00e49c69"} Dec 11 08:32:19 crc kubenswrapper[4860]: I1211 08:32:19.601849 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a82c6492-9c83-4860-b478-7172462f1be8" path="/var/lib/kubelet/pods/a82c6492-9c83-4860-b478-7172462f1be8/volumes" Dec 11 08:32:20 crc kubenswrapper[4860]: I1211 08:32:20.337902 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4b75f4be-3810-4273-a18e-381ab6c6ef2b","Type":"ContainerStarted","Data":"01f6cea7f0188139639afcb2cb2f02df9bbc4ab5772cd11f1eaa37446664df2c"} Dec 11 08:32:20 crc kubenswrapper[4860]: I1211 08:32:20.338330 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4b75f4be-3810-4273-a18e-381ab6c6ef2b","Type":"ContainerStarted","Data":"77cb7b9bb0bbcc5cc16f5ad5b82fd59fa14b18ba65d3b0cc262e206d946c6dc5"} Dec 11 08:32:20 crc kubenswrapper[4860]: I1211 08:32:20.372568 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.372547611 podStartE2EDuration="2.372547611s" podCreationTimestamp="2025-12-11 08:32:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:32:20.357776586 +0000 UTC m=+1273.086295651" watchObservedRunningTime="2025-12-11 08:32:20.372547611 +0000 UTC m=+1273.101066666" Dec 11 08:32:21 crc kubenswrapper[4860]: I1211 08:32:21.630971 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 11 08:32:23 crc kubenswrapper[4860]: I1211 08:32:23.768320 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 11 08:32:23 crc kubenswrapper[4860]: I1211 08:32:23.768685 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 11 08:32:25 crc kubenswrapper[4860]: I1211 08:32:25.639342 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 11 08:32:25 crc kubenswrapper[4860]: I1211 08:32:25.639680 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 11 08:32:26 crc kubenswrapper[4860]: I1211 08:32:26.631067 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 11 08:32:26 crc kubenswrapper[4860]: I1211 08:32:26.651876 4860 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="fcbd0284-56fd-4cb0-895e-78e6a571868a" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.200:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 11 08:32:26 crc kubenswrapper[4860]: I1211 08:32:26.652179 4860 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="fcbd0284-56fd-4cb0-895e-78e6a571868a" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.200:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 11 08:32:26 crc kubenswrapper[4860]: I1211 08:32:26.667361 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 11 08:32:27 crc kubenswrapper[4860]: I1211 08:32:27.487180 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 11 08:32:28 crc kubenswrapper[4860]: I1211 08:32:28.767634 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 11 08:32:28 crc kubenswrapper[4860]: I1211 08:32:28.767969 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 11 08:32:29 crc kubenswrapper[4860]: I1211 08:32:29.789822 4860 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="4b75f4be-3810-4273-a18e-381ab6c6ef2b" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.202:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 11 08:32:29 crc kubenswrapper[4860]: I1211 08:32:29.789825 4860 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="4b75f4be-3810-4273-a18e-381ab6c6ef2b" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.202:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 11 08:32:33 crc kubenswrapper[4860]: I1211 08:32:33.474477 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 11 08:32:35 crc kubenswrapper[4860]: I1211 08:32:35.644756 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 11 08:32:35 crc kubenswrapper[4860]: I1211 08:32:35.648750 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 11 08:32:35 crc kubenswrapper[4860]: I1211 08:32:35.659768 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 11 08:32:35 crc kubenswrapper[4860]: I1211 08:32:35.667140 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 11 08:32:36 crc kubenswrapper[4860]: I1211 08:32:36.550907 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 11 08:32:36 crc kubenswrapper[4860]: I1211 08:32:36.559019 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 11 08:32:38 crc kubenswrapper[4860]: I1211 08:32:38.775187 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 11 08:32:38 crc kubenswrapper[4860]: I1211 08:32:38.775845 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 11 08:32:38 crc kubenswrapper[4860]: I1211 08:32:38.785660 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 11 08:32:38 crc kubenswrapper[4860]: I1211 08:32:38.792286 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 11 08:32:47 crc kubenswrapper[4860]: I1211 08:32:47.368119 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 11 08:32:48 crc kubenswrapper[4860]: I1211 08:32:48.262682 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 11 08:32:51 crc kubenswrapper[4860]: I1211 08:32:51.854329 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="97634ec0-7953-4e38-a07e-95967ce874d7" containerName="rabbitmq" containerID="cri-o://c08aab93a48a1273f41e2191e9a12ab22ada37fed914adea9f44a9eb9eee90c9" gracePeriod=604796 Dec 11 08:32:52 crc kubenswrapper[4860]: I1211 08:32:52.632052 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="a004cb7b-ee08-44c0-9dc7-d0b69509fe13" containerName="rabbitmq" containerID="cri-o://c1b416ccd0ce4787808ac0bc1fe677baedfb2f3a87f1c9829d5cf793665cbe0c" gracePeriod=604796 Dec 11 08:32:56 crc kubenswrapper[4860]: I1211 08:32:56.284918 4860 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="97634ec0-7953-4e38-a07e-95967ce874d7" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.98:5671: connect: connection refused" Dec 11 08:32:56 crc kubenswrapper[4860]: I1211 08:32:56.382468 4860 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="a004cb7b-ee08-44c0-9dc7-d0b69509fe13" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.99:5671: connect: connection refused" Dec 11 08:32:58 crc kubenswrapper[4860]: I1211 08:32:58.438317 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 11 08:32:58 crc kubenswrapper[4860]: I1211 08:32:58.577065 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/97634ec0-7953-4e38-a07e-95967ce874d7-config-data\") pod \"97634ec0-7953-4e38-a07e-95967ce874d7\" (UID: \"97634ec0-7953-4e38-a07e-95967ce874d7\") " Dec 11 08:32:58 crc kubenswrapper[4860]: I1211 08:32:58.577171 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/97634ec0-7953-4e38-a07e-95967ce874d7-erlang-cookie-secret\") pod \"97634ec0-7953-4e38-a07e-95967ce874d7\" (UID: \"97634ec0-7953-4e38-a07e-95967ce874d7\") " Dec 11 08:32:58 crc kubenswrapper[4860]: I1211 08:32:58.577258 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/97634ec0-7953-4e38-a07e-95967ce874d7-rabbitmq-erlang-cookie\") pod \"97634ec0-7953-4e38-a07e-95967ce874d7\" (UID: \"97634ec0-7953-4e38-a07e-95967ce874d7\") " Dec 11 08:32:58 crc kubenswrapper[4860]: I1211 08:32:58.577350 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/97634ec0-7953-4e38-a07e-95967ce874d7-plugins-conf\") pod \"97634ec0-7953-4e38-a07e-95967ce874d7\" (UID: \"97634ec0-7953-4e38-a07e-95967ce874d7\") " Dec 11 08:32:58 crc kubenswrapper[4860]: I1211 08:32:58.577997 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/97634ec0-7953-4e38-a07e-95967ce874d7-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "97634ec0-7953-4e38-a07e-95967ce874d7" (UID: "97634ec0-7953-4e38-a07e-95967ce874d7"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:32:58 crc kubenswrapper[4860]: I1211 08:32:58.578113 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/97634ec0-7953-4e38-a07e-95967ce874d7-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "97634ec0-7953-4e38-a07e-95967ce874d7" (UID: "97634ec0-7953-4e38-a07e-95967ce874d7"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:32:58 crc kubenswrapper[4860]: I1211 08:32:58.578411 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/97634ec0-7953-4e38-a07e-95967ce874d7-rabbitmq-plugins\") pod \"97634ec0-7953-4e38-a07e-95967ce874d7\" (UID: \"97634ec0-7953-4e38-a07e-95967ce874d7\") " Dec 11 08:32:58 crc kubenswrapper[4860]: I1211 08:32:58.578437 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/97634ec0-7953-4e38-a07e-95967ce874d7-rabbitmq-confd\") pod \"97634ec0-7953-4e38-a07e-95967ce874d7\" (UID: \"97634ec0-7953-4e38-a07e-95967ce874d7\") " Dec 11 08:32:58 crc kubenswrapper[4860]: I1211 08:32:58.578492 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/97634ec0-7953-4e38-a07e-95967ce874d7-rabbitmq-tls\") pod \"97634ec0-7953-4e38-a07e-95967ce874d7\" (UID: \"97634ec0-7953-4e38-a07e-95967ce874d7\") " Dec 11 08:32:58 crc kubenswrapper[4860]: I1211 08:32:58.578559 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/97634ec0-7953-4e38-a07e-95967ce874d7-server-conf\") pod \"97634ec0-7953-4e38-a07e-95967ce874d7\" (UID: \"97634ec0-7953-4e38-a07e-95967ce874d7\") " Dec 11 08:32:58 crc kubenswrapper[4860]: I1211 08:32:58.579107 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/97634ec0-7953-4e38-a07e-95967ce874d7-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "97634ec0-7953-4e38-a07e-95967ce874d7" (UID: "97634ec0-7953-4e38-a07e-95967ce874d7"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:32:58 crc kubenswrapper[4860]: I1211 08:32:58.579518 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/97634ec0-7953-4e38-a07e-95967ce874d7-pod-info\") pod \"97634ec0-7953-4e38-a07e-95967ce874d7\" (UID: \"97634ec0-7953-4e38-a07e-95967ce874d7\") " Dec 11 08:32:58 crc kubenswrapper[4860]: I1211 08:32:58.579612 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"97634ec0-7953-4e38-a07e-95967ce874d7\" (UID: \"97634ec0-7953-4e38-a07e-95967ce874d7\") " Dec 11 08:32:58 crc kubenswrapper[4860]: I1211 08:32:58.579656 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jbrbn\" (UniqueName: \"kubernetes.io/projected/97634ec0-7953-4e38-a07e-95967ce874d7-kube-api-access-jbrbn\") pod \"97634ec0-7953-4e38-a07e-95967ce874d7\" (UID: \"97634ec0-7953-4e38-a07e-95967ce874d7\") " Dec 11 08:32:58 crc kubenswrapper[4860]: I1211 08:32:58.580415 4860 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/97634ec0-7953-4e38-a07e-95967ce874d7-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 11 08:32:58 crc kubenswrapper[4860]: I1211 08:32:58.580433 4860 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/97634ec0-7953-4e38-a07e-95967ce874d7-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 11 08:32:58 crc kubenswrapper[4860]: I1211 08:32:58.580446 4860 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/97634ec0-7953-4e38-a07e-95967ce874d7-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 11 08:32:58 crc kubenswrapper[4860]: I1211 08:32:58.583593 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/97634ec0-7953-4e38-a07e-95967ce874d7-pod-info" (OuterVolumeSpecName: "pod-info") pod "97634ec0-7953-4e38-a07e-95967ce874d7" (UID: "97634ec0-7953-4e38-a07e-95967ce874d7"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 11 08:32:58 crc kubenswrapper[4860]: I1211 08:32:58.583916 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/97634ec0-7953-4e38-a07e-95967ce874d7-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "97634ec0-7953-4e38-a07e-95967ce874d7" (UID: "97634ec0-7953-4e38-a07e-95967ce874d7"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:32:58 crc kubenswrapper[4860]: I1211 08:32:58.586189 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/97634ec0-7953-4e38-a07e-95967ce874d7-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "97634ec0-7953-4e38-a07e-95967ce874d7" (UID: "97634ec0-7953-4e38-a07e-95967ce874d7"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:32:58 crc kubenswrapper[4860]: I1211 08:32:58.587422 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "persistence") pod "97634ec0-7953-4e38-a07e-95967ce874d7" (UID: "97634ec0-7953-4e38-a07e-95967ce874d7"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 11 08:32:58 crc kubenswrapper[4860]: I1211 08:32:58.592835 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/97634ec0-7953-4e38-a07e-95967ce874d7-kube-api-access-jbrbn" (OuterVolumeSpecName: "kube-api-access-jbrbn") pod "97634ec0-7953-4e38-a07e-95967ce874d7" (UID: "97634ec0-7953-4e38-a07e-95967ce874d7"). InnerVolumeSpecName "kube-api-access-jbrbn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:32:58 crc kubenswrapper[4860]: I1211 08:32:58.609077 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/97634ec0-7953-4e38-a07e-95967ce874d7-config-data" (OuterVolumeSpecName: "config-data") pod "97634ec0-7953-4e38-a07e-95967ce874d7" (UID: "97634ec0-7953-4e38-a07e-95967ce874d7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:32:58 crc kubenswrapper[4860]: I1211 08:32:58.643884 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/97634ec0-7953-4e38-a07e-95967ce874d7-server-conf" (OuterVolumeSpecName: "server-conf") pod "97634ec0-7953-4e38-a07e-95967ce874d7" (UID: "97634ec0-7953-4e38-a07e-95967ce874d7"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:32:58 crc kubenswrapper[4860]: I1211 08:32:58.682391 4860 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Dec 11 08:32:58 crc kubenswrapper[4860]: I1211 08:32:58.682424 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jbrbn\" (UniqueName: \"kubernetes.io/projected/97634ec0-7953-4e38-a07e-95967ce874d7-kube-api-access-jbrbn\") on node \"crc\" DevicePath \"\"" Dec 11 08:32:58 crc kubenswrapper[4860]: I1211 08:32:58.682434 4860 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/97634ec0-7953-4e38-a07e-95967ce874d7-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 08:32:58 crc kubenswrapper[4860]: I1211 08:32:58.682444 4860 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/97634ec0-7953-4e38-a07e-95967ce874d7-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 11 08:32:58 crc kubenswrapper[4860]: I1211 08:32:58.682455 4860 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/97634ec0-7953-4e38-a07e-95967ce874d7-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 11 08:32:58 crc kubenswrapper[4860]: I1211 08:32:58.682464 4860 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/97634ec0-7953-4e38-a07e-95967ce874d7-server-conf\") on node \"crc\" DevicePath \"\"" Dec 11 08:32:58 crc kubenswrapper[4860]: I1211 08:32:58.682472 4860 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/97634ec0-7953-4e38-a07e-95967ce874d7-pod-info\") on node \"crc\" DevicePath \"\"" Dec 11 08:32:58 crc kubenswrapper[4860]: I1211 08:32:58.713573 4860 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Dec 11 08:32:58 crc kubenswrapper[4860]: I1211 08:32:58.752044 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/97634ec0-7953-4e38-a07e-95967ce874d7-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "97634ec0-7953-4e38-a07e-95967ce874d7" (UID: "97634ec0-7953-4e38-a07e-95967ce874d7"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:32:58 crc kubenswrapper[4860]: I1211 08:32:58.783913 4860 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/97634ec0-7953-4e38-a07e-95967ce874d7-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 11 08:32:58 crc kubenswrapper[4860]: I1211 08:32:58.783945 4860 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Dec 11 08:32:58 crc kubenswrapper[4860]: I1211 08:32:58.791166 4860 generic.go:334] "Generic (PLEG): container finished" podID="97634ec0-7953-4e38-a07e-95967ce874d7" containerID="c08aab93a48a1273f41e2191e9a12ab22ada37fed914adea9f44a9eb9eee90c9" exitCode=0 Dec 11 08:32:58 crc kubenswrapper[4860]: I1211 08:32:58.791217 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"97634ec0-7953-4e38-a07e-95967ce874d7","Type":"ContainerDied","Data":"c08aab93a48a1273f41e2191e9a12ab22ada37fed914adea9f44a9eb9eee90c9"} Dec 11 08:32:58 crc kubenswrapper[4860]: I1211 08:32:58.791251 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"97634ec0-7953-4e38-a07e-95967ce874d7","Type":"ContainerDied","Data":"01823976a22b0507fcfd26eed82dc3121aabbf54791145971f8499276ddbea83"} Dec 11 08:32:58 crc kubenswrapper[4860]: I1211 08:32:58.791271 4860 scope.go:117] "RemoveContainer" containerID="c08aab93a48a1273f41e2191e9a12ab22ada37fed914adea9f44a9eb9eee90c9" Dec 11 08:32:58 crc kubenswrapper[4860]: I1211 08:32:58.791451 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 11 08:32:58 crc kubenswrapper[4860]: I1211 08:32:58.837145 4860 scope.go:117] "RemoveContainer" containerID="c1aae3ef18a4011e77f7a6a7c3f8803d1d98f58a1d231ac7824be320f058f72a" Dec 11 08:32:58 crc kubenswrapper[4860]: I1211 08:32:58.850247 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 11 08:32:58 crc kubenswrapper[4860]: I1211 08:32:58.862448 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 11 08:32:58 crc kubenswrapper[4860]: I1211 08:32:58.882171 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 11 08:32:58 crc kubenswrapper[4860]: E1211 08:32:58.882868 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97634ec0-7953-4e38-a07e-95967ce874d7" containerName="rabbitmq" Dec 11 08:32:58 crc kubenswrapper[4860]: I1211 08:32:58.882888 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="97634ec0-7953-4e38-a07e-95967ce874d7" containerName="rabbitmq" Dec 11 08:32:58 crc kubenswrapper[4860]: E1211 08:32:58.882942 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97634ec0-7953-4e38-a07e-95967ce874d7" containerName="setup-container" Dec 11 08:32:58 crc kubenswrapper[4860]: I1211 08:32:58.882953 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="97634ec0-7953-4e38-a07e-95967ce874d7" containerName="setup-container" Dec 11 08:32:58 crc kubenswrapper[4860]: I1211 08:32:58.883174 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="97634ec0-7953-4e38-a07e-95967ce874d7" containerName="rabbitmq" Dec 11 08:32:58 crc kubenswrapper[4860]: I1211 08:32:58.886732 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 11 08:32:58 crc kubenswrapper[4860]: I1211 08:32:58.890862 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 11 08:32:58 crc kubenswrapper[4860]: I1211 08:32:58.891043 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 11 08:32:58 crc kubenswrapper[4860]: I1211 08:32:58.891272 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-6vxjr" Dec 11 08:32:58 crc kubenswrapper[4860]: I1211 08:32:58.891489 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 11 08:32:58 crc kubenswrapper[4860]: I1211 08:32:58.891880 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 11 08:32:58 crc kubenswrapper[4860]: I1211 08:32:58.892143 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 11 08:32:58 crc kubenswrapper[4860]: I1211 08:32:58.893988 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 11 08:32:58 crc kubenswrapper[4860]: I1211 08:32:58.904765 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 11 08:32:58 crc kubenswrapper[4860]: I1211 08:32:58.930501 4860 scope.go:117] "RemoveContainer" containerID="c08aab93a48a1273f41e2191e9a12ab22ada37fed914adea9f44a9eb9eee90c9" Dec 11 08:32:58 crc kubenswrapper[4860]: E1211 08:32:58.934109 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c08aab93a48a1273f41e2191e9a12ab22ada37fed914adea9f44a9eb9eee90c9\": container with ID starting with c08aab93a48a1273f41e2191e9a12ab22ada37fed914adea9f44a9eb9eee90c9 not found: ID does not exist" containerID="c08aab93a48a1273f41e2191e9a12ab22ada37fed914adea9f44a9eb9eee90c9" Dec 11 08:32:58 crc kubenswrapper[4860]: I1211 08:32:58.934168 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c08aab93a48a1273f41e2191e9a12ab22ada37fed914adea9f44a9eb9eee90c9"} err="failed to get container status \"c08aab93a48a1273f41e2191e9a12ab22ada37fed914adea9f44a9eb9eee90c9\": rpc error: code = NotFound desc = could not find container \"c08aab93a48a1273f41e2191e9a12ab22ada37fed914adea9f44a9eb9eee90c9\": container with ID starting with c08aab93a48a1273f41e2191e9a12ab22ada37fed914adea9f44a9eb9eee90c9 not found: ID does not exist" Dec 11 08:32:58 crc kubenswrapper[4860]: I1211 08:32:58.934198 4860 scope.go:117] "RemoveContainer" containerID="c1aae3ef18a4011e77f7a6a7c3f8803d1d98f58a1d231ac7824be320f058f72a" Dec 11 08:32:58 crc kubenswrapper[4860]: E1211 08:32:58.938890 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c1aae3ef18a4011e77f7a6a7c3f8803d1d98f58a1d231ac7824be320f058f72a\": container with ID starting with c1aae3ef18a4011e77f7a6a7c3f8803d1d98f58a1d231ac7824be320f058f72a not found: ID does not exist" containerID="c1aae3ef18a4011e77f7a6a7c3f8803d1d98f58a1d231ac7824be320f058f72a" Dec 11 08:32:58 crc kubenswrapper[4860]: I1211 08:32:58.938937 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c1aae3ef18a4011e77f7a6a7c3f8803d1d98f58a1d231ac7824be320f058f72a"} err="failed to get container status \"c1aae3ef18a4011e77f7a6a7c3f8803d1d98f58a1d231ac7824be320f058f72a\": rpc error: code = NotFound desc = could not find container \"c1aae3ef18a4011e77f7a6a7c3f8803d1d98f58a1d231ac7824be320f058f72a\": container with ID starting with c1aae3ef18a4011e77f7a6a7c3f8803d1d98f58a1d231ac7824be320f058f72a not found: ID does not exist" Dec 11 08:32:58 crc kubenswrapper[4860]: I1211 08:32:58.987888 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a1d839a1-4174-4abc-9e3d-5579603687e6-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"a1d839a1-4174-4abc-9e3d-5579603687e6\") " pod="openstack/rabbitmq-server-0" Dec 11 08:32:58 crc kubenswrapper[4860]: I1211 08:32:58.987953 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a1d839a1-4174-4abc-9e3d-5579603687e6-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"a1d839a1-4174-4abc-9e3d-5579603687e6\") " pod="openstack/rabbitmq-server-0" Dec 11 08:32:58 crc kubenswrapper[4860]: I1211 08:32:58.988025 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a1d839a1-4174-4abc-9e3d-5579603687e6-pod-info\") pod \"rabbitmq-server-0\" (UID: \"a1d839a1-4174-4abc-9e3d-5579603687e6\") " pod="openstack/rabbitmq-server-0" Dec 11 08:32:58 crc kubenswrapper[4860]: I1211 08:32:58.988051 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a1d839a1-4174-4abc-9e3d-5579603687e6-config-data\") pod \"rabbitmq-server-0\" (UID: \"a1d839a1-4174-4abc-9e3d-5579603687e6\") " pod="openstack/rabbitmq-server-0" Dec 11 08:32:58 crc kubenswrapper[4860]: I1211 08:32:58.988080 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a1d839a1-4174-4abc-9e3d-5579603687e6-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"a1d839a1-4174-4abc-9e3d-5579603687e6\") " pod="openstack/rabbitmq-server-0" Dec 11 08:32:58 crc kubenswrapper[4860]: I1211 08:32:58.988105 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a1d839a1-4174-4abc-9e3d-5579603687e6-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"a1d839a1-4174-4abc-9e3d-5579603687e6\") " pod="openstack/rabbitmq-server-0" Dec 11 08:32:58 crc kubenswrapper[4860]: I1211 08:32:58.988147 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-687sn\" (UniqueName: \"kubernetes.io/projected/a1d839a1-4174-4abc-9e3d-5579603687e6-kube-api-access-687sn\") pod \"rabbitmq-server-0\" (UID: \"a1d839a1-4174-4abc-9e3d-5579603687e6\") " pod="openstack/rabbitmq-server-0" Dec 11 08:32:58 crc kubenswrapper[4860]: I1211 08:32:58.988189 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a1d839a1-4174-4abc-9e3d-5579603687e6-server-conf\") pod \"rabbitmq-server-0\" (UID: \"a1d839a1-4174-4abc-9e3d-5579603687e6\") " pod="openstack/rabbitmq-server-0" Dec 11 08:32:58 crc kubenswrapper[4860]: I1211 08:32:58.988209 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a1d839a1-4174-4abc-9e3d-5579603687e6-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"a1d839a1-4174-4abc-9e3d-5579603687e6\") " pod="openstack/rabbitmq-server-0" Dec 11 08:32:58 crc kubenswrapper[4860]: I1211 08:32:58.988234 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"a1d839a1-4174-4abc-9e3d-5579603687e6\") " pod="openstack/rabbitmq-server-0" Dec 11 08:32:58 crc kubenswrapper[4860]: I1211 08:32:58.988280 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a1d839a1-4174-4abc-9e3d-5579603687e6-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"a1d839a1-4174-4abc-9e3d-5579603687e6\") " pod="openstack/rabbitmq-server-0" Dec 11 08:32:59 crc kubenswrapper[4860]: I1211 08:32:59.090845 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a1d839a1-4174-4abc-9e3d-5579603687e6-pod-info\") pod \"rabbitmq-server-0\" (UID: \"a1d839a1-4174-4abc-9e3d-5579603687e6\") " pod="openstack/rabbitmq-server-0" Dec 11 08:32:59 crc kubenswrapper[4860]: I1211 08:32:59.091243 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a1d839a1-4174-4abc-9e3d-5579603687e6-config-data\") pod \"rabbitmq-server-0\" (UID: \"a1d839a1-4174-4abc-9e3d-5579603687e6\") " pod="openstack/rabbitmq-server-0" Dec 11 08:32:59 crc kubenswrapper[4860]: I1211 08:32:59.091317 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a1d839a1-4174-4abc-9e3d-5579603687e6-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"a1d839a1-4174-4abc-9e3d-5579603687e6\") " pod="openstack/rabbitmq-server-0" Dec 11 08:32:59 crc kubenswrapper[4860]: I1211 08:32:59.091370 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a1d839a1-4174-4abc-9e3d-5579603687e6-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"a1d839a1-4174-4abc-9e3d-5579603687e6\") " pod="openstack/rabbitmq-server-0" Dec 11 08:32:59 crc kubenswrapper[4860]: I1211 08:32:59.091460 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-687sn\" (UniqueName: \"kubernetes.io/projected/a1d839a1-4174-4abc-9e3d-5579603687e6-kube-api-access-687sn\") pod \"rabbitmq-server-0\" (UID: \"a1d839a1-4174-4abc-9e3d-5579603687e6\") " pod="openstack/rabbitmq-server-0" Dec 11 08:32:59 crc kubenswrapper[4860]: I1211 08:32:59.091536 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a1d839a1-4174-4abc-9e3d-5579603687e6-server-conf\") pod \"rabbitmq-server-0\" (UID: \"a1d839a1-4174-4abc-9e3d-5579603687e6\") " pod="openstack/rabbitmq-server-0" Dec 11 08:32:59 crc kubenswrapper[4860]: I1211 08:32:59.091565 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a1d839a1-4174-4abc-9e3d-5579603687e6-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"a1d839a1-4174-4abc-9e3d-5579603687e6\") " pod="openstack/rabbitmq-server-0" Dec 11 08:32:59 crc kubenswrapper[4860]: I1211 08:32:59.091614 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"a1d839a1-4174-4abc-9e3d-5579603687e6\") " pod="openstack/rabbitmq-server-0" Dec 11 08:32:59 crc kubenswrapper[4860]: I1211 08:32:59.091697 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a1d839a1-4174-4abc-9e3d-5579603687e6-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"a1d839a1-4174-4abc-9e3d-5579603687e6\") " pod="openstack/rabbitmq-server-0" Dec 11 08:32:59 crc kubenswrapper[4860]: I1211 08:32:59.091729 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a1d839a1-4174-4abc-9e3d-5579603687e6-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"a1d839a1-4174-4abc-9e3d-5579603687e6\") " pod="openstack/rabbitmq-server-0" Dec 11 08:32:59 crc kubenswrapper[4860]: I1211 08:32:59.091766 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a1d839a1-4174-4abc-9e3d-5579603687e6-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"a1d839a1-4174-4abc-9e3d-5579603687e6\") " pod="openstack/rabbitmq-server-0" Dec 11 08:32:59 crc kubenswrapper[4860]: I1211 08:32:59.095591 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a1d839a1-4174-4abc-9e3d-5579603687e6-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"a1d839a1-4174-4abc-9e3d-5579603687e6\") " pod="openstack/rabbitmq-server-0" Dec 11 08:32:59 crc kubenswrapper[4860]: I1211 08:32:59.095882 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a1d839a1-4174-4abc-9e3d-5579603687e6-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"a1d839a1-4174-4abc-9e3d-5579603687e6\") " pod="openstack/rabbitmq-server-0" Dec 11 08:32:59 crc kubenswrapper[4860]: I1211 08:32:59.096115 4860 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"a1d839a1-4174-4abc-9e3d-5579603687e6\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/rabbitmq-server-0" Dec 11 08:32:59 crc kubenswrapper[4860]: I1211 08:32:59.096257 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a1d839a1-4174-4abc-9e3d-5579603687e6-config-data\") pod \"rabbitmq-server-0\" (UID: \"a1d839a1-4174-4abc-9e3d-5579603687e6\") " pod="openstack/rabbitmq-server-0" Dec 11 08:32:59 crc kubenswrapper[4860]: I1211 08:32:59.097596 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a1d839a1-4174-4abc-9e3d-5579603687e6-server-conf\") pod \"rabbitmq-server-0\" (UID: \"a1d839a1-4174-4abc-9e3d-5579603687e6\") " pod="openstack/rabbitmq-server-0" Dec 11 08:32:59 crc kubenswrapper[4860]: I1211 08:32:59.097990 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a1d839a1-4174-4abc-9e3d-5579603687e6-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"a1d839a1-4174-4abc-9e3d-5579603687e6\") " pod="openstack/rabbitmq-server-0" Dec 11 08:32:59 crc kubenswrapper[4860]: I1211 08:32:59.101285 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a1d839a1-4174-4abc-9e3d-5579603687e6-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"a1d839a1-4174-4abc-9e3d-5579603687e6\") " pod="openstack/rabbitmq-server-0" Dec 11 08:32:59 crc kubenswrapper[4860]: I1211 08:32:59.102009 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a1d839a1-4174-4abc-9e3d-5579603687e6-pod-info\") pod \"rabbitmq-server-0\" (UID: \"a1d839a1-4174-4abc-9e3d-5579603687e6\") " pod="openstack/rabbitmq-server-0" Dec 11 08:32:59 crc kubenswrapper[4860]: I1211 08:32:59.104232 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a1d839a1-4174-4abc-9e3d-5579603687e6-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"a1d839a1-4174-4abc-9e3d-5579603687e6\") " pod="openstack/rabbitmq-server-0" Dec 11 08:32:59 crc kubenswrapper[4860]: I1211 08:32:59.111092 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a1d839a1-4174-4abc-9e3d-5579603687e6-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"a1d839a1-4174-4abc-9e3d-5579603687e6\") " pod="openstack/rabbitmq-server-0" Dec 11 08:32:59 crc kubenswrapper[4860]: I1211 08:32:59.117119 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-687sn\" (UniqueName: \"kubernetes.io/projected/a1d839a1-4174-4abc-9e3d-5579603687e6-kube-api-access-687sn\") pod \"rabbitmq-server-0\" (UID: \"a1d839a1-4174-4abc-9e3d-5579603687e6\") " pod="openstack/rabbitmq-server-0" Dec 11 08:32:59 crc kubenswrapper[4860]: I1211 08:32:59.140251 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"a1d839a1-4174-4abc-9e3d-5579603687e6\") " pod="openstack/rabbitmq-server-0" Dec 11 08:32:59 crc kubenswrapper[4860]: I1211 08:32:59.233286 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 11 08:32:59 crc kubenswrapper[4860]: I1211 08:32:59.300698 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:32:59 crc kubenswrapper[4860]: I1211 08:32:59.396923 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"a004cb7b-ee08-44c0-9dc7-d0b69509fe13\" (UID: \"a004cb7b-ee08-44c0-9dc7-d0b69509fe13\") " Dec 11 08:32:59 crc kubenswrapper[4860]: I1211 08:32:59.397407 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a004cb7b-ee08-44c0-9dc7-d0b69509fe13-rabbitmq-confd\") pod \"a004cb7b-ee08-44c0-9dc7-d0b69509fe13\" (UID: \"a004cb7b-ee08-44c0-9dc7-d0b69509fe13\") " Dec 11 08:32:59 crc kubenswrapper[4860]: I1211 08:32:59.397447 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a004cb7b-ee08-44c0-9dc7-d0b69509fe13-pod-info\") pod \"a004cb7b-ee08-44c0-9dc7-d0b69509fe13\" (UID: \"a004cb7b-ee08-44c0-9dc7-d0b69509fe13\") " Dec 11 08:32:59 crc kubenswrapper[4860]: I1211 08:32:59.397510 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a004cb7b-ee08-44c0-9dc7-d0b69509fe13-rabbitmq-erlang-cookie\") pod \"a004cb7b-ee08-44c0-9dc7-d0b69509fe13\" (UID: \"a004cb7b-ee08-44c0-9dc7-d0b69509fe13\") " Dec 11 08:32:59 crc kubenswrapper[4860]: I1211 08:32:59.397564 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a004cb7b-ee08-44c0-9dc7-d0b69509fe13-config-data\") pod \"a004cb7b-ee08-44c0-9dc7-d0b69509fe13\" (UID: \"a004cb7b-ee08-44c0-9dc7-d0b69509fe13\") " Dec 11 08:32:59 crc kubenswrapper[4860]: I1211 08:32:59.397597 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a004cb7b-ee08-44c0-9dc7-d0b69509fe13-server-conf\") pod \"a004cb7b-ee08-44c0-9dc7-d0b69509fe13\" (UID: \"a004cb7b-ee08-44c0-9dc7-d0b69509fe13\") " Dec 11 08:32:59 crc kubenswrapper[4860]: I1211 08:32:59.397623 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a004cb7b-ee08-44c0-9dc7-d0b69509fe13-rabbitmq-plugins\") pod \"a004cb7b-ee08-44c0-9dc7-d0b69509fe13\" (UID: \"a004cb7b-ee08-44c0-9dc7-d0b69509fe13\") " Dec 11 08:32:59 crc kubenswrapper[4860]: I1211 08:32:59.397749 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a004cb7b-ee08-44c0-9dc7-d0b69509fe13-rabbitmq-tls\") pod \"a004cb7b-ee08-44c0-9dc7-d0b69509fe13\" (UID: \"a004cb7b-ee08-44c0-9dc7-d0b69509fe13\") " Dec 11 08:32:59 crc kubenswrapper[4860]: I1211 08:32:59.397803 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a004cb7b-ee08-44c0-9dc7-d0b69509fe13-plugins-conf\") pod \"a004cb7b-ee08-44c0-9dc7-d0b69509fe13\" (UID: \"a004cb7b-ee08-44c0-9dc7-d0b69509fe13\") " Dec 11 08:32:59 crc kubenswrapper[4860]: I1211 08:32:59.397840 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a004cb7b-ee08-44c0-9dc7-d0b69509fe13-erlang-cookie-secret\") pod \"a004cb7b-ee08-44c0-9dc7-d0b69509fe13\" (UID: \"a004cb7b-ee08-44c0-9dc7-d0b69509fe13\") " Dec 11 08:32:59 crc kubenswrapper[4860]: I1211 08:32:59.397872 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hkscb\" (UniqueName: \"kubernetes.io/projected/a004cb7b-ee08-44c0-9dc7-d0b69509fe13-kube-api-access-hkscb\") pod \"a004cb7b-ee08-44c0-9dc7-d0b69509fe13\" (UID: \"a004cb7b-ee08-44c0-9dc7-d0b69509fe13\") " Dec 11 08:32:59 crc kubenswrapper[4860]: I1211 08:32:59.400014 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a004cb7b-ee08-44c0-9dc7-d0b69509fe13-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "a004cb7b-ee08-44c0-9dc7-d0b69509fe13" (UID: "a004cb7b-ee08-44c0-9dc7-d0b69509fe13"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:32:59 crc kubenswrapper[4860]: I1211 08:32:59.402485 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a004cb7b-ee08-44c0-9dc7-d0b69509fe13-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "a004cb7b-ee08-44c0-9dc7-d0b69509fe13" (UID: "a004cb7b-ee08-44c0-9dc7-d0b69509fe13"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:32:59 crc kubenswrapper[4860]: I1211 08:32:59.404727 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a004cb7b-ee08-44c0-9dc7-d0b69509fe13-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "a004cb7b-ee08-44c0-9dc7-d0b69509fe13" (UID: "a004cb7b-ee08-44c0-9dc7-d0b69509fe13"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:32:59 crc kubenswrapper[4860]: I1211 08:32:59.406864 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a004cb7b-ee08-44c0-9dc7-d0b69509fe13-kube-api-access-hkscb" (OuterVolumeSpecName: "kube-api-access-hkscb") pod "a004cb7b-ee08-44c0-9dc7-d0b69509fe13" (UID: "a004cb7b-ee08-44c0-9dc7-d0b69509fe13"). InnerVolumeSpecName "kube-api-access-hkscb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:32:59 crc kubenswrapper[4860]: I1211 08:32:59.415826 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a004cb7b-ee08-44c0-9dc7-d0b69509fe13-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "a004cb7b-ee08-44c0-9dc7-d0b69509fe13" (UID: "a004cb7b-ee08-44c0-9dc7-d0b69509fe13"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:32:59 crc kubenswrapper[4860]: I1211 08:32:59.422344 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "persistence") pod "a004cb7b-ee08-44c0-9dc7-d0b69509fe13" (UID: "a004cb7b-ee08-44c0-9dc7-d0b69509fe13"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 11 08:32:59 crc kubenswrapper[4860]: I1211 08:32:59.432710 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/a004cb7b-ee08-44c0-9dc7-d0b69509fe13-pod-info" (OuterVolumeSpecName: "pod-info") pod "a004cb7b-ee08-44c0-9dc7-d0b69509fe13" (UID: "a004cb7b-ee08-44c0-9dc7-d0b69509fe13"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 11 08:32:59 crc kubenswrapper[4860]: I1211 08:32:59.443476 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a004cb7b-ee08-44c0-9dc7-d0b69509fe13-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "a004cb7b-ee08-44c0-9dc7-d0b69509fe13" (UID: "a004cb7b-ee08-44c0-9dc7-d0b69509fe13"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:32:59 crc kubenswrapper[4860]: I1211 08:32:59.474309 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a004cb7b-ee08-44c0-9dc7-d0b69509fe13-config-data" (OuterVolumeSpecName: "config-data") pod "a004cb7b-ee08-44c0-9dc7-d0b69509fe13" (UID: "a004cb7b-ee08-44c0-9dc7-d0b69509fe13"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:32:59 crc kubenswrapper[4860]: I1211 08:32:59.503214 4860 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a004cb7b-ee08-44c0-9dc7-d0b69509fe13-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 11 08:32:59 crc kubenswrapper[4860]: I1211 08:32:59.503258 4860 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a004cb7b-ee08-44c0-9dc7-d0b69509fe13-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 11 08:32:59 crc kubenswrapper[4860]: I1211 08:32:59.503271 4860 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a004cb7b-ee08-44c0-9dc7-d0b69509fe13-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 11 08:32:59 crc kubenswrapper[4860]: I1211 08:32:59.503284 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hkscb\" (UniqueName: \"kubernetes.io/projected/a004cb7b-ee08-44c0-9dc7-d0b69509fe13-kube-api-access-hkscb\") on node \"crc\" DevicePath \"\"" Dec 11 08:32:59 crc kubenswrapper[4860]: I1211 08:32:59.503316 4860 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Dec 11 08:32:59 crc kubenswrapper[4860]: I1211 08:32:59.503328 4860 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a004cb7b-ee08-44c0-9dc7-d0b69509fe13-pod-info\") on node \"crc\" DevicePath \"\"" Dec 11 08:32:59 crc kubenswrapper[4860]: I1211 08:32:59.503343 4860 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a004cb7b-ee08-44c0-9dc7-d0b69509fe13-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 11 08:32:59 crc kubenswrapper[4860]: I1211 08:32:59.503354 4860 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a004cb7b-ee08-44c0-9dc7-d0b69509fe13-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 08:32:59 crc kubenswrapper[4860]: I1211 08:32:59.503365 4860 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a004cb7b-ee08-44c0-9dc7-d0b69509fe13-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 11 08:32:59 crc kubenswrapper[4860]: I1211 08:32:59.518172 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a004cb7b-ee08-44c0-9dc7-d0b69509fe13-server-conf" (OuterVolumeSpecName: "server-conf") pod "a004cb7b-ee08-44c0-9dc7-d0b69509fe13" (UID: "a004cb7b-ee08-44c0-9dc7-d0b69509fe13"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:32:59 crc kubenswrapper[4860]: I1211 08:32:59.530549 4860 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Dec 11 08:32:59 crc kubenswrapper[4860]: I1211 08:32:59.606415 4860 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Dec 11 08:32:59 crc kubenswrapper[4860]: I1211 08:32:59.606771 4860 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a004cb7b-ee08-44c0-9dc7-d0b69509fe13-server-conf\") on node \"crc\" DevicePath \"\"" Dec 11 08:32:59 crc kubenswrapper[4860]: I1211 08:32:59.609058 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="97634ec0-7953-4e38-a07e-95967ce874d7" path="/var/lib/kubelet/pods/97634ec0-7953-4e38-a07e-95967ce874d7/volumes" Dec 11 08:32:59 crc kubenswrapper[4860]: I1211 08:32:59.613247 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 11 08:32:59 crc kubenswrapper[4860]: I1211 08:32:59.642947 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a004cb7b-ee08-44c0-9dc7-d0b69509fe13-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "a004cb7b-ee08-44c0-9dc7-d0b69509fe13" (UID: "a004cb7b-ee08-44c0-9dc7-d0b69509fe13"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:32:59 crc kubenswrapper[4860]: I1211 08:32:59.708725 4860 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a004cb7b-ee08-44c0-9dc7-d0b69509fe13-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 11 08:32:59 crc kubenswrapper[4860]: I1211 08:32:59.803695 4860 generic.go:334] "Generic (PLEG): container finished" podID="a004cb7b-ee08-44c0-9dc7-d0b69509fe13" containerID="c1b416ccd0ce4787808ac0bc1fe677baedfb2f3a87f1c9829d5cf793665cbe0c" exitCode=0 Dec 11 08:32:59 crc kubenswrapper[4860]: I1211 08:32:59.803803 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"a004cb7b-ee08-44c0-9dc7-d0b69509fe13","Type":"ContainerDied","Data":"c1b416ccd0ce4787808ac0bc1fe677baedfb2f3a87f1c9829d5cf793665cbe0c"} Dec 11 08:32:59 crc kubenswrapper[4860]: I1211 08:32:59.803841 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"a004cb7b-ee08-44c0-9dc7-d0b69509fe13","Type":"ContainerDied","Data":"ab72d932df3fb4339b63c207dbbb48deec5ee7b26febe2b3ac79346f2e9282fa"} Dec 11 08:32:59 crc kubenswrapper[4860]: I1211 08:32:59.803865 4860 scope.go:117] "RemoveContainer" containerID="c1b416ccd0ce4787808ac0bc1fe677baedfb2f3a87f1c9829d5cf793665cbe0c" Dec 11 08:32:59 crc kubenswrapper[4860]: I1211 08:32:59.804055 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:32:59 crc kubenswrapper[4860]: I1211 08:32:59.809722 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"a1d839a1-4174-4abc-9e3d-5579603687e6","Type":"ContainerStarted","Data":"f516c623cc8063b3e31f6443090992ad0803ffe204b84ed2067a92564ba2f38c"} Dec 11 08:32:59 crc kubenswrapper[4860]: I1211 08:32:59.837136 4860 scope.go:117] "RemoveContainer" containerID="7b2484370445a6fcc6a996d0a46f40caeeb99158af08c51c6b35e54bd1ef3322" Dec 11 08:32:59 crc kubenswrapper[4860]: I1211 08:32:59.852981 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 11 08:32:59 crc kubenswrapper[4860]: I1211 08:32:59.862990 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 11 08:32:59 crc kubenswrapper[4860]: I1211 08:32:59.886776 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 11 08:32:59 crc kubenswrapper[4860]: E1211 08:32:59.887241 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a004cb7b-ee08-44c0-9dc7-d0b69509fe13" containerName="rabbitmq" Dec 11 08:32:59 crc kubenswrapper[4860]: I1211 08:32:59.887259 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="a004cb7b-ee08-44c0-9dc7-d0b69509fe13" containerName="rabbitmq" Dec 11 08:32:59 crc kubenswrapper[4860]: E1211 08:32:59.887284 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a004cb7b-ee08-44c0-9dc7-d0b69509fe13" containerName="setup-container" Dec 11 08:32:59 crc kubenswrapper[4860]: I1211 08:32:59.887292 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="a004cb7b-ee08-44c0-9dc7-d0b69509fe13" containerName="setup-container" Dec 11 08:32:59 crc kubenswrapper[4860]: I1211 08:32:59.887522 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="a004cb7b-ee08-44c0-9dc7-d0b69509fe13" containerName="rabbitmq" Dec 11 08:32:59 crc kubenswrapper[4860]: I1211 08:32:59.888606 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:32:59 crc kubenswrapper[4860]: I1211 08:32:59.889540 4860 scope.go:117] "RemoveContainer" containerID="c1b416ccd0ce4787808ac0bc1fe677baedfb2f3a87f1c9829d5cf793665cbe0c" Dec 11 08:32:59 crc kubenswrapper[4860]: E1211 08:32:59.890200 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c1b416ccd0ce4787808ac0bc1fe677baedfb2f3a87f1c9829d5cf793665cbe0c\": container with ID starting with c1b416ccd0ce4787808ac0bc1fe677baedfb2f3a87f1c9829d5cf793665cbe0c not found: ID does not exist" containerID="c1b416ccd0ce4787808ac0bc1fe677baedfb2f3a87f1c9829d5cf793665cbe0c" Dec 11 08:32:59 crc kubenswrapper[4860]: I1211 08:32:59.890232 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c1b416ccd0ce4787808ac0bc1fe677baedfb2f3a87f1c9829d5cf793665cbe0c"} err="failed to get container status \"c1b416ccd0ce4787808ac0bc1fe677baedfb2f3a87f1c9829d5cf793665cbe0c\": rpc error: code = NotFound desc = could not find container \"c1b416ccd0ce4787808ac0bc1fe677baedfb2f3a87f1c9829d5cf793665cbe0c\": container with ID starting with c1b416ccd0ce4787808ac0bc1fe677baedfb2f3a87f1c9829d5cf793665cbe0c not found: ID does not exist" Dec 11 08:32:59 crc kubenswrapper[4860]: I1211 08:32:59.890259 4860 scope.go:117] "RemoveContainer" containerID="7b2484370445a6fcc6a996d0a46f40caeeb99158af08c51c6b35e54bd1ef3322" Dec 11 08:32:59 crc kubenswrapper[4860]: E1211 08:32:59.890707 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7b2484370445a6fcc6a996d0a46f40caeeb99158af08c51c6b35e54bd1ef3322\": container with ID starting with 7b2484370445a6fcc6a996d0a46f40caeeb99158af08c51c6b35e54bd1ef3322 not found: ID does not exist" containerID="7b2484370445a6fcc6a996d0a46f40caeeb99158af08c51c6b35e54bd1ef3322" Dec 11 08:32:59 crc kubenswrapper[4860]: I1211 08:32:59.890741 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b2484370445a6fcc6a996d0a46f40caeeb99158af08c51c6b35e54bd1ef3322"} err="failed to get container status \"7b2484370445a6fcc6a996d0a46f40caeeb99158af08c51c6b35e54bd1ef3322\": rpc error: code = NotFound desc = could not find container \"7b2484370445a6fcc6a996d0a46f40caeeb99158af08c51c6b35e54bd1ef3322\": container with ID starting with 7b2484370445a6fcc6a996d0a46f40caeeb99158af08c51c6b35e54bd1ef3322 not found: ID does not exist" Dec 11 08:32:59 crc kubenswrapper[4860]: I1211 08:32:59.891712 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 11 08:32:59 crc kubenswrapper[4860]: I1211 08:32:59.891927 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 11 08:32:59 crc kubenswrapper[4860]: I1211 08:32:59.892197 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 11 08:32:59 crc kubenswrapper[4860]: I1211 08:32:59.892600 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 11 08:32:59 crc kubenswrapper[4860]: I1211 08:32:59.892660 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 11 08:32:59 crc kubenswrapper[4860]: I1211 08:32:59.892804 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 11 08:32:59 crc kubenswrapper[4860]: I1211 08:32:59.892883 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-vmvgn" Dec 11 08:32:59 crc kubenswrapper[4860]: I1211 08:32:59.918939 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 11 08:33:00 crc kubenswrapper[4860]: I1211 08:33:00.018711 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/9cfe9bd2-a1fe-4cae-8f48-c007bfd95961-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"9cfe9bd2-a1fe-4cae-8f48-c007bfd95961\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:33:00 crc kubenswrapper[4860]: I1211 08:33:00.018789 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/9cfe9bd2-a1fe-4cae-8f48-c007bfd95961-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"9cfe9bd2-a1fe-4cae-8f48-c007bfd95961\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:33:00 crc kubenswrapper[4860]: I1211 08:33:00.018832 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/9cfe9bd2-a1fe-4cae-8f48-c007bfd95961-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"9cfe9bd2-a1fe-4cae-8f48-c007bfd95961\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:33:00 crc kubenswrapper[4860]: I1211 08:33:00.019184 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/9cfe9bd2-a1fe-4cae-8f48-c007bfd95961-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"9cfe9bd2-a1fe-4cae-8f48-c007bfd95961\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:33:00 crc kubenswrapper[4860]: I1211 08:33:00.019383 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/9cfe9bd2-a1fe-4cae-8f48-c007bfd95961-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"9cfe9bd2-a1fe-4cae-8f48-c007bfd95961\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:33:00 crc kubenswrapper[4860]: I1211 08:33:00.019430 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/9cfe9bd2-a1fe-4cae-8f48-c007bfd95961-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"9cfe9bd2-a1fe-4cae-8f48-c007bfd95961\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:33:00 crc kubenswrapper[4860]: I1211 08:33:00.019490 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9cfe9bd2-a1fe-4cae-8f48-c007bfd95961-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"9cfe9bd2-a1fe-4cae-8f48-c007bfd95961\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:33:00 crc kubenswrapper[4860]: I1211 08:33:00.019635 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ssm5q\" (UniqueName: \"kubernetes.io/projected/9cfe9bd2-a1fe-4cae-8f48-c007bfd95961-kube-api-access-ssm5q\") pod \"rabbitmq-cell1-server-0\" (UID: \"9cfe9bd2-a1fe-4cae-8f48-c007bfd95961\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:33:00 crc kubenswrapper[4860]: I1211 08:33:00.019695 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/9cfe9bd2-a1fe-4cae-8f48-c007bfd95961-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"9cfe9bd2-a1fe-4cae-8f48-c007bfd95961\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:33:00 crc kubenswrapper[4860]: I1211 08:33:00.019715 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/9cfe9bd2-a1fe-4cae-8f48-c007bfd95961-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"9cfe9bd2-a1fe-4cae-8f48-c007bfd95961\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:33:00 crc kubenswrapper[4860]: I1211 08:33:00.019748 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"9cfe9bd2-a1fe-4cae-8f48-c007bfd95961\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:33:00 crc kubenswrapper[4860]: I1211 08:33:00.122218 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9cfe9bd2-a1fe-4cae-8f48-c007bfd95961-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"9cfe9bd2-a1fe-4cae-8f48-c007bfd95961\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:33:00 crc kubenswrapper[4860]: I1211 08:33:00.122624 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ssm5q\" (UniqueName: \"kubernetes.io/projected/9cfe9bd2-a1fe-4cae-8f48-c007bfd95961-kube-api-access-ssm5q\") pod \"rabbitmq-cell1-server-0\" (UID: \"9cfe9bd2-a1fe-4cae-8f48-c007bfd95961\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:33:00 crc kubenswrapper[4860]: I1211 08:33:00.122718 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/9cfe9bd2-a1fe-4cae-8f48-c007bfd95961-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"9cfe9bd2-a1fe-4cae-8f48-c007bfd95961\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:33:00 crc kubenswrapper[4860]: I1211 08:33:00.123288 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/9cfe9bd2-a1fe-4cae-8f48-c007bfd95961-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"9cfe9bd2-a1fe-4cae-8f48-c007bfd95961\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:33:00 crc kubenswrapper[4860]: I1211 08:33:00.122748 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/9cfe9bd2-a1fe-4cae-8f48-c007bfd95961-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"9cfe9bd2-a1fe-4cae-8f48-c007bfd95961\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:33:00 crc kubenswrapper[4860]: I1211 08:33:00.123381 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9cfe9bd2-a1fe-4cae-8f48-c007bfd95961-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"9cfe9bd2-a1fe-4cae-8f48-c007bfd95961\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:33:00 crc kubenswrapper[4860]: I1211 08:33:00.123416 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"9cfe9bd2-a1fe-4cae-8f48-c007bfd95961\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:33:00 crc kubenswrapper[4860]: I1211 08:33:00.123393 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/9cfe9bd2-a1fe-4cae-8f48-c007bfd95961-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"9cfe9bd2-a1fe-4cae-8f48-c007bfd95961\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:33:00 crc kubenswrapper[4860]: I1211 08:33:00.124051 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/9cfe9bd2-a1fe-4cae-8f48-c007bfd95961-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"9cfe9bd2-a1fe-4cae-8f48-c007bfd95961\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:33:00 crc kubenswrapper[4860]: I1211 08:33:00.124099 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/9cfe9bd2-a1fe-4cae-8f48-c007bfd95961-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"9cfe9bd2-a1fe-4cae-8f48-c007bfd95961\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:33:00 crc kubenswrapper[4860]: I1211 08:33:00.124184 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/9cfe9bd2-a1fe-4cae-8f48-c007bfd95961-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"9cfe9bd2-a1fe-4cae-8f48-c007bfd95961\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:33:00 crc kubenswrapper[4860]: I1211 08:33:00.124304 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/9cfe9bd2-a1fe-4cae-8f48-c007bfd95961-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"9cfe9bd2-a1fe-4cae-8f48-c007bfd95961\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:33:00 crc kubenswrapper[4860]: I1211 08:33:00.124377 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/9cfe9bd2-a1fe-4cae-8f48-c007bfd95961-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"9cfe9bd2-a1fe-4cae-8f48-c007bfd95961\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:33:00 crc kubenswrapper[4860]: I1211 08:33:00.124403 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/9cfe9bd2-a1fe-4cae-8f48-c007bfd95961-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"9cfe9bd2-a1fe-4cae-8f48-c007bfd95961\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:33:00 crc kubenswrapper[4860]: I1211 08:33:00.123959 4860 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"9cfe9bd2-a1fe-4cae-8f48-c007bfd95961\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:33:00 crc kubenswrapper[4860]: I1211 08:33:00.126431 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/9cfe9bd2-a1fe-4cae-8f48-c007bfd95961-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"9cfe9bd2-a1fe-4cae-8f48-c007bfd95961\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:33:00 crc kubenswrapper[4860]: I1211 08:33:00.126882 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/9cfe9bd2-a1fe-4cae-8f48-c007bfd95961-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"9cfe9bd2-a1fe-4cae-8f48-c007bfd95961\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:33:00 crc kubenswrapper[4860]: I1211 08:33:00.136521 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/9cfe9bd2-a1fe-4cae-8f48-c007bfd95961-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"9cfe9bd2-a1fe-4cae-8f48-c007bfd95961\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:33:00 crc kubenswrapper[4860]: I1211 08:33:00.136614 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/9cfe9bd2-a1fe-4cae-8f48-c007bfd95961-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"9cfe9bd2-a1fe-4cae-8f48-c007bfd95961\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:33:00 crc kubenswrapper[4860]: I1211 08:33:00.136803 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/9cfe9bd2-a1fe-4cae-8f48-c007bfd95961-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"9cfe9bd2-a1fe-4cae-8f48-c007bfd95961\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:33:00 crc kubenswrapper[4860]: I1211 08:33:00.137861 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/9cfe9bd2-a1fe-4cae-8f48-c007bfd95961-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"9cfe9bd2-a1fe-4cae-8f48-c007bfd95961\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:33:00 crc kubenswrapper[4860]: I1211 08:33:00.140416 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ssm5q\" (UniqueName: \"kubernetes.io/projected/9cfe9bd2-a1fe-4cae-8f48-c007bfd95961-kube-api-access-ssm5q\") pod \"rabbitmq-cell1-server-0\" (UID: \"9cfe9bd2-a1fe-4cae-8f48-c007bfd95961\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:33:00 crc kubenswrapper[4860]: I1211 08:33:00.163733 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"9cfe9bd2-a1fe-4cae-8f48-c007bfd95961\") " pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:33:00 crc kubenswrapper[4860]: I1211 08:33:00.239172 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:33:00 crc kubenswrapper[4860]: I1211 08:33:00.587548 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 11 08:33:00 crc kubenswrapper[4860]: W1211 08:33:00.591947 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9cfe9bd2_a1fe_4cae_8f48_c007bfd95961.slice/crio-3bf0adc60002d21a8d36ba02a8ddaea1af0eee741bdb046b56fee877ae79a3b2 WatchSource:0}: Error finding container 3bf0adc60002d21a8d36ba02a8ddaea1af0eee741bdb046b56fee877ae79a3b2: Status 404 returned error can't find the container with id 3bf0adc60002d21a8d36ba02a8ddaea1af0eee741bdb046b56fee877ae79a3b2 Dec 11 08:33:00 crc kubenswrapper[4860]: I1211 08:33:00.823079 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"9cfe9bd2-a1fe-4cae-8f48-c007bfd95961","Type":"ContainerStarted","Data":"3bf0adc60002d21a8d36ba02a8ddaea1af0eee741bdb046b56fee877ae79a3b2"} Dec 11 08:33:00 crc kubenswrapper[4860]: I1211 08:33:00.955253 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5576978c7c-2m9pb"] Dec 11 08:33:00 crc kubenswrapper[4860]: I1211 08:33:00.957492 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5576978c7c-2m9pb" Dec 11 08:33:00 crc kubenswrapper[4860]: I1211 08:33:00.959869 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Dec 11 08:33:00 crc kubenswrapper[4860]: I1211 08:33:00.973296 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5576978c7c-2m9pb"] Dec 11 08:33:01 crc kubenswrapper[4860]: I1211 08:33:01.133114 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ecccf7e4-15f8-4a10-9662-c188d5ecc068-dns-svc\") pod \"dnsmasq-dns-5576978c7c-2m9pb\" (UID: \"ecccf7e4-15f8-4a10-9662-c188d5ecc068\") " pod="openstack/dnsmasq-dns-5576978c7c-2m9pb" Dec 11 08:33:01 crc kubenswrapper[4860]: I1211 08:33:01.133625 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vbclz\" (UniqueName: \"kubernetes.io/projected/ecccf7e4-15f8-4a10-9662-c188d5ecc068-kube-api-access-vbclz\") pod \"dnsmasq-dns-5576978c7c-2m9pb\" (UID: \"ecccf7e4-15f8-4a10-9662-c188d5ecc068\") " pod="openstack/dnsmasq-dns-5576978c7c-2m9pb" Dec 11 08:33:01 crc kubenswrapper[4860]: I1211 08:33:01.133694 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/ecccf7e4-15f8-4a10-9662-c188d5ecc068-openstack-edpm-ipam\") pod \"dnsmasq-dns-5576978c7c-2m9pb\" (UID: \"ecccf7e4-15f8-4a10-9662-c188d5ecc068\") " pod="openstack/dnsmasq-dns-5576978c7c-2m9pb" Dec 11 08:33:01 crc kubenswrapper[4860]: I1211 08:33:01.133757 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ecccf7e4-15f8-4a10-9662-c188d5ecc068-dns-swift-storage-0\") pod \"dnsmasq-dns-5576978c7c-2m9pb\" (UID: \"ecccf7e4-15f8-4a10-9662-c188d5ecc068\") " pod="openstack/dnsmasq-dns-5576978c7c-2m9pb" Dec 11 08:33:01 crc kubenswrapper[4860]: I1211 08:33:01.133807 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ecccf7e4-15f8-4a10-9662-c188d5ecc068-config\") pod \"dnsmasq-dns-5576978c7c-2m9pb\" (UID: \"ecccf7e4-15f8-4a10-9662-c188d5ecc068\") " pod="openstack/dnsmasq-dns-5576978c7c-2m9pb" Dec 11 08:33:01 crc kubenswrapper[4860]: I1211 08:33:01.133841 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ecccf7e4-15f8-4a10-9662-c188d5ecc068-ovsdbserver-sb\") pod \"dnsmasq-dns-5576978c7c-2m9pb\" (UID: \"ecccf7e4-15f8-4a10-9662-c188d5ecc068\") " pod="openstack/dnsmasq-dns-5576978c7c-2m9pb" Dec 11 08:33:01 crc kubenswrapper[4860]: I1211 08:33:01.133928 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ecccf7e4-15f8-4a10-9662-c188d5ecc068-ovsdbserver-nb\") pod \"dnsmasq-dns-5576978c7c-2m9pb\" (UID: \"ecccf7e4-15f8-4a10-9662-c188d5ecc068\") " pod="openstack/dnsmasq-dns-5576978c7c-2m9pb" Dec 11 08:33:01 crc kubenswrapper[4860]: I1211 08:33:01.235243 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ecccf7e4-15f8-4a10-9662-c188d5ecc068-dns-svc\") pod \"dnsmasq-dns-5576978c7c-2m9pb\" (UID: \"ecccf7e4-15f8-4a10-9662-c188d5ecc068\") " pod="openstack/dnsmasq-dns-5576978c7c-2m9pb" Dec 11 08:33:01 crc kubenswrapper[4860]: I1211 08:33:01.235317 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vbclz\" (UniqueName: \"kubernetes.io/projected/ecccf7e4-15f8-4a10-9662-c188d5ecc068-kube-api-access-vbclz\") pod \"dnsmasq-dns-5576978c7c-2m9pb\" (UID: \"ecccf7e4-15f8-4a10-9662-c188d5ecc068\") " pod="openstack/dnsmasq-dns-5576978c7c-2m9pb" Dec 11 08:33:01 crc kubenswrapper[4860]: I1211 08:33:01.235337 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/ecccf7e4-15f8-4a10-9662-c188d5ecc068-openstack-edpm-ipam\") pod \"dnsmasq-dns-5576978c7c-2m9pb\" (UID: \"ecccf7e4-15f8-4a10-9662-c188d5ecc068\") " pod="openstack/dnsmasq-dns-5576978c7c-2m9pb" Dec 11 08:33:01 crc kubenswrapper[4860]: I1211 08:33:01.235375 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ecccf7e4-15f8-4a10-9662-c188d5ecc068-dns-swift-storage-0\") pod \"dnsmasq-dns-5576978c7c-2m9pb\" (UID: \"ecccf7e4-15f8-4a10-9662-c188d5ecc068\") " pod="openstack/dnsmasq-dns-5576978c7c-2m9pb" Dec 11 08:33:01 crc kubenswrapper[4860]: I1211 08:33:01.235409 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ecccf7e4-15f8-4a10-9662-c188d5ecc068-config\") pod \"dnsmasq-dns-5576978c7c-2m9pb\" (UID: \"ecccf7e4-15f8-4a10-9662-c188d5ecc068\") " pod="openstack/dnsmasq-dns-5576978c7c-2m9pb" Dec 11 08:33:01 crc kubenswrapper[4860]: I1211 08:33:01.235429 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ecccf7e4-15f8-4a10-9662-c188d5ecc068-ovsdbserver-sb\") pod \"dnsmasq-dns-5576978c7c-2m9pb\" (UID: \"ecccf7e4-15f8-4a10-9662-c188d5ecc068\") " pod="openstack/dnsmasq-dns-5576978c7c-2m9pb" Dec 11 08:33:01 crc kubenswrapper[4860]: I1211 08:33:01.235487 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ecccf7e4-15f8-4a10-9662-c188d5ecc068-ovsdbserver-nb\") pod \"dnsmasq-dns-5576978c7c-2m9pb\" (UID: \"ecccf7e4-15f8-4a10-9662-c188d5ecc068\") " pod="openstack/dnsmasq-dns-5576978c7c-2m9pb" Dec 11 08:33:01 crc kubenswrapper[4860]: I1211 08:33:01.236338 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ecccf7e4-15f8-4a10-9662-c188d5ecc068-ovsdbserver-nb\") pod \"dnsmasq-dns-5576978c7c-2m9pb\" (UID: \"ecccf7e4-15f8-4a10-9662-c188d5ecc068\") " pod="openstack/dnsmasq-dns-5576978c7c-2m9pb" Dec 11 08:33:01 crc kubenswrapper[4860]: I1211 08:33:01.237428 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ecccf7e4-15f8-4a10-9662-c188d5ecc068-config\") pod \"dnsmasq-dns-5576978c7c-2m9pb\" (UID: \"ecccf7e4-15f8-4a10-9662-c188d5ecc068\") " pod="openstack/dnsmasq-dns-5576978c7c-2m9pb" Dec 11 08:33:01 crc kubenswrapper[4860]: I1211 08:33:01.237556 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ecccf7e4-15f8-4a10-9662-c188d5ecc068-dns-svc\") pod \"dnsmasq-dns-5576978c7c-2m9pb\" (UID: \"ecccf7e4-15f8-4a10-9662-c188d5ecc068\") " pod="openstack/dnsmasq-dns-5576978c7c-2m9pb" Dec 11 08:33:01 crc kubenswrapper[4860]: I1211 08:33:01.237741 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ecccf7e4-15f8-4a10-9662-c188d5ecc068-dns-swift-storage-0\") pod \"dnsmasq-dns-5576978c7c-2m9pb\" (UID: \"ecccf7e4-15f8-4a10-9662-c188d5ecc068\") " pod="openstack/dnsmasq-dns-5576978c7c-2m9pb" Dec 11 08:33:01 crc kubenswrapper[4860]: I1211 08:33:01.237799 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ecccf7e4-15f8-4a10-9662-c188d5ecc068-ovsdbserver-sb\") pod \"dnsmasq-dns-5576978c7c-2m9pb\" (UID: \"ecccf7e4-15f8-4a10-9662-c188d5ecc068\") " pod="openstack/dnsmasq-dns-5576978c7c-2m9pb" Dec 11 08:33:01 crc kubenswrapper[4860]: I1211 08:33:01.238529 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/ecccf7e4-15f8-4a10-9662-c188d5ecc068-openstack-edpm-ipam\") pod \"dnsmasq-dns-5576978c7c-2m9pb\" (UID: \"ecccf7e4-15f8-4a10-9662-c188d5ecc068\") " pod="openstack/dnsmasq-dns-5576978c7c-2m9pb" Dec 11 08:33:01 crc kubenswrapper[4860]: I1211 08:33:01.270031 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vbclz\" (UniqueName: \"kubernetes.io/projected/ecccf7e4-15f8-4a10-9662-c188d5ecc068-kube-api-access-vbclz\") pod \"dnsmasq-dns-5576978c7c-2m9pb\" (UID: \"ecccf7e4-15f8-4a10-9662-c188d5ecc068\") " pod="openstack/dnsmasq-dns-5576978c7c-2m9pb" Dec 11 08:33:01 crc kubenswrapper[4860]: I1211 08:33:01.302874 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5576978c7c-2m9pb" Dec 11 08:33:01 crc kubenswrapper[4860]: I1211 08:33:01.601279 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a004cb7b-ee08-44c0-9dc7-d0b69509fe13" path="/var/lib/kubelet/pods/a004cb7b-ee08-44c0-9dc7-d0b69509fe13/volumes" Dec 11 08:33:01 crc kubenswrapper[4860]: I1211 08:33:01.773177 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5576978c7c-2m9pb"] Dec 11 08:33:01 crc kubenswrapper[4860]: W1211 08:33:01.776364 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podecccf7e4_15f8_4a10_9662_c188d5ecc068.slice/crio-7bed0057c9d4a3cc2cd037fe0c07dd5ecbf8fc53b13a00e03f87c9ca90a0f1bc WatchSource:0}: Error finding container 7bed0057c9d4a3cc2cd037fe0c07dd5ecbf8fc53b13a00e03f87c9ca90a0f1bc: Status 404 returned error can't find the container with id 7bed0057c9d4a3cc2cd037fe0c07dd5ecbf8fc53b13a00e03f87c9ca90a0f1bc Dec 11 08:33:01 crc kubenswrapper[4860]: I1211 08:33:01.838693 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"a1d839a1-4174-4abc-9e3d-5579603687e6","Type":"ContainerStarted","Data":"1983955f161e6974353df114f1902b23c7065b295b78a3af0755e8cbbdcbb3a6"} Dec 11 08:33:01 crc kubenswrapper[4860]: I1211 08:33:01.839929 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5576978c7c-2m9pb" event={"ID":"ecccf7e4-15f8-4a10-9662-c188d5ecc068","Type":"ContainerStarted","Data":"7bed0057c9d4a3cc2cd037fe0c07dd5ecbf8fc53b13a00e03f87c9ca90a0f1bc"} Dec 11 08:33:02 crc kubenswrapper[4860]: I1211 08:33:02.852904 4860 generic.go:334] "Generic (PLEG): container finished" podID="ecccf7e4-15f8-4a10-9662-c188d5ecc068" containerID="663607f4f83dc73084a650362934ec4f5ca40a77c0d419cedf68471a77a18b12" exitCode=0 Dec 11 08:33:02 crc kubenswrapper[4860]: I1211 08:33:02.852995 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5576978c7c-2m9pb" event={"ID":"ecccf7e4-15f8-4a10-9662-c188d5ecc068","Type":"ContainerDied","Data":"663607f4f83dc73084a650362934ec4f5ca40a77c0d419cedf68471a77a18b12"} Dec 11 08:33:02 crc kubenswrapper[4860]: I1211 08:33:02.855625 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"9cfe9bd2-a1fe-4cae-8f48-c007bfd95961","Type":"ContainerStarted","Data":"fdfdcc293ae6a46b533c028cac1585240f64ab520d8db90fadd9acdae90d9bf8"} Dec 11 08:33:03 crc kubenswrapper[4860]: I1211 08:33:03.867176 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5576978c7c-2m9pb" event={"ID":"ecccf7e4-15f8-4a10-9662-c188d5ecc068","Type":"ContainerStarted","Data":"02ef99652be9a72d174fd83a2d766b321cdae7dd835f28989bbf269ec696161b"} Dec 11 08:33:03 crc kubenswrapper[4860]: I1211 08:33:03.888453 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5576978c7c-2m9pb" podStartSLOduration=3.888418346 podStartE2EDuration="3.888418346s" podCreationTimestamp="2025-12-11 08:33:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:33:03.887242898 +0000 UTC m=+1316.615761953" watchObservedRunningTime="2025-12-11 08:33:03.888418346 +0000 UTC m=+1316.616937441" Dec 11 08:33:04 crc kubenswrapper[4860]: I1211 08:33:04.876928 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5576978c7c-2m9pb" Dec 11 08:33:11 crc kubenswrapper[4860]: I1211 08:33:11.303924 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5576978c7c-2m9pb" Dec 11 08:33:11 crc kubenswrapper[4860]: I1211 08:33:11.433620 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c7b6c5df9-7zcv8"] Dec 11 08:33:11 crc kubenswrapper[4860]: I1211 08:33:11.433903 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5c7b6c5df9-7zcv8" podUID="33f4ccd8-b2d7-43f7-a6f6-daf6ea2fa5b8" containerName="dnsmasq-dns" containerID="cri-o://0f40eceff896f6453086ec0213e1ebfaa19e60a191b1173daa71b9121db12111" gracePeriod=10 Dec 11 08:33:11 crc kubenswrapper[4860]: I1211 08:33:11.617760 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8c6f6df99-zx6zh"] Dec 11 08:33:11 crc kubenswrapper[4860]: I1211 08:33:11.621706 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8c6f6df99-zx6zh" Dec 11 08:33:11 crc kubenswrapper[4860]: I1211 08:33:11.653866 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8c6f6df99-zx6zh"] Dec 11 08:33:11 crc kubenswrapper[4860]: I1211 08:33:11.684593 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2e567b4f-94b9-4662-825a-045e124c7948-ovsdbserver-nb\") pod \"dnsmasq-dns-8c6f6df99-zx6zh\" (UID: \"2e567b4f-94b9-4662-825a-045e124c7948\") " pod="openstack/dnsmasq-dns-8c6f6df99-zx6zh" Dec 11 08:33:11 crc kubenswrapper[4860]: I1211 08:33:11.684833 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/2e567b4f-94b9-4662-825a-045e124c7948-openstack-edpm-ipam\") pod \"dnsmasq-dns-8c6f6df99-zx6zh\" (UID: \"2e567b4f-94b9-4662-825a-045e124c7948\") " pod="openstack/dnsmasq-dns-8c6f6df99-zx6zh" Dec 11 08:33:11 crc kubenswrapper[4860]: I1211 08:33:11.684974 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2e567b4f-94b9-4662-825a-045e124c7948-config\") pod \"dnsmasq-dns-8c6f6df99-zx6zh\" (UID: \"2e567b4f-94b9-4662-825a-045e124c7948\") " pod="openstack/dnsmasq-dns-8c6f6df99-zx6zh" Dec 11 08:33:11 crc kubenswrapper[4860]: I1211 08:33:11.685024 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zlctn\" (UniqueName: \"kubernetes.io/projected/2e567b4f-94b9-4662-825a-045e124c7948-kube-api-access-zlctn\") pod \"dnsmasq-dns-8c6f6df99-zx6zh\" (UID: \"2e567b4f-94b9-4662-825a-045e124c7948\") " pod="openstack/dnsmasq-dns-8c6f6df99-zx6zh" Dec 11 08:33:11 crc kubenswrapper[4860]: I1211 08:33:11.685101 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2e567b4f-94b9-4662-825a-045e124c7948-dns-swift-storage-0\") pod \"dnsmasq-dns-8c6f6df99-zx6zh\" (UID: \"2e567b4f-94b9-4662-825a-045e124c7948\") " pod="openstack/dnsmasq-dns-8c6f6df99-zx6zh" Dec 11 08:33:11 crc kubenswrapper[4860]: I1211 08:33:11.685156 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2e567b4f-94b9-4662-825a-045e124c7948-ovsdbserver-sb\") pod \"dnsmasq-dns-8c6f6df99-zx6zh\" (UID: \"2e567b4f-94b9-4662-825a-045e124c7948\") " pod="openstack/dnsmasq-dns-8c6f6df99-zx6zh" Dec 11 08:33:11 crc kubenswrapper[4860]: I1211 08:33:11.685224 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2e567b4f-94b9-4662-825a-045e124c7948-dns-svc\") pod \"dnsmasq-dns-8c6f6df99-zx6zh\" (UID: \"2e567b4f-94b9-4662-825a-045e124c7948\") " pod="openstack/dnsmasq-dns-8c6f6df99-zx6zh" Dec 11 08:33:11 crc kubenswrapper[4860]: I1211 08:33:11.786856 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/2e567b4f-94b9-4662-825a-045e124c7948-openstack-edpm-ipam\") pod \"dnsmasq-dns-8c6f6df99-zx6zh\" (UID: \"2e567b4f-94b9-4662-825a-045e124c7948\") " pod="openstack/dnsmasq-dns-8c6f6df99-zx6zh" Dec 11 08:33:11 crc kubenswrapper[4860]: I1211 08:33:11.786930 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2e567b4f-94b9-4662-825a-045e124c7948-config\") pod \"dnsmasq-dns-8c6f6df99-zx6zh\" (UID: \"2e567b4f-94b9-4662-825a-045e124c7948\") " pod="openstack/dnsmasq-dns-8c6f6df99-zx6zh" Dec 11 08:33:11 crc kubenswrapper[4860]: I1211 08:33:11.786965 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zlctn\" (UniqueName: \"kubernetes.io/projected/2e567b4f-94b9-4662-825a-045e124c7948-kube-api-access-zlctn\") pod \"dnsmasq-dns-8c6f6df99-zx6zh\" (UID: \"2e567b4f-94b9-4662-825a-045e124c7948\") " pod="openstack/dnsmasq-dns-8c6f6df99-zx6zh" Dec 11 08:33:11 crc kubenswrapper[4860]: I1211 08:33:11.786985 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2e567b4f-94b9-4662-825a-045e124c7948-dns-swift-storage-0\") pod \"dnsmasq-dns-8c6f6df99-zx6zh\" (UID: \"2e567b4f-94b9-4662-825a-045e124c7948\") " pod="openstack/dnsmasq-dns-8c6f6df99-zx6zh" Dec 11 08:33:11 crc kubenswrapper[4860]: I1211 08:33:11.787007 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2e567b4f-94b9-4662-825a-045e124c7948-ovsdbserver-sb\") pod \"dnsmasq-dns-8c6f6df99-zx6zh\" (UID: \"2e567b4f-94b9-4662-825a-045e124c7948\") " pod="openstack/dnsmasq-dns-8c6f6df99-zx6zh" Dec 11 08:33:11 crc kubenswrapper[4860]: I1211 08:33:11.787028 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2e567b4f-94b9-4662-825a-045e124c7948-dns-svc\") pod \"dnsmasq-dns-8c6f6df99-zx6zh\" (UID: \"2e567b4f-94b9-4662-825a-045e124c7948\") " pod="openstack/dnsmasq-dns-8c6f6df99-zx6zh" Dec 11 08:33:11 crc kubenswrapper[4860]: I1211 08:33:11.787097 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2e567b4f-94b9-4662-825a-045e124c7948-ovsdbserver-nb\") pod \"dnsmasq-dns-8c6f6df99-zx6zh\" (UID: \"2e567b4f-94b9-4662-825a-045e124c7948\") " pod="openstack/dnsmasq-dns-8c6f6df99-zx6zh" Dec 11 08:33:11 crc kubenswrapper[4860]: I1211 08:33:11.788029 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2e567b4f-94b9-4662-825a-045e124c7948-config\") pod \"dnsmasq-dns-8c6f6df99-zx6zh\" (UID: \"2e567b4f-94b9-4662-825a-045e124c7948\") " pod="openstack/dnsmasq-dns-8c6f6df99-zx6zh" Dec 11 08:33:11 crc kubenswrapper[4860]: I1211 08:33:11.788874 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2e567b4f-94b9-4662-825a-045e124c7948-ovsdbserver-nb\") pod \"dnsmasq-dns-8c6f6df99-zx6zh\" (UID: \"2e567b4f-94b9-4662-825a-045e124c7948\") " pod="openstack/dnsmasq-dns-8c6f6df99-zx6zh" Dec 11 08:33:11 crc kubenswrapper[4860]: I1211 08:33:11.789063 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2e567b4f-94b9-4662-825a-045e124c7948-dns-svc\") pod \"dnsmasq-dns-8c6f6df99-zx6zh\" (UID: \"2e567b4f-94b9-4662-825a-045e124c7948\") " pod="openstack/dnsmasq-dns-8c6f6df99-zx6zh" Dec 11 08:33:11 crc kubenswrapper[4860]: I1211 08:33:11.789083 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2e567b4f-94b9-4662-825a-045e124c7948-dns-swift-storage-0\") pod \"dnsmasq-dns-8c6f6df99-zx6zh\" (UID: \"2e567b4f-94b9-4662-825a-045e124c7948\") " pod="openstack/dnsmasq-dns-8c6f6df99-zx6zh" Dec 11 08:33:11 crc kubenswrapper[4860]: I1211 08:33:11.789752 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/2e567b4f-94b9-4662-825a-045e124c7948-openstack-edpm-ipam\") pod \"dnsmasq-dns-8c6f6df99-zx6zh\" (UID: \"2e567b4f-94b9-4662-825a-045e124c7948\") " pod="openstack/dnsmasq-dns-8c6f6df99-zx6zh" Dec 11 08:33:11 crc kubenswrapper[4860]: I1211 08:33:11.789759 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2e567b4f-94b9-4662-825a-045e124c7948-ovsdbserver-sb\") pod \"dnsmasq-dns-8c6f6df99-zx6zh\" (UID: \"2e567b4f-94b9-4662-825a-045e124c7948\") " pod="openstack/dnsmasq-dns-8c6f6df99-zx6zh" Dec 11 08:33:11 crc kubenswrapper[4860]: I1211 08:33:11.840629 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zlctn\" (UniqueName: \"kubernetes.io/projected/2e567b4f-94b9-4662-825a-045e124c7948-kube-api-access-zlctn\") pod \"dnsmasq-dns-8c6f6df99-zx6zh\" (UID: \"2e567b4f-94b9-4662-825a-045e124c7948\") " pod="openstack/dnsmasq-dns-8c6f6df99-zx6zh" Dec 11 08:33:11 crc kubenswrapper[4860]: I1211 08:33:11.956501 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8c6f6df99-zx6zh" Dec 11 08:33:11 crc kubenswrapper[4860]: I1211 08:33:11.973709 4860 generic.go:334] "Generic (PLEG): container finished" podID="33f4ccd8-b2d7-43f7-a6f6-daf6ea2fa5b8" containerID="0f40eceff896f6453086ec0213e1ebfaa19e60a191b1173daa71b9121db12111" exitCode=0 Dec 11 08:33:11 crc kubenswrapper[4860]: I1211 08:33:11.973774 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c7b6c5df9-7zcv8" event={"ID":"33f4ccd8-b2d7-43f7-a6f6-daf6ea2fa5b8","Type":"ContainerDied","Data":"0f40eceff896f6453086ec0213e1ebfaa19e60a191b1173daa71b9121db12111"} Dec 11 08:33:11 crc kubenswrapper[4860]: I1211 08:33:11.973900 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c7b6c5df9-7zcv8" event={"ID":"33f4ccd8-b2d7-43f7-a6f6-daf6ea2fa5b8","Type":"ContainerDied","Data":"1c83d81db7c29d50bca861c14fde14a3cc6920ebec192ef3bb0fb2f2e272b71e"} Dec 11 08:33:11 crc kubenswrapper[4860]: I1211 08:33:11.973917 4860 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1c83d81db7c29d50bca861c14fde14a3cc6920ebec192ef3bb0fb2f2e272b71e" Dec 11 08:33:12 crc kubenswrapper[4860]: I1211 08:33:12.102769 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c7b6c5df9-7zcv8" Dec 11 08:33:12 crc kubenswrapper[4860]: I1211 08:33:12.201163 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/33f4ccd8-b2d7-43f7-a6f6-daf6ea2fa5b8-dns-svc\") pod \"33f4ccd8-b2d7-43f7-a6f6-daf6ea2fa5b8\" (UID: \"33f4ccd8-b2d7-43f7-a6f6-daf6ea2fa5b8\") " Dec 11 08:33:12 crc kubenswrapper[4860]: I1211 08:33:12.201201 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/33f4ccd8-b2d7-43f7-a6f6-daf6ea2fa5b8-ovsdbserver-nb\") pod \"33f4ccd8-b2d7-43f7-a6f6-daf6ea2fa5b8\" (UID: \"33f4ccd8-b2d7-43f7-a6f6-daf6ea2fa5b8\") " Dec 11 08:33:12 crc kubenswrapper[4860]: I1211 08:33:12.201361 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/33f4ccd8-b2d7-43f7-a6f6-daf6ea2fa5b8-ovsdbserver-sb\") pod \"33f4ccd8-b2d7-43f7-a6f6-daf6ea2fa5b8\" (UID: \"33f4ccd8-b2d7-43f7-a6f6-daf6ea2fa5b8\") " Dec 11 08:33:12 crc kubenswrapper[4860]: I1211 08:33:12.201523 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/33f4ccd8-b2d7-43f7-a6f6-daf6ea2fa5b8-dns-swift-storage-0\") pod \"33f4ccd8-b2d7-43f7-a6f6-daf6ea2fa5b8\" (UID: \"33f4ccd8-b2d7-43f7-a6f6-daf6ea2fa5b8\") " Dec 11 08:33:12 crc kubenswrapper[4860]: I1211 08:33:12.201596 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5n794\" (UniqueName: \"kubernetes.io/projected/33f4ccd8-b2d7-43f7-a6f6-daf6ea2fa5b8-kube-api-access-5n794\") pod \"33f4ccd8-b2d7-43f7-a6f6-daf6ea2fa5b8\" (UID: \"33f4ccd8-b2d7-43f7-a6f6-daf6ea2fa5b8\") " Dec 11 08:33:12 crc kubenswrapper[4860]: I1211 08:33:12.201634 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/33f4ccd8-b2d7-43f7-a6f6-daf6ea2fa5b8-config\") pod \"33f4ccd8-b2d7-43f7-a6f6-daf6ea2fa5b8\" (UID: \"33f4ccd8-b2d7-43f7-a6f6-daf6ea2fa5b8\") " Dec 11 08:33:12 crc kubenswrapper[4860]: I1211 08:33:12.218984 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/33f4ccd8-b2d7-43f7-a6f6-daf6ea2fa5b8-kube-api-access-5n794" (OuterVolumeSpecName: "kube-api-access-5n794") pod "33f4ccd8-b2d7-43f7-a6f6-daf6ea2fa5b8" (UID: "33f4ccd8-b2d7-43f7-a6f6-daf6ea2fa5b8"). InnerVolumeSpecName "kube-api-access-5n794". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:33:12 crc kubenswrapper[4860]: I1211 08:33:12.261391 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/33f4ccd8-b2d7-43f7-a6f6-daf6ea2fa5b8-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "33f4ccd8-b2d7-43f7-a6f6-daf6ea2fa5b8" (UID: "33f4ccd8-b2d7-43f7-a6f6-daf6ea2fa5b8"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:33:12 crc kubenswrapper[4860]: I1211 08:33:12.269352 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/33f4ccd8-b2d7-43f7-a6f6-daf6ea2fa5b8-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "33f4ccd8-b2d7-43f7-a6f6-daf6ea2fa5b8" (UID: "33f4ccd8-b2d7-43f7-a6f6-daf6ea2fa5b8"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:33:12 crc kubenswrapper[4860]: I1211 08:33:12.272206 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/33f4ccd8-b2d7-43f7-a6f6-daf6ea2fa5b8-config" (OuterVolumeSpecName: "config") pod "33f4ccd8-b2d7-43f7-a6f6-daf6ea2fa5b8" (UID: "33f4ccd8-b2d7-43f7-a6f6-daf6ea2fa5b8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:33:12 crc kubenswrapper[4860]: I1211 08:33:12.285517 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/33f4ccd8-b2d7-43f7-a6f6-daf6ea2fa5b8-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "33f4ccd8-b2d7-43f7-a6f6-daf6ea2fa5b8" (UID: "33f4ccd8-b2d7-43f7-a6f6-daf6ea2fa5b8"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:33:12 crc kubenswrapper[4860]: I1211 08:33:12.287828 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/33f4ccd8-b2d7-43f7-a6f6-daf6ea2fa5b8-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "33f4ccd8-b2d7-43f7-a6f6-daf6ea2fa5b8" (UID: "33f4ccd8-b2d7-43f7-a6f6-daf6ea2fa5b8"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:33:12 crc kubenswrapper[4860]: I1211 08:33:12.306142 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5n794\" (UniqueName: \"kubernetes.io/projected/33f4ccd8-b2d7-43f7-a6f6-daf6ea2fa5b8-kube-api-access-5n794\") on node \"crc\" DevicePath \"\"" Dec 11 08:33:12 crc kubenswrapper[4860]: I1211 08:33:12.306186 4860 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/33f4ccd8-b2d7-43f7-a6f6-daf6ea2fa5b8-config\") on node \"crc\" DevicePath \"\"" Dec 11 08:33:12 crc kubenswrapper[4860]: I1211 08:33:12.306196 4860 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/33f4ccd8-b2d7-43f7-a6f6-daf6ea2fa5b8-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 11 08:33:12 crc kubenswrapper[4860]: I1211 08:33:12.306206 4860 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/33f4ccd8-b2d7-43f7-a6f6-daf6ea2fa5b8-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 11 08:33:12 crc kubenswrapper[4860]: I1211 08:33:12.306215 4860 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/33f4ccd8-b2d7-43f7-a6f6-daf6ea2fa5b8-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 11 08:33:12 crc kubenswrapper[4860]: I1211 08:33:12.306224 4860 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/33f4ccd8-b2d7-43f7-a6f6-daf6ea2fa5b8-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 11 08:33:12 crc kubenswrapper[4860]: I1211 08:33:12.522831 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8c6f6df99-zx6zh"] Dec 11 08:33:12 crc kubenswrapper[4860]: I1211 08:33:12.988162 4860 generic.go:334] "Generic (PLEG): container finished" podID="2e567b4f-94b9-4662-825a-045e124c7948" containerID="d9245de76861f146c3ed6a3087d77ab66d551b2d8cbdd332a6bf96e569aaf9af" exitCode=0 Dec 11 08:33:12 crc kubenswrapper[4860]: I1211 08:33:12.988240 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8c6f6df99-zx6zh" event={"ID":"2e567b4f-94b9-4662-825a-045e124c7948","Type":"ContainerDied","Data":"d9245de76861f146c3ed6a3087d77ab66d551b2d8cbdd332a6bf96e569aaf9af"} Dec 11 08:33:12 crc kubenswrapper[4860]: I1211 08:33:12.988431 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c7b6c5df9-7zcv8" Dec 11 08:33:12 crc kubenswrapper[4860]: I1211 08:33:12.988465 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8c6f6df99-zx6zh" event={"ID":"2e567b4f-94b9-4662-825a-045e124c7948","Type":"ContainerStarted","Data":"ef3c4eed02dd8867561f1ce17e97290917a68ffb24defb4e618a7fe514dde4c6"} Dec 11 08:33:13 crc kubenswrapper[4860]: I1211 08:33:13.150226 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c7b6c5df9-7zcv8"] Dec 11 08:33:13 crc kubenswrapper[4860]: I1211 08:33:13.159597 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c7b6c5df9-7zcv8"] Dec 11 08:33:13 crc kubenswrapper[4860]: I1211 08:33:13.591304 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="33f4ccd8-b2d7-43f7-a6f6-daf6ea2fa5b8" path="/var/lib/kubelet/pods/33f4ccd8-b2d7-43f7-a6f6-daf6ea2fa5b8/volumes" Dec 11 08:33:14 crc kubenswrapper[4860]: I1211 08:33:14.000600 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8c6f6df99-zx6zh" event={"ID":"2e567b4f-94b9-4662-825a-045e124c7948","Type":"ContainerStarted","Data":"0983c135c737a62065175618d3ef86a6aa26707ac1f29baf5d693fce6adb495d"} Dec 11 08:33:14 crc kubenswrapper[4860]: I1211 08:33:14.000814 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-8c6f6df99-zx6zh" Dec 11 08:33:14 crc kubenswrapper[4860]: I1211 08:33:14.022453 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-8c6f6df99-zx6zh" podStartSLOduration=3.022429139 podStartE2EDuration="3.022429139s" podCreationTimestamp="2025-12-11 08:33:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:33:14.021855941 +0000 UTC m=+1326.750374996" watchObservedRunningTime="2025-12-11 08:33:14.022429139 +0000 UTC m=+1326.750948194" Dec 11 08:33:21 crc kubenswrapper[4860]: I1211 08:33:21.958924 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-8c6f6df99-zx6zh" Dec 11 08:33:22 crc kubenswrapper[4860]: I1211 08:33:22.020801 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5576978c7c-2m9pb"] Dec 11 08:33:22 crc kubenswrapper[4860]: I1211 08:33:22.021227 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5576978c7c-2m9pb" podUID="ecccf7e4-15f8-4a10-9662-c188d5ecc068" containerName="dnsmasq-dns" containerID="cri-o://02ef99652be9a72d174fd83a2d766b321cdae7dd835f28989bbf269ec696161b" gracePeriod=10 Dec 11 08:33:22 crc kubenswrapper[4860]: I1211 08:33:22.532655 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5576978c7c-2m9pb" Dec 11 08:33:22 crc kubenswrapper[4860]: I1211 08:33:22.577235 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ecccf7e4-15f8-4a10-9662-c188d5ecc068-config\") pod \"ecccf7e4-15f8-4a10-9662-c188d5ecc068\" (UID: \"ecccf7e4-15f8-4a10-9662-c188d5ecc068\") " Dec 11 08:33:22 crc kubenswrapper[4860]: I1211 08:33:22.577652 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ecccf7e4-15f8-4a10-9662-c188d5ecc068-ovsdbserver-nb\") pod \"ecccf7e4-15f8-4a10-9662-c188d5ecc068\" (UID: \"ecccf7e4-15f8-4a10-9662-c188d5ecc068\") " Dec 11 08:33:22 crc kubenswrapper[4860]: I1211 08:33:22.577779 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ecccf7e4-15f8-4a10-9662-c188d5ecc068-dns-svc\") pod \"ecccf7e4-15f8-4a10-9662-c188d5ecc068\" (UID: \"ecccf7e4-15f8-4a10-9662-c188d5ecc068\") " Dec 11 08:33:22 crc kubenswrapper[4860]: I1211 08:33:22.577920 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vbclz\" (UniqueName: \"kubernetes.io/projected/ecccf7e4-15f8-4a10-9662-c188d5ecc068-kube-api-access-vbclz\") pod \"ecccf7e4-15f8-4a10-9662-c188d5ecc068\" (UID: \"ecccf7e4-15f8-4a10-9662-c188d5ecc068\") " Dec 11 08:33:22 crc kubenswrapper[4860]: I1211 08:33:22.578425 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ecccf7e4-15f8-4a10-9662-c188d5ecc068-dns-swift-storage-0\") pod \"ecccf7e4-15f8-4a10-9662-c188d5ecc068\" (UID: \"ecccf7e4-15f8-4a10-9662-c188d5ecc068\") " Dec 11 08:33:22 crc kubenswrapper[4860]: I1211 08:33:22.578844 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/ecccf7e4-15f8-4a10-9662-c188d5ecc068-openstack-edpm-ipam\") pod \"ecccf7e4-15f8-4a10-9662-c188d5ecc068\" (UID: \"ecccf7e4-15f8-4a10-9662-c188d5ecc068\") " Dec 11 08:33:22 crc kubenswrapper[4860]: I1211 08:33:22.578947 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ecccf7e4-15f8-4a10-9662-c188d5ecc068-ovsdbserver-sb\") pod \"ecccf7e4-15f8-4a10-9662-c188d5ecc068\" (UID: \"ecccf7e4-15f8-4a10-9662-c188d5ecc068\") " Dec 11 08:33:22 crc kubenswrapper[4860]: I1211 08:33:22.597827 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ecccf7e4-15f8-4a10-9662-c188d5ecc068-kube-api-access-vbclz" (OuterVolumeSpecName: "kube-api-access-vbclz") pod "ecccf7e4-15f8-4a10-9662-c188d5ecc068" (UID: "ecccf7e4-15f8-4a10-9662-c188d5ecc068"). InnerVolumeSpecName "kube-api-access-vbclz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:33:22 crc kubenswrapper[4860]: I1211 08:33:22.637669 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ecccf7e4-15f8-4a10-9662-c188d5ecc068-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ecccf7e4-15f8-4a10-9662-c188d5ecc068" (UID: "ecccf7e4-15f8-4a10-9662-c188d5ecc068"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:33:22 crc kubenswrapper[4860]: I1211 08:33:22.641907 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ecccf7e4-15f8-4a10-9662-c188d5ecc068-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ecccf7e4-15f8-4a10-9662-c188d5ecc068" (UID: "ecccf7e4-15f8-4a10-9662-c188d5ecc068"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:33:22 crc kubenswrapper[4860]: I1211 08:33:22.644160 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ecccf7e4-15f8-4a10-9662-c188d5ecc068-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ecccf7e4-15f8-4a10-9662-c188d5ecc068" (UID: "ecccf7e4-15f8-4a10-9662-c188d5ecc068"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:33:22 crc kubenswrapper[4860]: I1211 08:33:22.647391 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ecccf7e4-15f8-4a10-9662-c188d5ecc068-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "ecccf7e4-15f8-4a10-9662-c188d5ecc068" (UID: "ecccf7e4-15f8-4a10-9662-c188d5ecc068"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:33:22 crc kubenswrapper[4860]: I1211 08:33:22.648411 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ecccf7e4-15f8-4a10-9662-c188d5ecc068-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "ecccf7e4-15f8-4a10-9662-c188d5ecc068" (UID: "ecccf7e4-15f8-4a10-9662-c188d5ecc068"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:33:22 crc kubenswrapper[4860]: I1211 08:33:22.655581 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ecccf7e4-15f8-4a10-9662-c188d5ecc068-config" (OuterVolumeSpecName: "config") pod "ecccf7e4-15f8-4a10-9662-c188d5ecc068" (UID: "ecccf7e4-15f8-4a10-9662-c188d5ecc068"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:33:22 crc kubenswrapper[4860]: I1211 08:33:22.680534 4860 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ecccf7e4-15f8-4a10-9662-c188d5ecc068-config\") on node \"crc\" DevicePath \"\"" Dec 11 08:33:22 crc kubenswrapper[4860]: I1211 08:33:22.680569 4860 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ecccf7e4-15f8-4a10-9662-c188d5ecc068-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 11 08:33:22 crc kubenswrapper[4860]: I1211 08:33:22.680627 4860 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ecccf7e4-15f8-4a10-9662-c188d5ecc068-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 11 08:33:22 crc kubenswrapper[4860]: I1211 08:33:22.680638 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vbclz\" (UniqueName: \"kubernetes.io/projected/ecccf7e4-15f8-4a10-9662-c188d5ecc068-kube-api-access-vbclz\") on node \"crc\" DevicePath \"\"" Dec 11 08:33:22 crc kubenswrapper[4860]: I1211 08:33:22.680730 4860 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ecccf7e4-15f8-4a10-9662-c188d5ecc068-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 11 08:33:22 crc kubenswrapper[4860]: I1211 08:33:22.680740 4860 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/ecccf7e4-15f8-4a10-9662-c188d5ecc068-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 11 08:33:22 crc kubenswrapper[4860]: I1211 08:33:22.680748 4860 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ecccf7e4-15f8-4a10-9662-c188d5ecc068-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 11 08:33:23 crc kubenswrapper[4860]: I1211 08:33:23.105033 4860 generic.go:334] "Generic (PLEG): container finished" podID="ecccf7e4-15f8-4a10-9662-c188d5ecc068" containerID="02ef99652be9a72d174fd83a2d766b321cdae7dd835f28989bbf269ec696161b" exitCode=0 Dec 11 08:33:23 crc kubenswrapper[4860]: I1211 08:33:23.105091 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5576978c7c-2m9pb" event={"ID":"ecccf7e4-15f8-4a10-9662-c188d5ecc068","Type":"ContainerDied","Data":"02ef99652be9a72d174fd83a2d766b321cdae7dd835f28989bbf269ec696161b"} Dec 11 08:33:23 crc kubenswrapper[4860]: I1211 08:33:23.105112 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5576978c7c-2m9pb" Dec 11 08:33:23 crc kubenswrapper[4860]: I1211 08:33:23.105137 4860 scope.go:117] "RemoveContainer" containerID="02ef99652be9a72d174fd83a2d766b321cdae7dd835f28989bbf269ec696161b" Dec 11 08:33:23 crc kubenswrapper[4860]: I1211 08:33:23.105124 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5576978c7c-2m9pb" event={"ID":"ecccf7e4-15f8-4a10-9662-c188d5ecc068","Type":"ContainerDied","Data":"7bed0057c9d4a3cc2cd037fe0c07dd5ecbf8fc53b13a00e03f87c9ca90a0f1bc"} Dec 11 08:33:23 crc kubenswrapper[4860]: I1211 08:33:23.134867 4860 scope.go:117] "RemoveContainer" containerID="663607f4f83dc73084a650362934ec4f5ca40a77c0d419cedf68471a77a18b12" Dec 11 08:33:23 crc kubenswrapper[4860]: I1211 08:33:23.140149 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5576978c7c-2m9pb"] Dec 11 08:33:23 crc kubenswrapper[4860]: I1211 08:33:23.150820 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5576978c7c-2m9pb"] Dec 11 08:33:23 crc kubenswrapper[4860]: I1211 08:33:23.174219 4860 scope.go:117] "RemoveContainer" containerID="02ef99652be9a72d174fd83a2d766b321cdae7dd835f28989bbf269ec696161b" Dec 11 08:33:23 crc kubenswrapper[4860]: E1211 08:33:23.175137 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"02ef99652be9a72d174fd83a2d766b321cdae7dd835f28989bbf269ec696161b\": container with ID starting with 02ef99652be9a72d174fd83a2d766b321cdae7dd835f28989bbf269ec696161b not found: ID does not exist" containerID="02ef99652be9a72d174fd83a2d766b321cdae7dd835f28989bbf269ec696161b" Dec 11 08:33:23 crc kubenswrapper[4860]: I1211 08:33:23.175182 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"02ef99652be9a72d174fd83a2d766b321cdae7dd835f28989bbf269ec696161b"} err="failed to get container status \"02ef99652be9a72d174fd83a2d766b321cdae7dd835f28989bbf269ec696161b\": rpc error: code = NotFound desc = could not find container \"02ef99652be9a72d174fd83a2d766b321cdae7dd835f28989bbf269ec696161b\": container with ID starting with 02ef99652be9a72d174fd83a2d766b321cdae7dd835f28989bbf269ec696161b not found: ID does not exist" Dec 11 08:33:23 crc kubenswrapper[4860]: I1211 08:33:23.175217 4860 scope.go:117] "RemoveContainer" containerID="663607f4f83dc73084a650362934ec4f5ca40a77c0d419cedf68471a77a18b12" Dec 11 08:33:23 crc kubenswrapper[4860]: E1211 08:33:23.175764 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"663607f4f83dc73084a650362934ec4f5ca40a77c0d419cedf68471a77a18b12\": container with ID starting with 663607f4f83dc73084a650362934ec4f5ca40a77c0d419cedf68471a77a18b12 not found: ID does not exist" containerID="663607f4f83dc73084a650362934ec4f5ca40a77c0d419cedf68471a77a18b12" Dec 11 08:33:23 crc kubenswrapper[4860]: I1211 08:33:23.175818 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"663607f4f83dc73084a650362934ec4f5ca40a77c0d419cedf68471a77a18b12"} err="failed to get container status \"663607f4f83dc73084a650362934ec4f5ca40a77c0d419cedf68471a77a18b12\": rpc error: code = NotFound desc = could not find container \"663607f4f83dc73084a650362934ec4f5ca40a77c0d419cedf68471a77a18b12\": container with ID starting with 663607f4f83dc73084a650362934ec4f5ca40a77c0d419cedf68471a77a18b12 not found: ID does not exist" Dec 11 08:33:23 crc kubenswrapper[4860]: I1211 08:33:23.595231 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ecccf7e4-15f8-4a10-9662-c188d5ecc068" path="/var/lib/kubelet/pods/ecccf7e4-15f8-4a10-9662-c188d5ecc068/volumes" Dec 11 08:33:34 crc kubenswrapper[4860]: I1211 08:33:34.223288 4860 generic.go:334] "Generic (PLEG): container finished" podID="a1d839a1-4174-4abc-9e3d-5579603687e6" containerID="1983955f161e6974353df114f1902b23c7065b295b78a3af0755e8cbbdcbb3a6" exitCode=0 Dec 11 08:33:34 crc kubenswrapper[4860]: I1211 08:33:34.223349 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"a1d839a1-4174-4abc-9e3d-5579603687e6","Type":"ContainerDied","Data":"1983955f161e6974353df114f1902b23c7065b295b78a3af0755e8cbbdcbb3a6"} Dec 11 08:33:34 crc kubenswrapper[4860]: I1211 08:33:34.518417 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2vqvr"] Dec 11 08:33:34 crc kubenswrapper[4860]: E1211 08:33:34.519232 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33f4ccd8-b2d7-43f7-a6f6-daf6ea2fa5b8" containerName="init" Dec 11 08:33:34 crc kubenswrapper[4860]: I1211 08:33:34.519253 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="33f4ccd8-b2d7-43f7-a6f6-daf6ea2fa5b8" containerName="init" Dec 11 08:33:34 crc kubenswrapper[4860]: E1211 08:33:34.519273 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ecccf7e4-15f8-4a10-9662-c188d5ecc068" containerName="dnsmasq-dns" Dec 11 08:33:34 crc kubenswrapper[4860]: I1211 08:33:34.519280 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="ecccf7e4-15f8-4a10-9662-c188d5ecc068" containerName="dnsmasq-dns" Dec 11 08:33:34 crc kubenswrapper[4860]: E1211 08:33:34.519295 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ecccf7e4-15f8-4a10-9662-c188d5ecc068" containerName="init" Dec 11 08:33:34 crc kubenswrapper[4860]: I1211 08:33:34.519301 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="ecccf7e4-15f8-4a10-9662-c188d5ecc068" containerName="init" Dec 11 08:33:34 crc kubenswrapper[4860]: E1211 08:33:34.519319 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33f4ccd8-b2d7-43f7-a6f6-daf6ea2fa5b8" containerName="dnsmasq-dns" Dec 11 08:33:34 crc kubenswrapper[4860]: I1211 08:33:34.519325 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="33f4ccd8-b2d7-43f7-a6f6-daf6ea2fa5b8" containerName="dnsmasq-dns" Dec 11 08:33:34 crc kubenswrapper[4860]: I1211 08:33:34.519498 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="33f4ccd8-b2d7-43f7-a6f6-daf6ea2fa5b8" containerName="dnsmasq-dns" Dec 11 08:33:34 crc kubenswrapper[4860]: I1211 08:33:34.519526 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="ecccf7e4-15f8-4a10-9662-c188d5ecc068" containerName="dnsmasq-dns" Dec 11 08:33:34 crc kubenswrapper[4860]: I1211 08:33:34.520303 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2vqvr" Dec 11 08:33:34 crc kubenswrapper[4860]: I1211 08:33:34.522943 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 11 08:33:34 crc kubenswrapper[4860]: I1211 08:33:34.523382 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 11 08:33:34 crc kubenswrapper[4860]: I1211 08:33:34.523718 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-7w7kh" Dec 11 08:33:34 crc kubenswrapper[4860]: I1211 08:33:34.526392 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 11 08:33:34 crc kubenswrapper[4860]: I1211 08:33:34.530739 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2vqvr"] Dec 11 08:33:34 crc kubenswrapper[4860]: I1211 08:33:34.651973 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b0ccaf6b-d322-440b-87c4-560b61547f52-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-2vqvr\" (UID: \"b0ccaf6b-d322-440b-87c4-560b61547f52\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2vqvr" Dec 11 08:33:34 crc kubenswrapper[4860]: I1211 08:33:34.652160 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b0ccaf6b-d322-440b-87c4-560b61547f52-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-2vqvr\" (UID: \"b0ccaf6b-d322-440b-87c4-560b61547f52\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2vqvr" Dec 11 08:33:34 crc kubenswrapper[4860]: I1211 08:33:34.652236 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0ccaf6b-d322-440b-87c4-560b61547f52-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-2vqvr\" (UID: \"b0ccaf6b-d322-440b-87c4-560b61547f52\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2vqvr" Dec 11 08:33:34 crc kubenswrapper[4860]: I1211 08:33:34.652339 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j5mvs\" (UniqueName: \"kubernetes.io/projected/b0ccaf6b-d322-440b-87c4-560b61547f52-kube-api-access-j5mvs\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-2vqvr\" (UID: \"b0ccaf6b-d322-440b-87c4-560b61547f52\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2vqvr" Dec 11 08:33:34 crc kubenswrapper[4860]: I1211 08:33:34.754307 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b0ccaf6b-d322-440b-87c4-560b61547f52-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-2vqvr\" (UID: \"b0ccaf6b-d322-440b-87c4-560b61547f52\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2vqvr" Dec 11 08:33:34 crc kubenswrapper[4860]: I1211 08:33:34.754357 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0ccaf6b-d322-440b-87c4-560b61547f52-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-2vqvr\" (UID: \"b0ccaf6b-d322-440b-87c4-560b61547f52\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2vqvr" Dec 11 08:33:34 crc kubenswrapper[4860]: I1211 08:33:34.754399 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j5mvs\" (UniqueName: \"kubernetes.io/projected/b0ccaf6b-d322-440b-87c4-560b61547f52-kube-api-access-j5mvs\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-2vqvr\" (UID: \"b0ccaf6b-d322-440b-87c4-560b61547f52\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2vqvr" Dec 11 08:33:34 crc kubenswrapper[4860]: I1211 08:33:34.754502 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b0ccaf6b-d322-440b-87c4-560b61547f52-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-2vqvr\" (UID: \"b0ccaf6b-d322-440b-87c4-560b61547f52\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2vqvr" Dec 11 08:33:34 crc kubenswrapper[4860]: I1211 08:33:34.759370 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b0ccaf6b-d322-440b-87c4-560b61547f52-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-2vqvr\" (UID: \"b0ccaf6b-d322-440b-87c4-560b61547f52\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2vqvr" Dec 11 08:33:34 crc kubenswrapper[4860]: I1211 08:33:34.759711 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0ccaf6b-d322-440b-87c4-560b61547f52-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-2vqvr\" (UID: \"b0ccaf6b-d322-440b-87c4-560b61547f52\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2vqvr" Dec 11 08:33:34 crc kubenswrapper[4860]: I1211 08:33:34.759852 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b0ccaf6b-d322-440b-87c4-560b61547f52-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-2vqvr\" (UID: \"b0ccaf6b-d322-440b-87c4-560b61547f52\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2vqvr" Dec 11 08:33:34 crc kubenswrapper[4860]: I1211 08:33:34.776635 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j5mvs\" (UniqueName: \"kubernetes.io/projected/b0ccaf6b-d322-440b-87c4-560b61547f52-kube-api-access-j5mvs\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-2vqvr\" (UID: \"b0ccaf6b-d322-440b-87c4-560b61547f52\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2vqvr" Dec 11 08:33:34 crc kubenswrapper[4860]: I1211 08:33:34.902718 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2vqvr" Dec 11 08:33:35 crc kubenswrapper[4860]: I1211 08:33:35.244294 4860 generic.go:334] "Generic (PLEG): container finished" podID="9cfe9bd2-a1fe-4cae-8f48-c007bfd95961" containerID="fdfdcc293ae6a46b533c028cac1585240f64ab520d8db90fadd9acdae90d9bf8" exitCode=0 Dec 11 08:33:35 crc kubenswrapper[4860]: I1211 08:33:35.244396 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"9cfe9bd2-a1fe-4cae-8f48-c007bfd95961","Type":"ContainerDied","Data":"fdfdcc293ae6a46b533c028cac1585240f64ab520d8db90fadd9acdae90d9bf8"} Dec 11 08:33:35 crc kubenswrapper[4860]: I1211 08:33:35.252905 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"a1d839a1-4174-4abc-9e3d-5579603687e6","Type":"ContainerStarted","Data":"ea7cacc819c20ffe15aec61c15dca5fb9f6ae613268fefb073cb5ed16bf8179b"} Dec 11 08:33:35 crc kubenswrapper[4860]: I1211 08:33:35.253150 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 11 08:33:35 crc kubenswrapper[4860]: I1211 08:33:35.313930 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=37.313913164 podStartE2EDuration="37.313913164s" podCreationTimestamp="2025-12-11 08:32:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:33:35.3033469 +0000 UTC m=+1348.031865945" watchObservedRunningTime="2025-12-11 08:33:35.313913164 +0000 UTC m=+1348.042432209" Dec 11 08:33:35 crc kubenswrapper[4860]: I1211 08:33:35.506902 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2vqvr"] Dec 11 08:33:35 crc kubenswrapper[4860]: W1211 08:33:35.525286 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb0ccaf6b_d322_440b_87c4_560b61547f52.slice/crio-9f6431fd80bdf1be19a1151672d42e0faadbfc3759c65b68c4c67ffe4d9e7123 WatchSource:0}: Error finding container 9f6431fd80bdf1be19a1151672d42e0faadbfc3759c65b68c4c67ffe4d9e7123: Status 404 returned error can't find the container with id 9f6431fd80bdf1be19a1151672d42e0faadbfc3759c65b68c4c67ffe4d9e7123 Dec 11 08:33:35 crc kubenswrapper[4860]: I1211 08:33:35.529299 4860 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 11 08:33:36 crc kubenswrapper[4860]: I1211 08:33:36.265724 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2vqvr" event={"ID":"b0ccaf6b-d322-440b-87c4-560b61547f52","Type":"ContainerStarted","Data":"9f6431fd80bdf1be19a1151672d42e0faadbfc3759c65b68c4c67ffe4d9e7123"} Dec 11 08:33:36 crc kubenswrapper[4860]: I1211 08:33:36.268221 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"9cfe9bd2-a1fe-4cae-8f48-c007bfd95961","Type":"ContainerStarted","Data":"6bd7205793b50a234344524198019dbe043a41efd410d53117f8cfdd718302de"} Dec 11 08:33:36 crc kubenswrapper[4860]: I1211 08:33:36.268835 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:33:36 crc kubenswrapper[4860]: I1211 08:33:36.298240 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=37.298215591 podStartE2EDuration="37.298215591s" podCreationTimestamp="2025-12-11 08:32:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 08:33:36.291772547 +0000 UTC m=+1349.020291592" watchObservedRunningTime="2025-12-11 08:33:36.298215591 +0000 UTC m=+1349.026734646" Dec 11 08:33:46 crc kubenswrapper[4860]: I1211 08:33:46.382545 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2vqvr" event={"ID":"b0ccaf6b-d322-440b-87c4-560b61547f52","Type":"ContainerStarted","Data":"405e6a8833532bfbd629161cdc29d6d0acd23394be8c48109e637bb748a496a1"} Dec 11 08:33:46 crc kubenswrapper[4860]: I1211 08:33:46.410014 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2vqvr" podStartSLOduration=2.049244671 podStartE2EDuration="12.409985302s" podCreationTimestamp="2025-12-11 08:33:34 +0000 UTC" firstStartedPulling="2025-12-11 08:33:35.528919151 +0000 UTC m=+1348.257438216" lastFinishedPulling="2025-12-11 08:33:45.889659792 +0000 UTC m=+1358.618178847" observedRunningTime="2025-12-11 08:33:46.407777892 +0000 UTC m=+1359.136296957" watchObservedRunningTime="2025-12-11 08:33:46.409985302 +0000 UTC m=+1359.138504357" Dec 11 08:33:49 crc kubenswrapper[4860]: I1211 08:33:49.236859 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 11 08:33:50 crc kubenswrapper[4860]: I1211 08:33:50.242905 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 11 08:33:58 crc kubenswrapper[4860]: I1211 08:33:58.560487 4860 generic.go:334] "Generic (PLEG): container finished" podID="b0ccaf6b-d322-440b-87c4-560b61547f52" containerID="405e6a8833532bfbd629161cdc29d6d0acd23394be8c48109e637bb748a496a1" exitCode=0 Dec 11 08:33:58 crc kubenswrapper[4860]: I1211 08:33:58.560668 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2vqvr" event={"ID":"b0ccaf6b-d322-440b-87c4-560b61547f52","Type":"ContainerDied","Data":"405e6a8833532bfbd629161cdc29d6d0acd23394be8c48109e637bb748a496a1"} Dec 11 08:34:00 crc kubenswrapper[4860]: I1211 08:34:00.017871 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2vqvr" Dec 11 08:34:00 crc kubenswrapper[4860]: I1211 08:34:00.118841 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j5mvs\" (UniqueName: \"kubernetes.io/projected/b0ccaf6b-d322-440b-87c4-560b61547f52-kube-api-access-j5mvs\") pod \"b0ccaf6b-d322-440b-87c4-560b61547f52\" (UID: \"b0ccaf6b-d322-440b-87c4-560b61547f52\") " Dec 11 08:34:00 crc kubenswrapper[4860]: I1211 08:34:00.119077 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0ccaf6b-d322-440b-87c4-560b61547f52-repo-setup-combined-ca-bundle\") pod \"b0ccaf6b-d322-440b-87c4-560b61547f52\" (UID: \"b0ccaf6b-d322-440b-87c4-560b61547f52\") " Dec 11 08:34:00 crc kubenswrapper[4860]: I1211 08:34:00.119812 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b0ccaf6b-d322-440b-87c4-560b61547f52-ssh-key\") pod \"b0ccaf6b-d322-440b-87c4-560b61547f52\" (UID: \"b0ccaf6b-d322-440b-87c4-560b61547f52\") " Dec 11 08:34:00 crc kubenswrapper[4860]: I1211 08:34:00.119942 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b0ccaf6b-d322-440b-87c4-560b61547f52-inventory\") pod \"b0ccaf6b-d322-440b-87c4-560b61547f52\" (UID: \"b0ccaf6b-d322-440b-87c4-560b61547f52\") " Dec 11 08:34:00 crc kubenswrapper[4860]: I1211 08:34:00.125163 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b0ccaf6b-d322-440b-87c4-560b61547f52-kube-api-access-j5mvs" (OuterVolumeSpecName: "kube-api-access-j5mvs") pod "b0ccaf6b-d322-440b-87c4-560b61547f52" (UID: "b0ccaf6b-d322-440b-87c4-560b61547f52"). InnerVolumeSpecName "kube-api-access-j5mvs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:34:00 crc kubenswrapper[4860]: I1211 08:34:00.125461 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b0ccaf6b-d322-440b-87c4-560b61547f52-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "b0ccaf6b-d322-440b-87c4-560b61547f52" (UID: "b0ccaf6b-d322-440b-87c4-560b61547f52"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:34:00 crc kubenswrapper[4860]: I1211 08:34:00.162011 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b0ccaf6b-d322-440b-87c4-560b61547f52-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "b0ccaf6b-d322-440b-87c4-560b61547f52" (UID: "b0ccaf6b-d322-440b-87c4-560b61547f52"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:34:00 crc kubenswrapper[4860]: I1211 08:34:00.173049 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b0ccaf6b-d322-440b-87c4-560b61547f52-inventory" (OuterVolumeSpecName: "inventory") pod "b0ccaf6b-d322-440b-87c4-560b61547f52" (UID: "b0ccaf6b-d322-440b-87c4-560b61547f52"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:34:00 crc kubenswrapper[4860]: I1211 08:34:00.222733 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j5mvs\" (UniqueName: \"kubernetes.io/projected/b0ccaf6b-d322-440b-87c4-560b61547f52-kube-api-access-j5mvs\") on node \"crc\" DevicePath \"\"" Dec 11 08:34:00 crc kubenswrapper[4860]: I1211 08:34:00.222786 4860 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0ccaf6b-d322-440b-87c4-560b61547f52-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 08:34:00 crc kubenswrapper[4860]: I1211 08:34:00.222807 4860 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b0ccaf6b-d322-440b-87c4-560b61547f52-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 11 08:34:00 crc kubenswrapper[4860]: I1211 08:34:00.222829 4860 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b0ccaf6b-d322-440b-87c4-560b61547f52-inventory\") on node \"crc\" DevicePath \"\"" Dec 11 08:34:00 crc kubenswrapper[4860]: I1211 08:34:00.587093 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2vqvr" event={"ID":"b0ccaf6b-d322-440b-87c4-560b61547f52","Type":"ContainerDied","Data":"9f6431fd80bdf1be19a1151672d42e0faadbfc3759c65b68c4c67ffe4d9e7123"} Dec 11 08:34:00 crc kubenswrapper[4860]: I1211 08:34:00.587153 4860 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9f6431fd80bdf1be19a1151672d42e0faadbfc3759c65b68c4c67ffe4d9e7123" Dec 11 08:34:00 crc kubenswrapper[4860]: I1211 08:34:00.587166 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2vqvr" Dec 11 08:34:00 crc kubenswrapper[4860]: I1211 08:34:00.701831 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-zb8rn"] Dec 11 08:34:00 crc kubenswrapper[4860]: E1211 08:34:00.702485 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0ccaf6b-d322-440b-87c4-560b61547f52" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 11 08:34:00 crc kubenswrapper[4860]: I1211 08:34:00.702513 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0ccaf6b-d322-440b-87c4-560b61547f52" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 11 08:34:00 crc kubenswrapper[4860]: I1211 08:34:00.702843 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="b0ccaf6b-d322-440b-87c4-560b61547f52" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 11 08:34:00 crc kubenswrapper[4860]: I1211 08:34:00.703871 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zb8rn" Dec 11 08:34:00 crc kubenswrapper[4860]: I1211 08:34:00.707999 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 11 08:34:00 crc kubenswrapper[4860]: I1211 08:34:00.708009 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 11 08:34:00 crc kubenswrapper[4860]: I1211 08:34:00.708331 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-7w7kh" Dec 11 08:34:00 crc kubenswrapper[4860]: I1211 08:34:00.712742 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 11 08:34:00 crc kubenswrapper[4860]: I1211 08:34:00.715414 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-zb8rn"] Dec 11 08:34:00 crc kubenswrapper[4860]: I1211 08:34:00.735160 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a84f5e60-25bb-4412-a38c-fa4e7a32c177-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-zb8rn\" (UID: \"a84f5e60-25bb-4412-a38c-fa4e7a32c177\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zb8rn" Dec 11 08:34:00 crc kubenswrapper[4860]: I1211 08:34:00.735702 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fqkrm\" (UniqueName: \"kubernetes.io/projected/a84f5e60-25bb-4412-a38c-fa4e7a32c177-kube-api-access-fqkrm\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-zb8rn\" (UID: \"a84f5e60-25bb-4412-a38c-fa4e7a32c177\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zb8rn" Dec 11 08:34:00 crc kubenswrapper[4860]: I1211 08:34:00.735839 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a84f5e60-25bb-4412-a38c-fa4e7a32c177-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-zb8rn\" (UID: \"a84f5e60-25bb-4412-a38c-fa4e7a32c177\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zb8rn" Dec 11 08:34:00 crc kubenswrapper[4860]: I1211 08:34:00.836589 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fqkrm\" (UniqueName: \"kubernetes.io/projected/a84f5e60-25bb-4412-a38c-fa4e7a32c177-kube-api-access-fqkrm\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-zb8rn\" (UID: \"a84f5e60-25bb-4412-a38c-fa4e7a32c177\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zb8rn" Dec 11 08:34:00 crc kubenswrapper[4860]: I1211 08:34:00.836667 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a84f5e60-25bb-4412-a38c-fa4e7a32c177-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-zb8rn\" (UID: \"a84f5e60-25bb-4412-a38c-fa4e7a32c177\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zb8rn" Dec 11 08:34:00 crc kubenswrapper[4860]: I1211 08:34:00.836731 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a84f5e60-25bb-4412-a38c-fa4e7a32c177-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-zb8rn\" (UID: \"a84f5e60-25bb-4412-a38c-fa4e7a32c177\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zb8rn" Dec 11 08:34:00 crc kubenswrapper[4860]: I1211 08:34:00.841556 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a84f5e60-25bb-4412-a38c-fa4e7a32c177-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-zb8rn\" (UID: \"a84f5e60-25bb-4412-a38c-fa4e7a32c177\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zb8rn" Dec 11 08:34:00 crc kubenswrapper[4860]: I1211 08:34:00.843371 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a84f5e60-25bb-4412-a38c-fa4e7a32c177-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-zb8rn\" (UID: \"a84f5e60-25bb-4412-a38c-fa4e7a32c177\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zb8rn" Dec 11 08:34:00 crc kubenswrapper[4860]: I1211 08:34:00.853716 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fqkrm\" (UniqueName: \"kubernetes.io/projected/a84f5e60-25bb-4412-a38c-fa4e7a32c177-kube-api-access-fqkrm\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-zb8rn\" (UID: \"a84f5e60-25bb-4412-a38c-fa4e7a32c177\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zb8rn" Dec 11 08:34:01 crc kubenswrapper[4860]: I1211 08:34:01.028285 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zb8rn" Dec 11 08:34:01 crc kubenswrapper[4860]: I1211 08:34:01.576013 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-zb8rn"] Dec 11 08:34:01 crc kubenswrapper[4860]: I1211 08:34:01.609807 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zb8rn" event={"ID":"a84f5e60-25bb-4412-a38c-fa4e7a32c177","Type":"ContainerStarted","Data":"7731977f84cdc548c1c605d1c4e6c852de3cf96ddc550da6fb9157d089d0ce6c"} Dec 11 08:34:02 crc kubenswrapper[4860]: I1211 08:34:02.678844 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zb8rn" podStartSLOduration=1.913927992 podStartE2EDuration="2.678810244s" podCreationTimestamp="2025-12-11 08:34:00 +0000 UTC" firstStartedPulling="2025-12-11 08:34:01.58278372 +0000 UTC m=+1374.311302785" lastFinishedPulling="2025-12-11 08:34:02.347665982 +0000 UTC m=+1375.076185037" observedRunningTime="2025-12-11 08:34:02.669979024 +0000 UTC m=+1375.398498089" watchObservedRunningTime="2025-12-11 08:34:02.678810244 +0000 UTC m=+1375.407329299" Dec 11 08:34:03 crc kubenswrapper[4860]: I1211 08:34:03.648812 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zb8rn" event={"ID":"a84f5e60-25bb-4412-a38c-fa4e7a32c177","Type":"ContainerStarted","Data":"ca3ffa14519ab5a3830b88a260481970d042374e9cbd2ac366627411d8ffea49"} Dec 11 08:34:05 crc kubenswrapper[4860]: I1211 08:34:05.673678 4860 generic.go:334] "Generic (PLEG): container finished" podID="a84f5e60-25bb-4412-a38c-fa4e7a32c177" containerID="ca3ffa14519ab5a3830b88a260481970d042374e9cbd2ac366627411d8ffea49" exitCode=0 Dec 11 08:34:05 crc kubenswrapper[4860]: I1211 08:34:05.673716 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zb8rn" event={"ID":"a84f5e60-25bb-4412-a38c-fa4e7a32c177","Type":"ContainerDied","Data":"ca3ffa14519ab5a3830b88a260481970d042374e9cbd2ac366627411d8ffea49"} Dec 11 08:34:07 crc kubenswrapper[4860]: I1211 08:34:07.124077 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zb8rn" Dec 11 08:34:07 crc kubenswrapper[4860]: I1211 08:34:07.289013 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a84f5e60-25bb-4412-a38c-fa4e7a32c177-inventory\") pod \"a84f5e60-25bb-4412-a38c-fa4e7a32c177\" (UID: \"a84f5e60-25bb-4412-a38c-fa4e7a32c177\") " Dec 11 08:34:07 crc kubenswrapper[4860]: I1211 08:34:07.289154 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqkrm\" (UniqueName: \"kubernetes.io/projected/a84f5e60-25bb-4412-a38c-fa4e7a32c177-kube-api-access-fqkrm\") pod \"a84f5e60-25bb-4412-a38c-fa4e7a32c177\" (UID: \"a84f5e60-25bb-4412-a38c-fa4e7a32c177\") " Dec 11 08:34:07 crc kubenswrapper[4860]: I1211 08:34:07.289310 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a84f5e60-25bb-4412-a38c-fa4e7a32c177-ssh-key\") pod \"a84f5e60-25bb-4412-a38c-fa4e7a32c177\" (UID: \"a84f5e60-25bb-4412-a38c-fa4e7a32c177\") " Dec 11 08:34:07 crc kubenswrapper[4860]: I1211 08:34:07.303827 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a84f5e60-25bb-4412-a38c-fa4e7a32c177-kube-api-access-fqkrm" (OuterVolumeSpecName: "kube-api-access-fqkrm") pod "a84f5e60-25bb-4412-a38c-fa4e7a32c177" (UID: "a84f5e60-25bb-4412-a38c-fa4e7a32c177"). InnerVolumeSpecName "kube-api-access-fqkrm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:34:07 crc kubenswrapper[4860]: I1211 08:34:07.319766 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a84f5e60-25bb-4412-a38c-fa4e7a32c177-inventory" (OuterVolumeSpecName: "inventory") pod "a84f5e60-25bb-4412-a38c-fa4e7a32c177" (UID: "a84f5e60-25bb-4412-a38c-fa4e7a32c177"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:34:07 crc kubenswrapper[4860]: I1211 08:34:07.338015 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a84f5e60-25bb-4412-a38c-fa4e7a32c177-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "a84f5e60-25bb-4412-a38c-fa4e7a32c177" (UID: "a84f5e60-25bb-4412-a38c-fa4e7a32c177"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:34:07 crc kubenswrapper[4860]: I1211 08:34:07.391906 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqkrm\" (UniqueName: \"kubernetes.io/projected/a84f5e60-25bb-4412-a38c-fa4e7a32c177-kube-api-access-fqkrm\") on node \"crc\" DevicePath \"\"" Dec 11 08:34:07 crc kubenswrapper[4860]: I1211 08:34:07.392159 4860 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a84f5e60-25bb-4412-a38c-fa4e7a32c177-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 11 08:34:07 crc kubenswrapper[4860]: I1211 08:34:07.392277 4860 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a84f5e60-25bb-4412-a38c-fa4e7a32c177-inventory\") on node \"crc\" DevicePath \"\"" Dec 11 08:34:07 crc kubenswrapper[4860]: I1211 08:34:07.697749 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zb8rn" event={"ID":"a84f5e60-25bb-4412-a38c-fa4e7a32c177","Type":"ContainerDied","Data":"7731977f84cdc548c1c605d1c4e6c852de3cf96ddc550da6fb9157d089d0ce6c"} Dec 11 08:34:07 crc kubenswrapper[4860]: I1211 08:34:07.697793 4860 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7731977f84cdc548c1c605d1c4e6c852de3cf96ddc550da6fb9157d089d0ce6c" Dec 11 08:34:07 crc kubenswrapper[4860]: I1211 08:34:07.697818 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zb8rn" Dec 11 08:34:07 crc kubenswrapper[4860]: I1211 08:34:07.777022 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-tm5lx"] Dec 11 08:34:07 crc kubenswrapper[4860]: E1211 08:34:07.777586 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a84f5e60-25bb-4412-a38c-fa4e7a32c177" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 11 08:34:07 crc kubenswrapper[4860]: I1211 08:34:07.777613 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="a84f5e60-25bb-4412-a38c-fa4e7a32c177" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 11 08:34:07 crc kubenswrapper[4860]: I1211 08:34:07.777836 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="a84f5e60-25bb-4412-a38c-fa4e7a32c177" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 11 08:34:07 crc kubenswrapper[4860]: I1211 08:34:07.778617 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-tm5lx" Dec 11 08:34:07 crc kubenswrapper[4860]: I1211 08:34:07.781208 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-7w7kh" Dec 11 08:34:07 crc kubenswrapper[4860]: I1211 08:34:07.782090 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 11 08:34:07 crc kubenswrapper[4860]: I1211 08:34:07.782214 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 11 08:34:07 crc kubenswrapper[4860]: I1211 08:34:07.783104 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 11 08:34:07 crc kubenswrapper[4860]: I1211 08:34:07.788279 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-tm5lx"] Dec 11 08:34:07 crc kubenswrapper[4860]: I1211 08:34:07.908059 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hxx58\" (UniqueName: \"kubernetes.io/projected/aa177563-2177-4361-b45a-7ffe6f88da96-kube-api-access-hxx58\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-tm5lx\" (UID: \"aa177563-2177-4361-b45a-7ffe6f88da96\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-tm5lx" Dec 11 08:34:07 crc kubenswrapper[4860]: I1211 08:34:07.908135 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa177563-2177-4361-b45a-7ffe6f88da96-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-tm5lx\" (UID: \"aa177563-2177-4361-b45a-7ffe6f88da96\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-tm5lx" Dec 11 08:34:07 crc kubenswrapper[4860]: I1211 08:34:07.908199 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/aa177563-2177-4361-b45a-7ffe6f88da96-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-tm5lx\" (UID: \"aa177563-2177-4361-b45a-7ffe6f88da96\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-tm5lx" Dec 11 08:34:07 crc kubenswrapper[4860]: I1211 08:34:07.908352 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/aa177563-2177-4361-b45a-7ffe6f88da96-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-tm5lx\" (UID: \"aa177563-2177-4361-b45a-7ffe6f88da96\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-tm5lx" Dec 11 08:34:08 crc kubenswrapper[4860]: I1211 08:34:08.011090 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/aa177563-2177-4361-b45a-7ffe6f88da96-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-tm5lx\" (UID: \"aa177563-2177-4361-b45a-7ffe6f88da96\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-tm5lx" Dec 11 08:34:08 crc kubenswrapper[4860]: I1211 08:34:08.011679 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hxx58\" (UniqueName: \"kubernetes.io/projected/aa177563-2177-4361-b45a-7ffe6f88da96-kube-api-access-hxx58\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-tm5lx\" (UID: \"aa177563-2177-4361-b45a-7ffe6f88da96\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-tm5lx" Dec 11 08:34:08 crc kubenswrapper[4860]: I1211 08:34:08.011741 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa177563-2177-4361-b45a-7ffe6f88da96-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-tm5lx\" (UID: \"aa177563-2177-4361-b45a-7ffe6f88da96\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-tm5lx" Dec 11 08:34:08 crc kubenswrapper[4860]: I1211 08:34:08.011787 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/aa177563-2177-4361-b45a-7ffe6f88da96-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-tm5lx\" (UID: \"aa177563-2177-4361-b45a-7ffe6f88da96\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-tm5lx" Dec 11 08:34:08 crc kubenswrapper[4860]: I1211 08:34:08.015932 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/aa177563-2177-4361-b45a-7ffe6f88da96-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-tm5lx\" (UID: \"aa177563-2177-4361-b45a-7ffe6f88da96\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-tm5lx" Dec 11 08:34:08 crc kubenswrapper[4860]: I1211 08:34:08.016601 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa177563-2177-4361-b45a-7ffe6f88da96-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-tm5lx\" (UID: \"aa177563-2177-4361-b45a-7ffe6f88da96\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-tm5lx" Dec 11 08:34:08 crc kubenswrapper[4860]: I1211 08:34:08.016851 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/aa177563-2177-4361-b45a-7ffe6f88da96-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-tm5lx\" (UID: \"aa177563-2177-4361-b45a-7ffe6f88da96\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-tm5lx" Dec 11 08:34:08 crc kubenswrapper[4860]: I1211 08:34:08.035074 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hxx58\" (UniqueName: \"kubernetes.io/projected/aa177563-2177-4361-b45a-7ffe6f88da96-kube-api-access-hxx58\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-tm5lx\" (UID: \"aa177563-2177-4361-b45a-7ffe6f88da96\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-tm5lx" Dec 11 08:34:08 crc kubenswrapper[4860]: I1211 08:34:08.125900 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-tm5lx" Dec 11 08:34:08 crc kubenswrapper[4860]: I1211 08:34:08.701594 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-tm5lx"] Dec 11 08:34:08 crc kubenswrapper[4860]: I1211 08:34:08.713823 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-tm5lx" event={"ID":"aa177563-2177-4361-b45a-7ffe6f88da96","Type":"ContainerStarted","Data":"76124c006e33ee0eaf691050c63c335bccc525e09d15bee5968f95a573981e91"} Dec 11 08:34:09 crc kubenswrapper[4860]: I1211 08:34:09.725410 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-tm5lx" event={"ID":"aa177563-2177-4361-b45a-7ffe6f88da96","Type":"ContainerStarted","Data":"8408c746ccaa99847ae7661947c36d21ed338ca4cc372cab77281756df6b2a95"} Dec 11 08:34:09 crc kubenswrapper[4860]: I1211 08:34:09.776529 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-tm5lx" podStartSLOduration=2.030889547 podStartE2EDuration="2.776505829s" podCreationTimestamp="2025-12-11 08:34:07 +0000 UTC" firstStartedPulling="2025-12-11 08:34:08.704725113 +0000 UTC m=+1381.433244168" lastFinishedPulling="2025-12-11 08:34:09.450341395 +0000 UTC m=+1382.178860450" observedRunningTime="2025-12-11 08:34:09.760729459 +0000 UTC m=+1382.489248574" watchObservedRunningTime="2025-12-11 08:34:09.776505829 +0000 UTC m=+1382.505024884" Dec 11 08:34:30 crc kubenswrapper[4860]: I1211 08:34:30.764492 4860 scope.go:117] "RemoveContainer" containerID="ab9f83a10a88c0bcbaba9d0b040e40de344ecc46ca54bb9cd5a1f2984b476452" Dec 11 08:34:38 crc kubenswrapper[4860]: I1211 08:34:38.795047 4860 patch_prober.go:28] interesting pod/machine-config-daemon-99qgp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 08:34:38 crc kubenswrapper[4860]: I1211 08:34:38.795828 4860 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 08:34:54 crc kubenswrapper[4860]: I1211 08:34:54.454450 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-hv59h"] Dec 11 08:34:54 crc kubenswrapper[4860]: I1211 08:34:54.456997 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hv59h" Dec 11 08:34:54 crc kubenswrapper[4860]: I1211 08:34:54.465232 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hv59h"] Dec 11 08:34:54 crc kubenswrapper[4860]: I1211 08:34:54.535913 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5b2df11-7726-4de5-9e72-6dff618c20ad-catalog-content\") pod \"redhat-operators-hv59h\" (UID: \"d5b2df11-7726-4de5-9e72-6dff618c20ad\") " pod="openshift-marketplace/redhat-operators-hv59h" Dec 11 08:34:54 crc kubenswrapper[4860]: I1211 08:34:54.536062 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5b2df11-7726-4de5-9e72-6dff618c20ad-utilities\") pod \"redhat-operators-hv59h\" (UID: \"d5b2df11-7726-4de5-9e72-6dff618c20ad\") " pod="openshift-marketplace/redhat-operators-hv59h" Dec 11 08:34:54 crc kubenswrapper[4860]: I1211 08:34:54.536230 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rk7qs\" (UniqueName: \"kubernetes.io/projected/d5b2df11-7726-4de5-9e72-6dff618c20ad-kube-api-access-rk7qs\") pod \"redhat-operators-hv59h\" (UID: \"d5b2df11-7726-4de5-9e72-6dff618c20ad\") " pod="openshift-marketplace/redhat-operators-hv59h" Dec 11 08:34:54 crc kubenswrapper[4860]: I1211 08:34:54.637832 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rk7qs\" (UniqueName: \"kubernetes.io/projected/d5b2df11-7726-4de5-9e72-6dff618c20ad-kube-api-access-rk7qs\") pod \"redhat-operators-hv59h\" (UID: \"d5b2df11-7726-4de5-9e72-6dff618c20ad\") " pod="openshift-marketplace/redhat-operators-hv59h" Dec 11 08:34:54 crc kubenswrapper[4860]: I1211 08:34:54.637968 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5b2df11-7726-4de5-9e72-6dff618c20ad-catalog-content\") pod \"redhat-operators-hv59h\" (UID: \"d5b2df11-7726-4de5-9e72-6dff618c20ad\") " pod="openshift-marketplace/redhat-operators-hv59h" Dec 11 08:34:54 crc kubenswrapper[4860]: I1211 08:34:54.638023 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5b2df11-7726-4de5-9e72-6dff618c20ad-utilities\") pod \"redhat-operators-hv59h\" (UID: \"d5b2df11-7726-4de5-9e72-6dff618c20ad\") " pod="openshift-marketplace/redhat-operators-hv59h" Dec 11 08:34:54 crc kubenswrapper[4860]: I1211 08:34:54.638798 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5b2df11-7726-4de5-9e72-6dff618c20ad-utilities\") pod \"redhat-operators-hv59h\" (UID: \"d5b2df11-7726-4de5-9e72-6dff618c20ad\") " pod="openshift-marketplace/redhat-operators-hv59h" Dec 11 08:34:54 crc kubenswrapper[4860]: I1211 08:34:54.638859 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5b2df11-7726-4de5-9e72-6dff618c20ad-catalog-content\") pod \"redhat-operators-hv59h\" (UID: \"d5b2df11-7726-4de5-9e72-6dff618c20ad\") " pod="openshift-marketplace/redhat-operators-hv59h" Dec 11 08:34:54 crc kubenswrapper[4860]: I1211 08:34:54.658907 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rk7qs\" (UniqueName: \"kubernetes.io/projected/d5b2df11-7726-4de5-9e72-6dff618c20ad-kube-api-access-rk7qs\") pod \"redhat-operators-hv59h\" (UID: \"d5b2df11-7726-4de5-9e72-6dff618c20ad\") " pod="openshift-marketplace/redhat-operators-hv59h" Dec 11 08:34:54 crc kubenswrapper[4860]: I1211 08:34:54.818674 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hv59h" Dec 11 08:34:55 crc kubenswrapper[4860]: I1211 08:34:55.328035 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hv59h"] Dec 11 08:34:56 crc kubenswrapper[4860]: I1211 08:34:56.233079 4860 generic.go:334] "Generic (PLEG): container finished" podID="d5b2df11-7726-4de5-9e72-6dff618c20ad" containerID="78e4dfa75770c48cbf60bf7903f24a422ab1c6587da88c3babb05769cd9017d6" exitCode=0 Dec 11 08:34:56 crc kubenswrapper[4860]: I1211 08:34:56.233197 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hv59h" event={"ID":"d5b2df11-7726-4de5-9e72-6dff618c20ad","Type":"ContainerDied","Data":"78e4dfa75770c48cbf60bf7903f24a422ab1c6587da88c3babb05769cd9017d6"} Dec 11 08:34:56 crc kubenswrapper[4860]: I1211 08:34:56.233464 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hv59h" event={"ID":"d5b2df11-7726-4de5-9e72-6dff618c20ad","Type":"ContainerStarted","Data":"1cb9b0fa5690d9f292156d26597a326b9d6a591546d00f46846544b3d47db80c"} Dec 11 08:34:58 crc kubenswrapper[4860]: I1211 08:34:58.255297 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hv59h" event={"ID":"d5b2df11-7726-4de5-9e72-6dff618c20ad","Type":"ContainerStarted","Data":"d3fbc6b232b4482a2ee393cd621c928d2e0f10877d9cfde2b943c7db37d6451c"} Dec 11 08:34:59 crc kubenswrapper[4860]: I1211 08:34:59.267702 4860 generic.go:334] "Generic (PLEG): container finished" podID="d5b2df11-7726-4de5-9e72-6dff618c20ad" containerID="d3fbc6b232b4482a2ee393cd621c928d2e0f10877d9cfde2b943c7db37d6451c" exitCode=0 Dec 11 08:34:59 crc kubenswrapper[4860]: I1211 08:34:59.267826 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hv59h" event={"ID":"d5b2df11-7726-4de5-9e72-6dff618c20ad","Type":"ContainerDied","Data":"d3fbc6b232b4482a2ee393cd621c928d2e0f10877d9cfde2b943c7db37d6451c"} Dec 11 08:35:00 crc kubenswrapper[4860]: I1211 08:35:00.281059 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hv59h" event={"ID":"d5b2df11-7726-4de5-9e72-6dff618c20ad","Type":"ContainerStarted","Data":"fa0d1ca2d55aedf12e68f4f96feb3752d99dbbbd7ef2613d463bdea45b116f11"} Dec 11 08:35:00 crc kubenswrapper[4860]: I1211 08:35:00.311964 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-hv59h" podStartSLOduration=2.754595762 podStartE2EDuration="6.311946671s" podCreationTimestamp="2025-12-11 08:34:54 +0000 UTC" firstStartedPulling="2025-12-11 08:34:56.235901677 +0000 UTC m=+1428.964420742" lastFinishedPulling="2025-12-11 08:34:59.793252596 +0000 UTC m=+1432.521771651" observedRunningTime="2025-12-11 08:35:00.303602617 +0000 UTC m=+1433.032121672" watchObservedRunningTime="2025-12-11 08:35:00.311946671 +0000 UTC m=+1433.040465726" Dec 11 08:35:04 crc kubenswrapper[4860]: I1211 08:35:04.819352 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-hv59h" Dec 11 08:35:04 crc kubenswrapper[4860]: I1211 08:35:04.819903 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-hv59h" Dec 11 08:35:05 crc kubenswrapper[4860]: I1211 08:35:05.871496 4860 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-hv59h" podUID="d5b2df11-7726-4de5-9e72-6dff618c20ad" containerName="registry-server" probeResult="failure" output=< Dec 11 08:35:05 crc kubenswrapper[4860]: timeout: failed to connect service ":50051" within 1s Dec 11 08:35:05 crc kubenswrapper[4860]: > Dec 11 08:35:08 crc kubenswrapper[4860]: I1211 08:35:08.795036 4860 patch_prober.go:28] interesting pod/machine-config-daemon-99qgp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 08:35:08 crc kubenswrapper[4860]: I1211 08:35:08.795364 4860 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 08:35:14 crc kubenswrapper[4860]: I1211 08:35:14.880700 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-hv59h" Dec 11 08:35:14 crc kubenswrapper[4860]: I1211 08:35:14.930579 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-hv59h" Dec 11 08:35:15 crc kubenswrapper[4860]: I1211 08:35:15.116672 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hv59h"] Dec 11 08:35:16 crc kubenswrapper[4860]: I1211 08:35:16.439485 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-hv59h" podUID="d5b2df11-7726-4de5-9e72-6dff618c20ad" containerName="registry-server" containerID="cri-o://fa0d1ca2d55aedf12e68f4f96feb3752d99dbbbd7ef2613d463bdea45b116f11" gracePeriod=2 Dec 11 08:35:16 crc kubenswrapper[4860]: I1211 08:35:16.930743 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hv59h" Dec 11 08:35:17 crc kubenswrapper[4860]: I1211 08:35:17.120391 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rk7qs\" (UniqueName: \"kubernetes.io/projected/d5b2df11-7726-4de5-9e72-6dff618c20ad-kube-api-access-rk7qs\") pod \"d5b2df11-7726-4de5-9e72-6dff618c20ad\" (UID: \"d5b2df11-7726-4de5-9e72-6dff618c20ad\") " Dec 11 08:35:17 crc kubenswrapper[4860]: I1211 08:35:17.120725 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5b2df11-7726-4de5-9e72-6dff618c20ad-utilities\") pod \"d5b2df11-7726-4de5-9e72-6dff618c20ad\" (UID: \"d5b2df11-7726-4de5-9e72-6dff618c20ad\") " Dec 11 08:35:17 crc kubenswrapper[4860]: I1211 08:35:17.120778 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5b2df11-7726-4de5-9e72-6dff618c20ad-catalog-content\") pod \"d5b2df11-7726-4de5-9e72-6dff618c20ad\" (UID: \"d5b2df11-7726-4de5-9e72-6dff618c20ad\") " Dec 11 08:35:17 crc kubenswrapper[4860]: I1211 08:35:17.123701 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d5b2df11-7726-4de5-9e72-6dff618c20ad-utilities" (OuterVolumeSpecName: "utilities") pod "d5b2df11-7726-4de5-9e72-6dff618c20ad" (UID: "d5b2df11-7726-4de5-9e72-6dff618c20ad"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:35:17 crc kubenswrapper[4860]: I1211 08:35:17.160212 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d5b2df11-7726-4de5-9e72-6dff618c20ad-kube-api-access-rk7qs" (OuterVolumeSpecName: "kube-api-access-rk7qs") pod "d5b2df11-7726-4de5-9e72-6dff618c20ad" (UID: "d5b2df11-7726-4de5-9e72-6dff618c20ad"). InnerVolumeSpecName "kube-api-access-rk7qs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:35:17 crc kubenswrapper[4860]: I1211 08:35:17.223076 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rk7qs\" (UniqueName: \"kubernetes.io/projected/d5b2df11-7726-4de5-9e72-6dff618c20ad-kube-api-access-rk7qs\") on node \"crc\" DevicePath \"\"" Dec 11 08:35:17 crc kubenswrapper[4860]: I1211 08:35:17.223123 4860 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5b2df11-7726-4de5-9e72-6dff618c20ad-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 08:35:17 crc kubenswrapper[4860]: I1211 08:35:17.240633 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d5b2df11-7726-4de5-9e72-6dff618c20ad-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d5b2df11-7726-4de5-9e72-6dff618c20ad" (UID: "d5b2df11-7726-4de5-9e72-6dff618c20ad"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:35:17 crc kubenswrapper[4860]: I1211 08:35:17.324550 4860 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5b2df11-7726-4de5-9e72-6dff618c20ad-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 08:35:17 crc kubenswrapper[4860]: I1211 08:35:17.454218 4860 generic.go:334] "Generic (PLEG): container finished" podID="d5b2df11-7726-4de5-9e72-6dff618c20ad" containerID="fa0d1ca2d55aedf12e68f4f96feb3752d99dbbbd7ef2613d463bdea45b116f11" exitCode=0 Dec 11 08:35:17 crc kubenswrapper[4860]: I1211 08:35:17.454276 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hv59h" event={"ID":"d5b2df11-7726-4de5-9e72-6dff618c20ad","Type":"ContainerDied","Data":"fa0d1ca2d55aedf12e68f4f96feb3752d99dbbbd7ef2613d463bdea45b116f11"} Dec 11 08:35:17 crc kubenswrapper[4860]: I1211 08:35:17.454291 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hv59h" Dec 11 08:35:17 crc kubenswrapper[4860]: I1211 08:35:17.454311 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hv59h" event={"ID":"d5b2df11-7726-4de5-9e72-6dff618c20ad","Type":"ContainerDied","Data":"1cb9b0fa5690d9f292156d26597a326b9d6a591546d00f46846544b3d47db80c"} Dec 11 08:35:17 crc kubenswrapper[4860]: I1211 08:35:17.454333 4860 scope.go:117] "RemoveContainer" containerID="fa0d1ca2d55aedf12e68f4f96feb3752d99dbbbd7ef2613d463bdea45b116f11" Dec 11 08:35:17 crc kubenswrapper[4860]: I1211 08:35:17.496052 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hv59h"] Dec 11 08:35:17 crc kubenswrapper[4860]: I1211 08:35:17.496124 4860 scope.go:117] "RemoveContainer" containerID="d3fbc6b232b4482a2ee393cd621c928d2e0f10877d9cfde2b943c7db37d6451c" Dec 11 08:35:17 crc kubenswrapper[4860]: I1211 08:35:17.505406 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-hv59h"] Dec 11 08:35:17 crc kubenswrapper[4860]: I1211 08:35:17.526572 4860 scope.go:117] "RemoveContainer" containerID="78e4dfa75770c48cbf60bf7903f24a422ab1c6587da88c3babb05769cd9017d6" Dec 11 08:35:17 crc kubenswrapper[4860]: I1211 08:35:17.573948 4860 scope.go:117] "RemoveContainer" containerID="fa0d1ca2d55aedf12e68f4f96feb3752d99dbbbd7ef2613d463bdea45b116f11" Dec 11 08:35:17 crc kubenswrapper[4860]: E1211 08:35:17.574963 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fa0d1ca2d55aedf12e68f4f96feb3752d99dbbbd7ef2613d463bdea45b116f11\": container with ID starting with fa0d1ca2d55aedf12e68f4f96feb3752d99dbbbd7ef2613d463bdea45b116f11 not found: ID does not exist" containerID="fa0d1ca2d55aedf12e68f4f96feb3752d99dbbbd7ef2613d463bdea45b116f11" Dec 11 08:35:17 crc kubenswrapper[4860]: I1211 08:35:17.575097 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fa0d1ca2d55aedf12e68f4f96feb3752d99dbbbd7ef2613d463bdea45b116f11"} err="failed to get container status \"fa0d1ca2d55aedf12e68f4f96feb3752d99dbbbd7ef2613d463bdea45b116f11\": rpc error: code = NotFound desc = could not find container \"fa0d1ca2d55aedf12e68f4f96feb3752d99dbbbd7ef2613d463bdea45b116f11\": container with ID starting with fa0d1ca2d55aedf12e68f4f96feb3752d99dbbbd7ef2613d463bdea45b116f11 not found: ID does not exist" Dec 11 08:35:17 crc kubenswrapper[4860]: I1211 08:35:17.575238 4860 scope.go:117] "RemoveContainer" containerID="d3fbc6b232b4482a2ee393cd621c928d2e0f10877d9cfde2b943c7db37d6451c" Dec 11 08:35:17 crc kubenswrapper[4860]: E1211 08:35:17.575683 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d3fbc6b232b4482a2ee393cd621c928d2e0f10877d9cfde2b943c7db37d6451c\": container with ID starting with d3fbc6b232b4482a2ee393cd621c928d2e0f10877d9cfde2b943c7db37d6451c not found: ID does not exist" containerID="d3fbc6b232b4482a2ee393cd621c928d2e0f10877d9cfde2b943c7db37d6451c" Dec 11 08:35:17 crc kubenswrapper[4860]: I1211 08:35:17.575797 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d3fbc6b232b4482a2ee393cd621c928d2e0f10877d9cfde2b943c7db37d6451c"} err="failed to get container status \"d3fbc6b232b4482a2ee393cd621c928d2e0f10877d9cfde2b943c7db37d6451c\": rpc error: code = NotFound desc = could not find container \"d3fbc6b232b4482a2ee393cd621c928d2e0f10877d9cfde2b943c7db37d6451c\": container with ID starting with d3fbc6b232b4482a2ee393cd621c928d2e0f10877d9cfde2b943c7db37d6451c not found: ID does not exist" Dec 11 08:35:17 crc kubenswrapper[4860]: I1211 08:35:17.575896 4860 scope.go:117] "RemoveContainer" containerID="78e4dfa75770c48cbf60bf7903f24a422ab1c6587da88c3babb05769cd9017d6" Dec 11 08:35:17 crc kubenswrapper[4860]: E1211 08:35:17.576268 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"78e4dfa75770c48cbf60bf7903f24a422ab1c6587da88c3babb05769cd9017d6\": container with ID starting with 78e4dfa75770c48cbf60bf7903f24a422ab1c6587da88c3babb05769cd9017d6 not found: ID does not exist" containerID="78e4dfa75770c48cbf60bf7903f24a422ab1c6587da88c3babb05769cd9017d6" Dec 11 08:35:17 crc kubenswrapper[4860]: I1211 08:35:17.576383 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"78e4dfa75770c48cbf60bf7903f24a422ab1c6587da88c3babb05769cd9017d6"} err="failed to get container status \"78e4dfa75770c48cbf60bf7903f24a422ab1c6587da88c3babb05769cd9017d6\": rpc error: code = NotFound desc = could not find container \"78e4dfa75770c48cbf60bf7903f24a422ab1c6587da88c3babb05769cd9017d6\": container with ID starting with 78e4dfa75770c48cbf60bf7903f24a422ab1c6587da88c3babb05769cd9017d6 not found: ID does not exist" Dec 11 08:35:17 crc kubenswrapper[4860]: I1211 08:35:17.599867 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d5b2df11-7726-4de5-9e72-6dff618c20ad" path="/var/lib/kubelet/pods/d5b2df11-7726-4de5-9e72-6dff618c20ad/volumes" Dec 11 08:35:27 crc kubenswrapper[4860]: I1211 08:35:27.540794 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-mqj28"] Dec 11 08:35:27 crc kubenswrapper[4860]: E1211 08:35:27.541806 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5b2df11-7726-4de5-9e72-6dff618c20ad" containerName="extract-content" Dec 11 08:35:27 crc kubenswrapper[4860]: I1211 08:35:27.541823 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5b2df11-7726-4de5-9e72-6dff618c20ad" containerName="extract-content" Dec 11 08:35:27 crc kubenswrapper[4860]: E1211 08:35:27.541859 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5b2df11-7726-4de5-9e72-6dff618c20ad" containerName="extract-utilities" Dec 11 08:35:27 crc kubenswrapper[4860]: I1211 08:35:27.541865 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5b2df11-7726-4de5-9e72-6dff618c20ad" containerName="extract-utilities" Dec 11 08:35:27 crc kubenswrapper[4860]: E1211 08:35:27.541886 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5b2df11-7726-4de5-9e72-6dff618c20ad" containerName="registry-server" Dec 11 08:35:27 crc kubenswrapper[4860]: I1211 08:35:27.541892 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5b2df11-7726-4de5-9e72-6dff618c20ad" containerName="registry-server" Dec 11 08:35:27 crc kubenswrapper[4860]: I1211 08:35:27.542097 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="d5b2df11-7726-4de5-9e72-6dff618c20ad" containerName="registry-server" Dec 11 08:35:27 crc kubenswrapper[4860]: I1211 08:35:27.543619 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mqj28" Dec 11 08:35:27 crc kubenswrapper[4860]: I1211 08:35:27.551347 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mqj28"] Dec 11 08:35:27 crc kubenswrapper[4860]: I1211 08:35:27.589555 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c4e3abdc-f78c-48da-a9fc-d2027fcb29b4-catalog-content\") pod \"certified-operators-mqj28\" (UID: \"c4e3abdc-f78c-48da-a9fc-d2027fcb29b4\") " pod="openshift-marketplace/certified-operators-mqj28" Dec 11 08:35:27 crc kubenswrapper[4860]: I1211 08:35:27.589674 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v9svh\" (UniqueName: \"kubernetes.io/projected/c4e3abdc-f78c-48da-a9fc-d2027fcb29b4-kube-api-access-v9svh\") pod \"certified-operators-mqj28\" (UID: \"c4e3abdc-f78c-48da-a9fc-d2027fcb29b4\") " pod="openshift-marketplace/certified-operators-mqj28" Dec 11 08:35:27 crc kubenswrapper[4860]: I1211 08:35:27.589799 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c4e3abdc-f78c-48da-a9fc-d2027fcb29b4-utilities\") pod \"certified-operators-mqj28\" (UID: \"c4e3abdc-f78c-48da-a9fc-d2027fcb29b4\") " pod="openshift-marketplace/certified-operators-mqj28" Dec 11 08:35:27 crc kubenswrapper[4860]: I1211 08:35:27.691663 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v9svh\" (UniqueName: \"kubernetes.io/projected/c4e3abdc-f78c-48da-a9fc-d2027fcb29b4-kube-api-access-v9svh\") pod \"certified-operators-mqj28\" (UID: \"c4e3abdc-f78c-48da-a9fc-d2027fcb29b4\") " pod="openshift-marketplace/certified-operators-mqj28" Dec 11 08:35:27 crc kubenswrapper[4860]: I1211 08:35:27.692153 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c4e3abdc-f78c-48da-a9fc-d2027fcb29b4-utilities\") pod \"certified-operators-mqj28\" (UID: \"c4e3abdc-f78c-48da-a9fc-d2027fcb29b4\") " pod="openshift-marketplace/certified-operators-mqj28" Dec 11 08:35:27 crc kubenswrapper[4860]: I1211 08:35:27.692359 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c4e3abdc-f78c-48da-a9fc-d2027fcb29b4-catalog-content\") pod \"certified-operators-mqj28\" (UID: \"c4e3abdc-f78c-48da-a9fc-d2027fcb29b4\") " pod="openshift-marketplace/certified-operators-mqj28" Dec 11 08:35:27 crc kubenswrapper[4860]: I1211 08:35:27.692938 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c4e3abdc-f78c-48da-a9fc-d2027fcb29b4-catalog-content\") pod \"certified-operators-mqj28\" (UID: \"c4e3abdc-f78c-48da-a9fc-d2027fcb29b4\") " pod="openshift-marketplace/certified-operators-mqj28" Dec 11 08:35:27 crc kubenswrapper[4860]: I1211 08:35:27.693056 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c4e3abdc-f78c-48da-a9fc-d2027fcb29b4-utilities\") pod \"certified-operators-mqj28\" (UID: \"c4e3abdc-f78c-48da-a9fc-d2027fcb29b4\") " pod="openshift-marketplace/certified-operators-mqj28" Dec 11 08:35:27 crc kubenswrapper[4860]: I1211 08:35:27.752608 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v9svh\" (UniqueName: \"kubernetes.io/projected/c4e3abdc-f78c-48da-a9fc-d2027fcb29b4-kube-api-access-v9svh\") pod \"certified-operators-mqj28\" (UID: \"c4e3abdc-f78c-48da-a9fc-d2027fcb29b4\") " pod="openshift-marketplace/certified-operators-mqj28" Dec 11 08:35:27 crc kubenswrapper[4860]: I1211 08:35:27.873136 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mqj28" Dec 11 08:35:28 crc kubenswrapper[4860]: I1211 08:35:28.952007 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mqj28"] Dec 11 08:35:29 crc kubenswrapper[4860]: I1211 08:35:29.568097 4860 generic.go:334] "Generic (PLEG): container finished" podID="c4e3abdc-f78c-48da-a9fc-d2027fcb29b4" containerID="9c780ffd4df6ec69bea07ca64ad4f51c7dd459e1fbbdab2bffeab2afcad6d071" exitCode=0 Dec 11 08:35:29 crc kubenswrapper[4860]: I1211 08:35:29.568163 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mqj28" event={"ID":"c4e3abdc-f78c-48da-a9fc-d2027fcb29b4","Type":"ContainerDied","Data":"9c780ffd4df6ec69bea07ca64ad4f51c7dd459e1fbbdab2bffeab2afcad6d071"} Dec 11 08:35:29 crc kubenswrapper[4860]: I1211 08:35:29.568422 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mqj28" event={"ID":"c4e3abdc-f78c-48da-a9fc-d2027fcb29b4","Type":"ContainerStarted","Data":"4e8df5b11eb4373e6f10e75ae49878d661c503844a69c11a989ec88554e78b6d"} Dec 11 08:35:30 crc kubenswrapper[4860]: I1211 08:35:30.583141 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mqj28" event={"ID":"c4e3abdc-f78c-48da-a9fc-d2027fcb29b4","Type":"ContainerStarted","Data":"29244305349e1f01d70c59bd89f07f8e3d374931a45a9c3e8fd877e9aa3d48bd"} Dec 11 08:35:30 crc kubenswrapper[4860]: I1211 08:35:30.842623 4860 scope.go:117] "RemoveContainer" containerID="3038d16b28f6b130939667f3daca0bc372de9ac7d60de3e167f8e56da8339937" Dec 11 08:35:31 crc kubenswrapper[4860]: I1211 08:35:31.625444 4860 generic.go:334] "Generic (PLEG): container finished" podID="c4e3abdc-f78c-48da-a9fc-d2027fcb29b4" containerID="29244305349e1f01d70c59bd89f07f8e3d374931a45a9c3e8fd877e9aa3d48bd" exitCode=0 Dec 11 08:35:31 crc kubenswrapper[4860]: I1211 08:35:31.625543 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mqj28" event={"ID":"c4e3abdc-f78c-48da-a9fc-d2027fcb29b4","Type":"ContainerDied","Data":"29244305349e1f01d70c59bd89f07f8e3d374931a45a9c3e8fd877e9aa3d48bd"} Dec 11 08:35:32 crc kubenswrapper[4860]: I1211 08:35:32.637934 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mqj28" event={"ID":"c4e3abdc-f78c-48da-a9fc-d2027fcb29b4","Type":"ContainerStarted","Data":"4f7ef57f442b3748df35b7139bd2832226ceeb68d2d134db9411b98f47b2eeed"} Dec 11 08:35:32 crc kubenswrapper[4860]: I1211 08:35:32.661236 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-mqj28" podStartSLOduration=2.998657943 podStartE2EDuration="5.661214235s" podCreationTimestamp="2025-12-11 08:35:27 +0000 UTC" firstStartedPulling="2025-12-11 08:35:29.570055249 +0000 UTC m=+1462.298574314" lastFinishedPulling="2025-12-11 08:35:32.232611541 +0000 UTC m=+1464.961130606" observedRunningTime="2025-12-11 08:35:32.653840602 +0000 UTC m=+1465.382359677" watchObservedRunningTime="2025-12-11 08:35:32.661214235 +0000 UTC m=+1465.389733290" Dec 11 08:35:37 crc kubenswrapper[4860]: I1211 08:35:37.873488 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-mqj28" Dec 11 08:35:37 crc kubenswrapper[4860]: I1211 08:35:37.873868 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-mqj28" Dec 11 08:35:37 crc kubenswrapper[4860]: I1211 08:35:37.935449 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-mqj28" Dec 11 08:35:38 crc kubenswrapper[4860]: I1211 08:35:38.782583 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-mqj28" Dec 11 08:35:38 crc kubenswrapper[4860]: I1211 08:35:38.795068 4860 patch_prober.go:28] interesting pod/machine-config-daemon-99qgp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 08:35:38 crc kubenswrapper[4860]: I1211 08:35:38.795141 4860 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 08:35:38 crc kubenswrapper[4860]: I1211 08:35:38.795203 4860 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" Dec 11 08:35:38 crc kubenswrapper[4860]: I1211 08:35:38.795964 4860 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2c5567d203277d4236f1f4cd1d73721ecd8809240167b07eac4086db3323fa9f"} pod="openshift-machine-config-operator/machine-config-daemon-99qgp" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 11 08:35:38 crc kubenswrapper[4860]: I1211 08:35:38.796033 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" containerName="machine-config-daemon" containerID="cri-o://2c5567d203277d4236f1f4cd1d73721ecd8809240167b07eac4086db3323fa9f" gracePeriod=600 Dec 11 08:35:38 crc kubenswrapper[4860]: I1211 08:35:38.847984 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mqj28"] Dec 11 08:35:39 crc kubenswrapper[4860]: I1211 08:35:39.724001 4860 generic.go:334] "Generic (PLEG): container finished" podID="31c30642-6e60-41b4-a477-0d802424e0aa" containerID="2c5567d203277d4236f1f4cd1d73721ecd8809240167b07eac4086db3323fa9f" exitCode=0 Dec 11 08:35:39 crc kubenswrapper[4860]: I1211 08:35:39.724032 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" event={"ID":"31c30642-6e60-41b4-a477-0d802424e0aa","Type":"ContainerDied","Data":"2c5567d203277d4236f1f4cd1d73721ecd8809240167b07eac4086db3323fa9f"} Dec 11 08:35:39 crc kubenswrapper[4860]: I1211 08:35:39.724398 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" event={"ID":"31c30642-6e60-41b4-a477-0d802424e0aa","Type":"ContainerStarted","Data":"af76095360a11870b804fb7f16c40dad604dbf93152022249293687b1519ca1b"} Dec 11 08:35:39 crc kubenswrapper[4860]: I1211 08:35:39.724437 4860 scope.go:117] "RemoveContainer" containerID="6be7e2592a74fa92891c9f53dda9c3fdad0d78e647aef9f0b578adb34ea14caa" Dec 11 08:35:40 crc kubenswrapper[4860]: I1211 08:35:40.738560 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-mqj28" podUID="c4e3abdc-f78c-48da-a9fc-d2027fcb29b4" containerName="registry-server" containerID="cri-o://4f7ef57f442b3748df35b7139bd2832226ceeb68d2d134db9411b98f47b2eeed" gracePeriod=2 Dec 11 08:35:41 crc kubenswrapper[4860]: I1211 08:35:41.383525 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mqj28" Dec 11 08:35:41 crc kubenswrapper[4860]: I1211 08:35:41.508903 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v9svh\" (UniqueName: \"kubernetes.io/projected/c4e3abdc-f78c-48da-a9fc-d2027fcb29b4-kube-api-access-v9svh\") pod \"c4e3abdc-f78c-48da-a9fc-d2027fcb29b4\" (UID: \"c4e3abdc-f78c-48da-a9fc-d2027fcb29b4\") " Dec 11 08:35:41 crc kubenswrapper[4860]: I1211 08:35:41.509077 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c4e3abdc-f78c-48da-a9fc-d2027fcb29b4-catalog-content\") pod \"c4e3abdc-f78c-48da-a9fc-d2027fcb29b4\" (UID: \"c4e3abdc-f78c-48da-a9fc-d2027fcb29b4\") " Dec 11 08:35:41 crc kubenswrapper[4860]: I1211 08:35:41.509184 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c4e3abdc-f78c-48da-a9fc-d2027fcb29b4-utilities\") pod \"c4e3abdc-f78c-48da-a9fc-d2027fcb29b4\" (UID: \"c4e3abdc-f78c-48da-a9fc-d2027fcb29b4\") " Dec 11 08:35:41 crc kubenswrapper[4860]: I1211 08:35:41.510354 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c4e3abdc-f78c-48da-a9fc-d2027fcb29b4-utilities" (OuterVolumeSpecName: "utilities") pod "c4e3abdc-f78c-48da-a9fc-d2027fcb29b4" (UID: "c4e3abdc-f78c-48da-a9fc-d2027fcb29b4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:35:41 crc kubenswrapper[4860]: I1211 08:35:41.520750 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c4e3abdc-f78c-48da-a9fc-d2027fcb29b4-kube-api-access-v9svh" (OuterVolumeSpecName: "kube-api-access-v9svh") pod "c4e3abdc-f78c-48da-a9fc-d2027fcb29b4" (UID: "c4e3abdc-f78c-48da-a9fc-d2027fcb29b4"). InnerVolumeSpecName "kube-api-access-v9svh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:35:41 crc kubenswrapper[4860]: I1211 08:35:41.566391 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c4e3abdc-f78c-48da-a9fc-d2027fcb29b4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c4e3abdc-f78c-48da-a9fc-d2027fcb29b4" (UID: "c4e3abdc-f78c-48da-a9fc-d2027fcb29b4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:35:41 crc kubenswrapper[4860]: I1211 08:35:41.611912 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v9svh\" (UniqueName: \"kubernetes.io/projected/c4e3abdc-f78c-48da-a9fc-d2027fcb29b4-kube-api-access-v9svh\") on node \"crc\" DevicePath \"\"" Dec 11 08:35:41 crc kubenswrapper[4860]: I1211 08:35:41.611944 4860 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c4e3abdc-f78c-48da-a9fc-d2027fcb29b4-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 08:35:41 crc kubenswrapper[4860]: I1211 08:35:41.611954 4860 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c4e3abdc-f78c-48da-a9fc-d2027fcb29b4-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 08:35:41 crc kubenswrapper[4860]: I1211 08:35:41.750601 4860 generic.go:334] "Generic (PLEG): container finished" podID="c4e3abdc-f78c-48da-a9fc-d2027fcb29b4" containerID="4f7ef57f442b3748df35b7139bd2832226ceeb68d2d134db9411b98f47b2eeed" exitCode=0 Dec 11 08:35:41 crc kubenswrapper[4860]: I1211 08:35:41.750663 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mqj28" event={"ID":"c4e3abdc-f78c-48da-a9fc-d2027fcb29b4","Type":"ContainerDied","Data":"4f7ef57f442b3748df35b7139bd2832226ceeb68d2d134db9411b98f47b2eeed"} Dec 11 08:35:41 crc kubenswrapper[4860]: I1211 08:35:41.750692 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mqj28" event={"ID":"c4e3abdc-f78c-48da-a9fc-d2027fcb29b4","Type":"ContainerDied","Data":"4e8df5b11eb4373e6f10e75ae49878d661c503844a69c11a989ec88554e78b6d"} Dec 11 08:35:41 crc kubenswrapper[4860]: I1211 08:35:41.750735 4860 scope.go:117] "RemoveContainer" containerID="4f7ef57f442b3748df35b7139bd2832226ceeb68d2d134db9411b98f47b2eeed" Dec 11 08:35:41 crc kubenswrapper[4860]: I1211 08:35:41.750765 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mqj28" Dec 11 08:35:41 crc kubenswrapper[4860]: I1211 08:35:41.779269 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mqj28"] Dec 11 08:35:41 crc kubenswrapper[4860]: I1211 08:35:41.784973 4860 scope.go:117] "RemoveContainer" containerID="29244305349e1f01d70c59bd89f07f8e3d374931a45a9c3e8fd877e9aa3d48bd" Dec 11 08:35:41 crc kubenswrapper[4860]: I1211 08:35:41.791091 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-mqj28"] Dec 11 08:35:41 crc kubenswrapper[4860]: I1211 08:35:41.808568 4860 scope.go:117] "RemoveContainer" containerID="9c780ffd4df6ec69bea07ca64ad4f51c7dd459e1fbbdab2bffeab2afcad6d071" Dec 11 08:35:41 crc kubenswrapper[4860]: I1211 08:35:41.883042 4860 scope.go:117] "RemoveContainer" containerID="4f7ef57f442b3748df35b7139bd2832226ceeb68d2d134db9411b98f47b2eeed" Dec 11 08:35:41 crc kubenswrapper[4860]: E1211 08:35:41.883613 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4f7ef57f442b3748df35b7139bd2832226ceeb68d2d134db9411b98f47b2eeed\": container with ID starting with 4f7ef57f442b3748df35b7139bd2832226ceeb68d2d134db9411b98f47b2eeed not found: ID does not exist" containerID="4f7ef57f442b3748df35b7139bd2832226ceeb68d2d134db9411b98f47b2eeed" Dec 11 08:35:41 crc kubenswrapper[4860]: I1211 08:35:41.883657 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4f7ef57f442b3748df35b7139bd2832226ceeb68d2d134db9411b98f47b2eeed"} err="failed to get container status \"4f7ef57f442b3748df35b7139bd2832226ceeb68d2d134db9411b98f47b2eeed\": rpc error: code = NotFound desc = could not find container \"4f7ef57f442b3748df35b7139bd2832226ceeb68d2d134db9411b98f47b2eeed\": container with ID starting with 4f7ef57f442b3748df35b7139bd2832226ceeb68d2d134db9411b98f47b2eeed not found: ID does not exist" Dec 11 08:35:41 crc kubenswrapper[4860]: I1211 08:35:41.883691 4860 scope.go:117] "RemoveContainer" containerID="29244305349e1f01d70c59bd89f07f8e3d374931a45a9c3e8fd877e9aa3d48bd" Dec 11 08:35:41 crc kubenswrapper[4860]: E1211 08:35:41.884118 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"29244305349e1f01d70c59bd89f07f8e3d374931a45a9c3e8fd877e9aa3d48bd\": container with ID starting with 29244305349e1f01d70c59bd89f07f8e3d374931a45a9c3e8fd877e9aa3d48bd not found: ID does not exist" containerID="29244305349e1f01d70c59bd89f07f8e3d374931a45a9c3e8fd877e9aa3d48bd" Dec 11 08:35:41 crc kubenswrapper[4860]: I1211 08:35:41.884144 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"29244305349e1f01d70c59bd89f07f8e3d374931a45a9c3e8fd877e9aa3d48bd"} err="failed to get container status \"29244305349e1f01d70c59bd89f07f8e3d374931a45a9c3e8fd877e9aa3d48bd\": rpc error: code = NotFound desc = could not find container \"29244305349e1f01d70c59bd89f07f8e3d374931a45a9c3e8fd877e9aa3d48bd\": container with ID starting with 29244305349e1f01d70c59bd89f07f8e3d374931a45a9c3e8fd877e9aa3d48bd not found: ID does not exist" Dec 11 08:35:41 crc kubenswrapper[4860]: I1211 08:35:41.884158 4860 scope.go:117] "RemoveContainer" containerID="9c780ffd4df6ec69bea07ca64ad4f51c7dd459e1fbbdab2bffeab2afcad6d071" Dec 11 08:35:41 crc kubenswrapper[4860]: E1211 08:35:41.884581 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9c780ffd4df6ec69bea07ca64ad4f51c7dd459e1fbbdab2bffeab2afcad6d071\": container with ID starting with 9c780ffd4df6ec69bea07ca64ad4f51c7dd459e1fbbdab2bffeab2afcad6d071 not found: ID does not exist" containerID="9c780ffd4df6ec69bea07ca64ad4f51c7dd459e1fbbdab2bffeab2afcad6d071" Dec 11 08:35:41 crc kubenswrapper[4860]: I1211 08:35:41.884605 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c780ffd4df6ec69bea07ca64ad4f51c7dd459e1fbbdab2bffeab2afcad6d071"} err="failed to get container status \"9c780ffd4df6ec69bea07ca64ad4f51c7dd459e1fbbdab2bffeab2afcad6d071\": rpc error: code = NotFound desc = could not find container \"9c780ffd4df6ec69bea07ca64ad4f51c7dd459e1fbbdab2bffeab2afcad6d071\": container with ID starting with 9c780ffd4df6ec69bea07ca64ad4f51c7dd459e1fbbdab2bffeab2afcad6d071 not found: ID does not exist" Dec 11 08:35:43 crc kubenswrapper[4860]: I1211 08:35:43.593758 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c4e3abdc-f78c-48da-a9fc-d2027fcb29b4" path="/var/lib/kubelet/pods/c4e3abdc-f78c-48da-a9fc-d2027fcb29b4/volumes" Dec 11 08:36:13 crc kubenswrapper[4860]: I1211 08:36:13.695518 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-btpkj"] Dec 11 08:36:13 crc kubenswrapper[4860]: E1211 08:36:13.696534 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4e3abdc-f78c-48da-a9fc-d2027fcb29b4" containerName="extract-utilities" Dec 11 08:36:13 crc kubenswrapper[4860]: I1211 08:36:13.696549 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4e3abdc-f78c-48da-a9fc-d2027fcb29b4" containerName="extract-utilities" Dec 11 08:36:13 crc kubenswrapper[4860]: E1211 08:36:13.696587 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4e3abdc-f78c-48da-a9fc-d2027fcb29b4" containerName="registry-server" Dec 11 08:36:13 crc kubenswrapper[4860]: I1211 08:36:13.696598 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4e3abdc-f78c-48da-a9fc-d2027fcb29b4" containerName="registry-server" Dec 11 08:36:13 crc kubenswrapper[4860]: E1211 08:36:13.696611 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4e3abdc-f78c-48da-a9fc-d2027fcb29b4" containerName="extract-content" Dec 11 08:36:13 crc kubenswrapper[4860]: I1211 08:36:13.696619 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4e3abdc-f78c-48da-a9fc-d2027fcb29b4" containerName="extract-content" Dec 11 08:36:13 crc kubenswrapper[4860]: I1211 08:36:13.696872 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4e3abdc-f78c-48da-a9fc-d2027fcb29b4" containerName="registry-server" Dec 11 08:36:13 crc kubenswrapper[4860]: I1211 08:36:13.698504 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-btpkj" Dec 11 08:36:13 crc kubenswrapper[4860]: I1211 08:36:13.717375 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-btpkj"] Dec 11 08:36:13 crc kubenswrapper[4860]: I1211 08:36:13.847228 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c7cb44e7-8a83-4089-9808-962e06331d46-catalog-content\") pod \"redhat-marketplace-btpkj\" (UID: \"c7cb44e7-8a83-4089-9808-962e06331d46\") " pod="openshift-marketplace/redhat-marketplace-btpkj" Dec 11 08:36:13 crc kubenswrapper[4860]: I1211 08:36:13.847585 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c7cb44e7-8a83-4089-9808-962e06331d46-utilities\") pod \"redhat-marketplace-btpkj\" (UID: \"c7cb44e7-8a83-4089-9808-962e06331d46\") " pod="openshift-marketplace/redhat-marketplace-btpkj" Dec 11 08:36:13 crc kubenswrapper[4860]: I1211 08:36:13.847802 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s5t82\" (UniqueName: \"kubernetes.io/projected/c7cb44e7-8a83-4089-9808-962e06331d46-kube-api-access-s5t82\") pod \"redhat-marketplace-btpkj\" (UID: \"c7cb44e7-8a83-4089-9808-962e06331d46\") " pod="openshift-marketplace/redhat-marketplace-btpkj" Dec 11 08:36:13 crc kubenswrapper[4860]: I1211 08:36:13.949714 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c7cb44e7-8a83-4089-9808-962e06331d46-catalog-content\") pod \"redhat-marketplace-btpkj\" (UID: \"c7cb44e7-8a83-4089-9808-962e06331d46\") " pod="openshift-marketplace/redhat-marketplace-btpkj" Dec 11 08:36:13 crc kubenswrapper[4860]: I1211 08:36:13.949807 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c7cb44e7-8a83-4089-9808-962e06331d46-utilities\") pod \"redhat-marketplace-btpkj\" (UID: \"c7cb44e7-8a83-4089-9808-962e06331d46\") " pod="openshift-marketplace/redhat-marketplace-btpkj" Dec 11 08:36:13 crc kubenswrapper[4860]: I1211 08:36:13.949923 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s5t82\" (UniqueName: \"kubernetes.io/projected/c7cb44e7-8a83-4089-9808-962e06331d46-kube-api-access-s5t82\") pod \"redhat-marketplace-btpkj\" (UID: \"c7cb44e7-8a83-4089-9808-962e06331d46\") " pod="openshift-marketplace/redhat-marketplace-btpkj" Dec 11 08:36:13 crc kubenswrapper[4860]: I1211 08:36:13.950859 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c7cb44e7-8a83-4089-9808-962e06331d46-catalog-content\") pod \"redhat-marketplace-btpkj\" (UID: \"c7cb44e7-8a83-4089-9808-962e06331d46\") " pod="openshift-marketplace/redhat-marketplace-btpkj" Dec 11 08:36:13 crc kubenswrapper[4860]: I1211 08:36:13.950872 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c7cb44e7-8a83-4089-9808-962e06331d46-utilities\") pod \"redhat-marketplace-btpkj\" (UID: \"c7cb44e7-8a83-4089-9808-962e06331d46\") " pod="openshift-marketplace/redhat-marketplace-btpkj" Dec 11 08:36:13 crc kubenswrapper[4860]: I1211 08:36:13.978531 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s5t82\" (UniqueName: \"kubernetes.io/projected/c7cb44e7-8a83-4089-9808-962e06331d46-kube-api-access-s5t82\") pod \"redhat-marketplace-btpkj\" (UID: \"c7cb44e7-8a83-4089-9808-962e06331d46\") " pod="openshift-marketplace/redhat-marketplace-btpkj" Dec 11 08:36:14 crc kubenswrapper[4860]: I1211 08:36:14.021849 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-btpkj" Dec 11 08:36:14 crc kubenswrapper[4860]: I1211 08:36:14.534929 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-btpkj"] Dec 11 08:36:15 crc kubenswrapper[4860]: I1211 08:36:15.124311 4860 generic.go:334] "Generic (PLEG): container finished" podID="c7cb44e7-8a83-4089-9808-962e06331d46" containerID="98e3df909e47c00adba3f51610d58ef66e69f86da8e440df91bc4a29f453df8c" exitCode=0 Dec 11 08:36:15 crc kubenswrapper[4860]: I1211 08:36:15.124394 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-btpkj" event={"ID":"c7cb44e7-8a83-4089-9808-962e06331d46","Type":"ContainerDied","Data":"98e3df909e47c00adba3f51610d58ef66e69f86da8e440df91bc4a29f453df8c"} Dec 11 08:36:15 crc kubenswrapper[4860]: I1211 08:36:15.124822 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-btpkj" event={"ID":"c7cb44e7-8a83-4089-9808-962e06331d46","Type":"ContainerStarted","Data":"1a6aec1d55c302ee3fe708219d9ecb3bfb0c1588ba9aaa10a4a2a7db8ddda8af"} Dec 11 08:36:19 crc kubenswrapper[4860]: I1211 08:36:19.172690 4860 generic.go:334] "Generic (PLEG): container finished" podID="c7cb44e7-8a83-4089-9808-962e06331d46" containerID="c8ff299078e66d69e63dc809bfc5bdaea7ccefc748a76ae5ad4fa979acdfd05d" exitCode=0 Dec 11 08:36:19 crc kubenswrapper[4860]: I1211 08:36:19.172762 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-btpkj" event={"ID":"c7cb44e7-8a83-4089-9808-962e06331d46","Type":"ContainerDied","Data":"c8ff299078e66d69e63dc809bfc5bdaea7ccefc748a76ae5ad4fa979acdfd05d"} Dec 11 08:36:21 crc kubenswrapper[4860]: I1211 08:36:21.196653 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-btpkj" event={"ID":"c7cb44e7-8a83-4089-9808-962e06331d46","Type":"ContainerStarted","Data":"110219f2985918352e3e0803cb4402c5c9235be6da7ac3902b42732c35af9313"} Dec 11 08:36:21 crc kubenswrapper[4860]: I1211 08:36:21.219724 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-btpkj" podStartSLOduration=3.243528483 podStartE2EDuration="8.219706014s" podCreationTimestamp="2025-12-11 08:36:13 +0000 UTC" firstStartedPulling="2025-12-11 08:36:15.127446242 +0000 UTC m=+1507.855965327" lastFinishedPulling="2025-12-11 08:36:20.103623803 +0000 UTC m=+1512.832142858" observedRunningTime="2025-12-11 08:36:21.217186304 +0000 UTC m=+1513.945705359" watchObservedRunningTime="2025-12-11 08:36:21.219706014 +0000 UTC m=+1513.948225059" Dec 11 08:36:24 crc kubenswrapper[4860]: I1211 08:36:24.022188 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-btpkj" Dec 11 08:36:24 crc kubenswrapper[4860]: I1211 08:36:24.022546 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-btpkj" Dec 11 08:36:24 crc kubenswrapper[4860]: I1211 08:36:24.077784 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-btpkj" Dec 11 08:36:30 crc kubenswrapper[4860]: I1211 08:36:30.969439 4860 scope.go:117] "RemoveContainer" containerID="fcaf6875741f0622368b408b61196853ecbfe68ddd9ec0f7cf84241df6b96926" Dec 11 08:36:31 crc kubenswrapper[4860]: I1211 08:36:31.000695 4860 scope.go:117] "RemoveContainer" containerID="0aab949e04f4b3a4703fdc0bff0fb82f6a451c3c24e3537820a25282e91ff2c8" Dec 11 08:36:31 crc kubenswrapper[4860]: I1211 08:36:31.023813 4860 scope.go:117] "RemoveContainer" containerID="5c83bda1a7e1ac8d4f912f91a01f263592a538d5209e99c9a8bf509f9d11da72" Dec 11 08:36:31 crc kubenswrapper[4860]: I1211 08:36:31.045028 4860 scope.go:117] "RemoveContainer" containerID="b2623116284268f161ab7f24e48bff42f3f1d8de1c286d66200076c14429dc28" Dec 11 08:36:31 crc kubenswrapper[4860]: I1211 08:36:31.070347 4860 scope.go:117] "RemoveContainer" containerID="fd59343d1602e30d80627c28c579dc28a98e01ed5bdc68b6006a2b053ba08373" Dec 11 08:36:31 crc kubenswrapper[4860]: I1211 08:36:31.100281 4860 scope.go:117] "RemoveContainer" containerID="e0910e8fa92a5f01a5e9fcf10245b4f567cd1f91da1a21ce559bfaf3ce103f95" Dec 11 08:36:31 crc kubenswrapper[4860]: I1211 08:36:31.125293 4860 scope.go:117] "RemoveContainer" containerID="948c71446dedd81878e5444ab6a4dbd7fbfe0ce699ae4e007ae7d21b188368ec" Dec 11 08:36:31 crc kubenswrapper[4860]: I1211 08:36:31.286960 4860 scope.go:117] "RemoveContainer" containerID="cfac87528b0888c0e951b979a1fd4da21776995c318a81737d10b48d2d6886c6" Dec 11 08:36:31 crc kubenswrapper[4860]: E1211 08:36:31.315825 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cfac87528b0888c0e951b979a1fd4da21776995c318a81737d10b48d2d6886c6\": container with ID starting with cfac87528b0888c0e951b979a1fd4da21776995c318a81737d10b48d2d6886c6 not found: ID does not exist" containerID="cfac87528b0888c0e951b979a1fd4da21776995c318a81737d10b48d2d6886c6" Dec 11 08:36:34 crc kubenswrapper[4860]: I1211 08:36:34.087932 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-btpkj" Dec 11 08:36:34 crc kubenswrapper[4860]: I1211 08:36:34.135296 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-btpkj"] Dec 11 08:36:34 crc kubenswrapper[4860]: I1211 08:36:34.348967 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-btpkj" podUID="c7cb44e7-8a83-4089-9808-962e06331d46" containerName="registry-server" containerID="cri-o://110219f2985918352e3e0803cb4402c5c9235be6da7ac3902b42732c35af9313" gracePeriod=2 Dec 11 08:36:35 crc kubenswrapper[4860]: I1211 08:36:35.311079 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-btpkj" Dec 11 08:36:35 crc kubenswrapper[4860]: I1211 08:36:35.358951 4860 generic.go:334] "Generic (PLEG): container finished" podID="c7cb44e7-8a83-4089-9808-962e06331d46" containerID="110219f2985918352e3e0803cb4402c5c9235be6da7ac3902b42732c35af9313" exitCode=0 Dec 11 08:36:35 crc kubenswrapper[4860]: I1211 08:36:35.358992 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-btpkj" event={"ID":"c7cb44e7-8a83-4089-9808-962e06331d46","Type":"ContainerDied","Data":"110219f2985918352e3e0803cb4402c5c9235be6da7ac3902b42732c35af9313"} Dec 11 08:36:35 crc kubenswrapper[4860]: I1211 08:36:35.359048 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-btpkj" event={"ID":"c7cb44e7-8a83-4089-9808-962e06331d46","Type":"ContainerDied","Data":"1a6aec1d55c302ee3fe708219d9ecb3bfb0c1588ba9aaa10a4a2a7db8ddda8af"} Dec 11 08:36:35 crc kubenswrapper[4860]: I1211 08:36:35.359052 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-btpkj" Dec 11 08:36:35 crc kubenswrapper[4860]: I1211 08:36:35.359137 4860 scope.go:117] "RemoveContainer" containerID="110219f2985918352e3e0803cb4402c5c9235be6da7ac3902b42732c35af9313" Dec 11 08:36:35 crc kubenswrapper[4860]: I1211 08:36:35.378465 4860 scope.go:117] "RemoveContainer" containerID="c8ff299078e66d69e63dc809bfc5bdaea7ccefc748a76ae5ad4fa979acdfd05d" Dec 11 08:36:35 crc kubenswrapper[4860]: I1211 08:36:35.409480 4860 scope.go:117] "RemoveContainer" containerID="98e3df909e47c00adba3f51610d58ef66e69f86da8e440df91bc4a29f453df8c" Dec 11 08:36:35 crc kubenswrapper[4860]: I1211 08:36:35.409622 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c7cb44e7-8a83-4089-9808-962e06331d46-catalog-content\") pod \"c7cb44e7-8a83-4089-9808-962e06331d46\" (UID: \"c7cb44e7-8a83-4089-9808-962e06331d46\") " Dec 11 08:36:35 crc kubenswrapper[4860]: I1211 08:36:35.409789 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s5t82\" (UniqueName: \"kubernetes.io/projected/c7cb44e7-8a83-4089-9808-962e06331d46-kube-api-access-s5t82\") pod \"c7cb44e7-8a83-4089-9808-962e06331d46\" (UID: \"c7cb44e7-8a83-4089-9808-962e06331d46\") " Dec 11 08:36:35 crc kubenswrapper[4860]: I1211 08:36:35.409972 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c7cb44e7-8a83-4089-9808-962e06331d46-utilities\") pod \"c7cb44e7-8a83-4089-9808-962e06331d46\" (UID: \"c7cb44e7-8a83-4089-9808-962e06331d46\") " Dec 11 08:36:35 crc kubenswrapper[4860]: I1211 08:36:35.411349 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c7cb44e7-8a83-4089-9808-962e06331d46-utilities" (OuterVolumeSpecName: "utilities") pod "c7cb44e7-8a83-4089-9808-962e06331d46" (UID: "c7cb44e7-8a83-4089-9808-962e06331d46"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:36:35 crc kubenswrapper[4860]: I1211 08:36:35.416670 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c7cb44e7-8a83-4089-9808-962e06331d46-kube-api-access-s5t82" (OuterVolumeSpecName: "kube-api-access-s5t82") pod "c7cb44e7-8a83-4089-9808-962e06331d46" (UID: "c7cb44e7-8a83-4089-9808-962e06331d46"). InnerVolumeSpecName "kube-api-access-s5t82". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:36:35 crc kubenswrapper[4860]: I1211 08:36:35.434417 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c7cb44e7-8a83-4089-9808-962e06331d46-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c7cb44e7-8a83-4089-9808-962e06331d46" (UID: "c7cb44e7-8a83-4089-9808-962e06331d46"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:36:35 crc kubenswrapper[4860]: I1211 08:36:35.512978 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s5t82\" (UniqueName: \"kubernetes.io/projected/c7cb44e7-8a83-4089-9808-962e06331d46-kube-api-access-s5t82\") on node \"crc\" DevicePath \"\"" Dec 11 08:36:35 crc kubenswrapper[4860]: I1211 08:36:35.513052 4860 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c7cb44e7-8a83-4089-9808-962e06331d46-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 08:36:35 crc kubenswrapper[4860]: I1211 08:36:35.513065 4860 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c7cb44e7-8a83-4089-9808-962e06331d46-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 08:36:35 crc kubenswrapper[4860]: I1211 08:36:35.521544 4860 scope.go:117] "RemoveContainer" containerID="110219f2985918352e3e0803cb4402c5c9235be6da7ac3902b42732c35af9313" Dec 11 08:36:35 crc kubenswrapper[4860]: E1211 08:36:35.524930 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"110219f2985918352e3e0803cb4402c5c9235be6da7ac3902b42732c35af9313\": container with ID starting with 110219f2985918352e3e0803cb4402c5c9235be6da7ac3902b42732c35af9313 not found: ID does not exist" containerID="110219f2985918352e3e0803cb4402c5c9235be6da7ac3902b42732c35af9313" Dec 11 08:36:35 crc kubenswrapper[4860]: I1211 08:36:35.524981 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"110219f2985918352e3e0803cb4402c5c9235be6da7ac3902b42732c35af9313"} err="failed to get container status \"110219f2985918352e3e0803cb4402c5c9235be6da7ac3902b42732c35af9313\": rpc error: code = NotFound desc = could not find container \"110219f2985918352e3e0803cb4402c5c9235be6da7ac3902b42732c35af9313\": container with ID starting with 110219f2985918352e3e0803cb4402c5c9235be6da7ac3902b42732c35af9313 not found: ID does not exist" Dec 11 08:36:35 crc kubenswrapper[4860]: I1211 08:36:35.525018 4860 scope.go:117] "RemoveContainer" containerID="c8ff299078e66d69e63dc809bfc5bdaea7ccefc748a76ae5ad4fa979acdfd05d" Dec 11 08:36:35 crc kubenswrapper[4860]: E1211 08:36:35.525700 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c8ff299078e66d69e63dc809bfc5bdaea7ccefc748a76ae5ad4fa979acdfd05d\": container with ID starting with c8ff299078e66d69e63dc809bfc5bdaea7ccefc748a76ae5ad4fa979acdfd05d not found: ID does not exist" containerID="c8ff299078e66d69e63dc809bfc5bdaea7ccefc748a76ae5ad4fa979acdfd05d" Dec 11 08:36:35 crc kubenswrapper[4860]: I1211 08:36:35.525738 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c8ff299078e66d69e63dc809bfc5bdaea7ccefc748a76ae5ad4fa979acdfd05d"} err="failed to get container status \"c8ff299078e66d69e63dc809bfc5bdaea7ccefc748a76ae5ad4fa979acdfd05d\": rpc error: code = NotFound desc = could not find container \"c8ff299078e66d69e63dc809bfc5bdaea7ccefc748a76ae5ad4fa979acdfd05d\": container with ID starting with c8ff299078e66d69e63dc809bfc5bdaea7ccefc748a76ae5ad4fa979acdfd05d not found: ID does not exist" Dec 11 08:36:35 crc kubenswrapper[4860]: I1211 08:36:35.525766 4860 scope.go:117] "RemoveContainer" containerID="98e3df909e47c00adba3f51610d58ef66e69f86da8e440df91bc4a29f453df8c" Dec 11 08:36:35 crc kubenswrapper[4860]: E1211 08:36:35.526150 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"98e3df909e47c00adba3f51610d58ef66e69f86da8e440df91bc4a29f453df8c\": container with ID starting with 98e3df909e47c00adba3f51610d58ef66e69f86da8e440df91bc4a29f453df8c not found: ID does not exist" containerID="98e3df909e47c00adba3f51610d58ef66e69f86da8e440df91bc4a29f453df8c" Dec 11 08:36:35 crc kubenswrapper[4860]: I1211 08:36:35.526189 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"98e3df909e47c00adba3f51610d58ef66e69f86da8e440df91bc4a29f453df8c"} err="failed to get container status \"98e3df909e47c00adba3f51610d58ef66e69f86da8e440df91bc4a29f453df8c\": rpc error: code = NotFound desc = could not find container \"98e3df909e47c00adba3f51610d58ef66e69f86da8e440df91bc4a29f453df8c\": container with ID starting with 98e3df909e47c00adba3f51610d58ef66e69f86da8e440df91bc4a29f453df8c not found: ID does not exist" Dec 11 08:36:35 crc kubenswrapper[4860]: I1211 08:36:35.688963 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-btpkj"] Dec 11 08:36:35 crc kubenswrapper[4860]: I1211 08:36:35.705166 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-btpkj"] Dec 11 08:36:37 crc kubenswrapper[4860]: I1211 08:36:37.590826 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c7cb44e7-8a83-4089-9808-962e06331d46" path="/var/lib/kubelet/pods/c7cb44e7-8a83-4089-9808-962e06331d46/volumes" Dec 11 08:37:16 crc kubenswrapper[4860]: I1211 08:37:16.755896 4860 generic.go:334] "Generic (PLEG): container finished" podID="aa177563-2177-4361-b45a-7ffe6f88da96" containerID="8408c746ccaa99847ae7661947c36d21ed338ca4cc372cab77281756df6b2a95" exitCode=0 Dec 11 08:37:16 crc kubenswrapper[4860]: I1211 08:37:16.755991 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-tm5lx" event={"ID":"aa177563-2177-4361-b45a-7ffe6f88da96","Type":"ContainerDied","Data":"8408c746ccaa99847ae7661947c36d21ed338ca4cc372cab77281756df6b2a95"} Dec 11 08:37:18 crc kubenswrapper[4860]: I1211 08:37:18.260339 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-tm5lx" Dec 11 08:37:18 crc kubenswrapper[4860]: I1211 08:37:18.296417 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa177563-2177-4361-b45a-7ffe6f88da96-bootstrap-combined-ca-bundle\") pod \"aa177563-2177-4361-b45a-7ffe6f88da96\" (UID: \"aa177563-2177-4361-b45a-7ffe6f88da96\") " Dec 11 08:37:18 crc kubenswrapper[4860]: I1211 08:37:18.296491 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hxx58\" (UniqueName: \"kubernetes.io/projected/aa177563-2177-4361-b45a-7ffe6f88da96-kube-api-access-hxx58\") pod \"aa177563-2177-4361-b45a-7ffe6f88da96\" (UID: \"aa177563-2177-4361-b45a-7ffe6f88da96\") " Dec 11 08:37:18 crc kubenswrapper[4860]: I1211 08:37:18.296852 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/aa177563-2177-4361-b45a-7ffe6f88da96-inventory\") pod \"aa177563-2177-4361-b45a-7ffe6f88da96\" (UID: \"aa177563-2177-4361-b45a-7ffe6f88da96\") " Dec 11 08:37:18 crc kubenswrapper[4860]: I1211 08:37:18.296881 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/aa177563-2177-4361-b45a-7ffe6f88da96-ssh-key\") pod \"aa177563-2177-4361-b45a-7ffe6f88da96\" (UID: \"aa177563-2177-4361-b45a-7ffe6f88da96\") " Dec 11 08:37:18 crc kubenswrapper[4860]: I1211 08:37:18.304411 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa177563-2177-4361-b45a-7ffe6f88da96-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "aa177563-2177-4361-b45a-7ffe6f88da96" (UID: "aa177563-2177-4361-b45a-7ffe6f88da96"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:37:18 crc kubenswrapper[4860]: I1211 08:37:18.306827 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa177563-2177-4361-b45a-7ffe6f88da96-kube-api-access-hxx58" (OuterVolumeSpecName: "kube-api-access-hxx58") pod "aa177563-2177-4361-b45a-7ffe6f88da96" (UID: "aa177563-2177-4361-b45a-7ffe6f88da96"). InnerVolumeSpecName "kube-api-access-hxx58". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:37:18 crc kubenswrapper[4860]: I1211 08:37:18.330945 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa177563-2177-4361-b45a-7ffe6f88da96-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "aa177563-2177-4361-b45a-7ffe6f88da96" (UID: "aa177563-2177-4361-b45a-7ffe6f88da96"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:37:18 crc kubenswrapper[4860]: I1211 08:37:18.349402 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa177563-2177-4361-b45a-7ffe6f88da96-inventory" (OuterVolumeSpecName: "inventory") pod "aa177563-2177-4361-b45a-7ffe6f88da96" (UID: "aa177563-2177-4361-b45a-7ffe6f88da96"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:37:18 crc kubenswrapper[4860]: I1211 08:37:18.399816 4860 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa177563-2177-4361-b45a-7ffe6f88da96-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:18 crc kubenswrapper[4860]: I1211 08:37:18.399848 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hxx58\" (UniqueName: \"kubernetes.io/projected/aa177563-2177-4361-b45a-7ffe6f88da96-kube-api-access-hxx58\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:18 crc kubenswrapper[4860]: I1211 08:37:18.399857 4860 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/aa177563-2177-4361-b45a-7ffe6f88da96-inventory\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:18 crc kubenswrapper[4860]: I1211 08:37:18.399869 4860 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/aa177563-2177-4361-b45a-7ffe6f88da96-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 11 08:37:18 crc kubenswrapper[4860]: I1211 08:37:18.784884 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-tm5lx" event={"ID":"aa177563-2177-4361-b45a-7ffe6f88da96","Type":"ContainerDied","Data":"76124c006e33ee0eaf691050c63c335bccc525e09d15bee5968f95a573981e91"} Dec 11 08:37:18 crc kubenswrapper[4860]: I1211 08:37:18.785201 4860 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="76124c006e33ee0eaf691050c63c335bccc525e09d15bee5968f95a573981e91" Dec 11 08:37:18 crc kubenswrapper[4860]: I1211 08:37:18.785269 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-tm5lx" Dec 11 08:37:18 crc kubenswrapper[4860]: I1211 08:37:18.899178 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-zvzpf"] Dec 11 08:37:18 crc kubenswrapper[4860]: E1211 08:37:18.899631 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa177563-2177-4361-b45a-7ffe6f88da96" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 11 08:37:18 crc kubenswrapper[4860]: I1211 08:37:18.899665 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa177563-2177-4361-b45a-7ffe6f88da96" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 11 08:37:18 crc kubenswrapper[4860]: E1211 08:37:18.899687 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7cb44e7-8a83-4089-9808-962e06331d46" containerName="extract-content" Dec 11 08:37:18 crc kubenswrapper[4860]: I1211 08:37:18.899694 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7cb44e7-8a83-4089-9808-962e06331d46" containerName="extract-content" Dec 11 08:37:18 crc kubenswrapper[4860]: E1211 08:37:18.899720 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7cb44e7-8a83-4089-9808-962e06331d46" containerName="registry-server" Dec 11 08:37:18 crc kubenswrapper[4860]: I1211 08:37:18.899726 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7cb44e7-8a83-4089-9808-962e06331d46" containerName="registry-server" Dec 11 08:37:18 crc kubenswrapper[4860]: E1211 08:37:18.899754 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7cb44e7-8a83-4089-9808-962e06331d46" containerName="extract-utilities" Dec 11 08:37:18 crc kubenswrapper[4860]: I1211 08:37:18.899763 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7cb44e7-8a83-4089-9808-962e06331d46" containerName="extract-utilities" Dec 11 08:37:18 crc kubenswrapper[4860]: I1211 08:37:18.899967 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="c7cb44e7-8a83-4089-9808-962e06331d46" containerName="registry-server" Dec 11 08:37:18 crc kubenswrapper[4860]: I1211 08:37:18.899991 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa177563-2177-4361-b45a-7ffe6f88da96" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 11 08:37:18 crc kubenswrapper[4860]: I1211 08:37:18.900807 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-zvzpf" Dec 11 08:37:18 crc kubenswrapper[4860]: I1211 08:37:18.906172 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 11 08:37:18 crc kubenswrapper[4860]: I1211 08:37:18.906765 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 11 08:37:18 crc kubenswrapper[4860]: I1211 08:37:18.907119 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-7w7kh" Dec 11 08:37:18 crc kubenswrapper[4860]: I1211 08:37:18.907167 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 11 08:37:18 crc kubenswrapper[4860]: I1211 08:37:18.908796 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1fb47321-3b60-4961-abf2-87ee1d57a58a-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-zvzpf\" (UID: \"1fb47321-3b60-4961-abf2-87ee1d57a58a\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-zvzpf" Dec 11 08:37:18 crc kubenswrapper[4860]: I1211 08:37:18.908888 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9lnhm\" (UniqueName: \"kubernetes.io/projected/1fb47321-3b60-4961-abf2-87ee1d57a58a-kube-api-access-9lnhm\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-zvzpf\" (UID: \"1fb47321-3b60-4961-abf2-87ee1d57a58a\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-zvzpf" Dec 11 08:37:18 crc kubenswrapper[4860]: I1211 08:37:18.908937 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1fb47321-3b60-4961-abf2-87ee1d57a58a-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-zvzpf\" (UID: \"1fb47321-3b60-4961-abf2-87ee1d57a58a\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-zvzpf" Dec 11 08:37:18 crc kubenswrapper[4860]: I1211 08:37:18.922564 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-zvzpf"] Dec 11 08:37:19 crc kubenswrapper[4860]: I1211 08:37:19.011228 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1fb47321-3b60-4961-abf2-87ee1d57a58a-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-zvzpf\" (UID: \"1fb47321-3b60-4961-abf2-87ee1d57a58a\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-zvzpf" Dec 11 08:37:19 crc kubenswrapper[4860]: I1211 08:37:19.011448 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1fb47321-3b60-4961-abf2-87ee1d57a58a-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-zvzpf\" (UID: \"1fb47321-3b60-4961-abf2-87ee1d57a58a\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-zvzpf" Dec 11 08:37:19 crc kubenswrapper[4860]: I1211 08:37:19.011563 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9lnhm\" (UniqueName: \"kubernetes.io/projected/1fb47321-3b60-4961-abf2-87ee1d57a58a-kube-api-access-9lnhm\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-zvzpf\" (UID: \"1fb47321-3b60-4961-abf2-87ee1d57a58a\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-zvzpf" Dec 11 08:37:19 crc kubenswrapper[4860]: I1211 08:37:19.027295 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1fb47321-3b60-4961-abf2-87ee1d57a58a-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-zvzpf\" (UID: \"1fb47321-3b60-4961-abf2-87ee1d57a58a\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-zvzpf" Dec 11 08:37:19 crc kubenswrapper[4860]: I1211 08:37:19.027405 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1fb47321-3b60-4961-abf2-87ee1d57a58a-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-zvzpf\" (UID: \"1fb47321-3b60-4961-abf2-87ee1d57a58a\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-zvzpf" Dec 11 08:37:19 crc kubenswrapper[4860]: I1211 08:37:19.041868 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9lnhm\" (UniqueName: \"kubernetes.io/projected/1fb47321-3b60-4961-abf2-87ee1d57a58a-kube-api-access-9lnhm\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-zvzpf\" (UID: \"1fb47321-3b60-4961-abf2-87ee1d57a58a\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-zvzpf" Dec 11 08:37:19 crc kubenswrapper[4860]: I1211 08:37:19.222561 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-zvzpf" Dec 11 08:37:19 crc kubenswrapper[4860]: I1211 08:37:19.791937 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-zvzpf"] Dec 11 08:37:20 crc kubenswrapper[4860]: I1211 08:37:20.809079 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-zvzpf" event={"ID":"1fb47321-3b60-4961-abf2-87ee1d57a58a","Type":"ContainerStarted","Data":"7f24df562e3ab9e437f7e743c6332c5d629f7a97ae813bf268e24bec5ff7f7f9"} Dec 11 08:37:20 crc kubenswrapper[4860]: I1211 08:37:20.809409 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-zvzpf" event={"ID":"1fb47321-3b60-4961-abf2-87ee1d57a58a","Type":"ContainerStarted","Data":"192ca75208fce9b3e7c211f1e57dc6b3dae105916e0ca8c8823a4bcda95c826d"} Dec 11 08:37:20 crc kubenswrapper[4860]: I1211 08:37:20.836456 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-zvzpf" podStartSLOduration=2.229824505 podStartE2EDuration="2.836428973s" podCreationTimestamp="2025-12-11 08:37:18 +0000 UTC" firstStartedPulling="2025-12-11 08:37:19.801774599 +0000 UTC m=+1572.530293654" lastFinishedPulling="2025-12-11 08:37:20.408379067 +0000 UTC m=+1573.136898122" observedRunningTime="2025-12-11 08:37:20.827028667 +0000 UTC m=+1573.555547732" watchObservedRunningTime="2025-12-11 08:37:20.836428973 +0000 UTC m=+1573.564948038" Dec 11 08:37:31 crc kubenswrapper[4860]: I1211 08:37:31.401072 4860 scope.go:117] "RemoveContainer" containerID="b72ee3f3e1ae8d023bd56cc1da250ac39f56b858ce74442ccc4f0e77c28e7c20" Dec 11 08:37:31 crc kubenswrapper[4860]: I1211 08:37:31.439136 4860 scope.go:117] "RemoveContainer" containerID="b291a77f964230ad3604e0f8070421f9d71a6daf17b5e1350871337ab4bd9f19" Dec 11 08:37:31 crc kubenswrapper[4860]: I1211 08:37:31.478889 4860 scope.go:117] "RemoveContainer" containerID="ffea859d11da0e998a8570da5a7edb29c1c59d7fd1d23e319c83bfbdfc654520" Dec 11 08:37:31 crc kubenswrapper[4860]: I1211 08:37:31.501532 4860 scope.go:117] "RemoveContainer" containerID="14f6f9a01ac6e9725c9b1bade90deaab2ab59447d0821c52bbfed989797caa75" Dec 11 08:38:08 crc kubenswrapper[4860]: I1211 08:38:08.794967 4860 patch_prober.go:28] interesting pod/machine-config-daemon-99qgp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 08:38:08 crc kubenswrapper[4860]: I1211 08:38:08.795574 4860 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 08:38:23 crc kubenswrapper[4860]: I1211 08:38:23.044277 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-f52d-account-create-update-8rklk"] Dec 11 08:38:23 crc kubenswrapper[4860]: I1211 08:38:23.058427 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-66fgz"] Dec 11 08:38:23 crc kubenswrapper[4860]: I1211 08:38:23.068131 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-d5e8-account-create-update-76hlk"] Dec 11 08:38:23 crc kubenswrapper[4860]: I1211 08:38:23.076416 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-f52d-account-create-update-8rklk"] Dec 11 08:38:23 crc kubenswrapper[4860]: I1211 08:38:23.084917 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-d5e8-account-create-update-76hlk"] Dec 11 08:38:23 crc kubenswrapper[4860]: I1211 08:38:23.092708 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-66fgz"] Dec 11 08:38:23 crc kubenswrapper[4860]: I1211 08:38:23.590699 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25bead29-f8d6-4539-97f8-9c3be58fbf64" path="/var/lib/kubelet/pods/25bead29-f8d6-4539-97f8-9c3be58fbf64/volumes" Dec 11 08:38:23 crc kubenswrapper[4860]: I1211 08:38:23.591404 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="35b0da4d-26cd-4018-877b-5e8cbe0d4176" path="/var/lib/kubelet/pods/35b0da4d-26cd-4018-877b-5e8cbe0d4176/volumes" Dec 11 08:38:23 crc kubenswrapper[4860]: I1211 08:38:23.592075 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="79aae41e-c682-42c8-8bbc-5d060804d535" path="/var/lib/kubelet/pods/79aae41e-c682-42c8-8bbc-5d060804d535/volumes" Dec 11 08:38:24 crc kubenswrapper[4860]: I1211 08:38:24.027717 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-x7qkx"] Dec 11 08:38:24 crc kubenswrapper[4860]: I1211 08:38:24.036848 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-84lgh"] Dec 11 08:38:24 crc kubenswrapper[4860]: I1211 08:38:24.045797 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-x7qkx"] Dec 11 08:38:24 crc kubenswrapper[4860]: I1211 08:38:24.054099 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-84lgh"] Dec 11 08:38:25 crc kubenswrapper[4860]: I1211 08:38:25.041992 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-86d4-account-create-update-khw2q"] Dec 11 08:38:25 crc kubenswrapper[4860]: I1211 08:38:25.051099 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-86d4-account-create-update-khw2q"] Dec 11 08:38:25 crc kubenswrapper[4860]: I1211 08:38:25.591644 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="60da3a16-5fba-47fc-8aef-7c7e942fa69b" path="/var/lib/kubelet/pods/60da3a16-5fba-47fc-8aef-7c7e942fa69b/volumes" Dec 11 08:38:25 crc kubenswrapper[4860]: I1211 08:38:25.592739 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d449e0eb-9165-4f6b-b848-c0c6ea8bfa35" path="/var/lib/kubelet/pods/d449e0eb-9165-4f6b-b848-c0c6ea8bfa35/volumes" Dec 11 08:38:25 crc kubenswrapper[4860]: I1211 08:38:25.593264 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f765a5cb-6b7f-40ce-95fe-8c1abf5ce4d5" path="/var/lib/kubelet/pods/f765a5cb-6b7f-40ce-95fe-8c1abf5ce4d5/volumes" Dec 11 08:38:31 crc kubenswrapper[4860]: I1211 08:38:31.592891 4860 scope.go:117] "RemoveContainer" containerID="f1438a46a2a9690d43847281c607ef6c47f38f8e05dfbaabe34adf6a253bfafe" Dec 11 08:38:31 crc kubenswrapper[4860]: I1211 08:38:31.644521 4860 scope.go:117] "RemoveContainer" containerID="44765f35f50e19467b780e5977831f691d7c724b607c8daec431551703717522" Dec 11 08:38:31 crc kubenswrapper[4860]: I1211 08:38:31.676189 4860 scope.go:117] "RemoveContainer" containerID="1aa442d3e55c72e0c256ef664cf020cd34cd97189a1619b6819b4ae1a5696ea7" Dec 11 08:38:31 crc kubenswrapper[4860]: I1211 08:38:31.721310 4860 scope.go:117] "RemoveContainer" containerID="a1dffde6d5b34e4277a3235833c9cfa23befa0d0541a1ecc3976f3b8de6a843f" Dec 11 08:38:31 crc kubenswrapper[4860]: I1211 08:38:31.777603 4860 scope.go:117] "RemoveContainer" containerID="deafcd4e90a9c8eb4c53882a697b7855a4976e819e86959545ccf4411f6e5a3a" Dec 11 08:38:31 crc kubenswrapper[4860]: I1211 08:38:31.837508 4860 scope.go:117] "RemoveContainer" containerID="0f40eceff896f6453086ec0213e1ebfaa19e60a191b1173daa71b9121db12111" Dec 11 08:38:31 crc kubenswrapper[4860]: I1211 08:38:31.884398 4860 scope.go:117] "RemoveContainer" containerID="c54a5f6501d64c77e977792e254bfbe24fadb74783de561e2d747ad3ad7140dd" Dec 11 08:38:31 crc kubenswrapper[4860]: I1211 08:38:31.903746 4860 scope.go:117] "RemoveContainer" containerID="739ec4e15f43896e65eeb1f0e6bbf9b0b64946cd2a3a5547c6a8c39b795bf61d" Dec 11 08:38:38 crc kubenswrapper[4860]: I1211 08:38:38.795674 4860 patch_prober.go:28] interesting pod/machine-config-daemon-99qgp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 08:38:38 crc kubenswrapper[4860]: I1211 08:38:38.796249 4860 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 08:38:48 crc kubenswrapper[4860]: I1211 08:38:48.037124 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-7567-account-create-update-5mzbx"] Dec 11 08:38:48 crc kubenswrapper[4860]: I1211 08:38:48.047902 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-dqvkz"] Dec 11 08:38:48 crc kubenswrapper[4860]: I1211 08:38:48.057672 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-c4mc4"] Dec 11 08:38:48 crc kubenswrapper[4860]: I1211 08:38:48.064827 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-970d-account-create-update-65w8k"] Dec 11 08:38:48 crc kubenswrapper[4860]: I1211 08:38:48.071951 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-7567-account-create-update-5mzbx"] Dec 11 08:38:48 crc kubenswrapper[4860]: I1211 08:38:48.078470 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-dqvkz"] Dec 11 08:38:48 crc kubenswrapper[4860]: I1211 08:38:48.085119 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-970d-account-create-update-65w8k"] Dec 11 08:38:48 crc kubenswrapper[4860]: I1211 08:38:48.091569 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-c4mc4"] Dec 11 08:38:49 crc kubenswrapper[4860]: I1211 08:38:49.590435 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="07fbfcf2-916b-4cd8-b603-48c55736a6b0" path="/var/lib/kubelet/pods/07fbfcf2-916b-4cd8-b603-48c55736a6b0/volumes" Dec 11 08:38:49 crc kubenswrapper[4860]: I1211 08:38:49.591590 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="54f31848-ee29-43be-8363-0242dbe36164" path="/var/lib/kubelet/pods/54f31848-ee29-43be-8363-0242dbe36164/volumes" Dec 11 08:38:49 crc kubenswrapper[4860]: I1211 08:38:49.592426 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5af63911-2f28-40ca-9d02-721f5871f52d" path="/var/lib/kubelet/pods/5af63911-2f28-40ca-9d02-721f5871f52d/volumes" Dec 11 08:38:49 crc kubenswrapper[4860]: I1211 08:38:49.593281 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bea23d9-3be0-4a68-8ec1-8f21bd7feb80" path="/var/lib/kubelet/pods/7bea23d9-3be0-4a68-8ec1-8f21bd7feb80/volumes" Dec 11 08:38:53 crc kubenswrapper[4860]: I1211 08:38:53.066882 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-6822-account-create-update-tjqd7"] Dec 11 08:38:53 crc kubenswrapper[4860]: I1211 08:38:53.082807 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-6822-account-create-update-tjqd7"] Dec 11 08:38:53 crc kubenswrapper[4860]: I1211 08:38:53.093498 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-lrszm"] Dec 11 08:38:53 crc kubenswrapper[4860]: I1211 08:38:53.107149 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-lrszm"] Dec 11 08:38:53 crc kubenswrapper[4860]: I1211 08:38:53.591067 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="989e314b-f91d-4ceb-ba16-79d180225afe" path="/var/lib/kubelet/pods/989e314b-f91d-4ceb-ba16-79d180225afe/volumes" Dec 11 08:38:53 crc kubenswrapper[4860]: I1211 08:38:53.591721 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e8941d7b-8d06-4ec6-8da3-101e001f824f" path="/var/lib/kubelet/pods/e8941d7b-8d06-4ec6-8da3-101e001f824f/volumes" Dec 11 08:38:56 crc kubenswrapper[4860]: I1211 08:38:56.764258 4860 generic.go:334] "Generic (PLEG): container finished" podID="1fb47321-3b60-4961-abf2-87ee1d57a58a" containerID="7f24df562e3ab9e437f7e743c6332c5d629f7a97ae813bf268e24bec5ff7f7f9" exitCode=0 Dec 11 08:38:56 crc kubenswrapper[4860]: I1211 08:38:56.764333 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-zvzpf" event={"ID":"1fb47321-3b60-4961-abf2-87ee1d57a58a","Type":"ContainerDied","Data":"7f24df562e3ab9e437f7e743c6332c5d629f7a97ae813bf268e24bec5ff7f7f9"} Dec 11 08:38:57 crc kubenswrapper[4860]: I1211 08:38:57.035968 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-wkrv5"] Dec 11 08:38:57 crc kubenswrapper[4860]: I1211 08:38:57.056147 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-wkrv5"] Dec 11 08:38:57 crc kubenswrapper[4860]: I1211 08:38:57.598612 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c602556f-fbc0-4e66-bd77-0d5b029e81bb" path="/var/lib/kubelet/pods/c602556f-fbc0-4e66-bd77-0d5b029e81bb/volumes" Dec 11 08:38:58 crc kubenswrapper[4860]: I1211 08:38:58.301121 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-zvzpf" Dec 11 08:38:58 crc kubenswrapper[4860]: I1211 08:38:58.442481 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9lnhm\" (UniqueName: \"kubernetes.io/projected/1fb47321-3b60-4961-abf2-87ee1d57a58a-kube-api-access-9lnhm\") pod \"1fb47321-3b60-4961-abf2-87ee1d57a58a\" (UID: \"1fb47321-3b60-4961-abf2-87ee1d57a58a\") " Dec 11 08:38:58 crc kubenswrapper[4860]: I1211 08:38:58.442608 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1fb47321-3b60-4961-abf2-87ee1d57a58a-ssh-key\") pod \"1fb47321-3b60-4961-abf2-87ee1d57a58a\" (UID: \"1fb47321-3b60-4961-abf2-87ee1d57a58a\") " Dec 11 08:38:58 crc kubenswrapper[4860]: I1211 08:38:58.442817 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1fb47321-3b60-4961-abf2-87ee1d57a58a-inventory\") pod \"1fb47321-3b60-4961-abf2-87ee1d57a58a\" (UID: \"1fb47321-3b60-4961-abf2-87ee1d57a58a\") " Dec 11 08:38:58 crc kubenswrapper[4860]: I1211 08:38:58.450267 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1fb47321-3b60-4961-abf2-87ee1d57a58a-kube-api-access-9lnhm" (OuterVolumeSpecName: "kube-api-access-9lnhm") pod "1fb47321-3b60-4961-abf2-87ee1d57a58a" (UID: "1fb47321-3b60-4961-abf2-87ee1d57a58a"). InnerVolumeSpecName "kube-api-access-9lnhm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:38:58 crc kubenswrapper[4860]: I1211 08:38:58.480461 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1fb47321-3b60-4961-abf2-87ee1d57a58a-inventory" (OuterVolumeSpecName: "inventory") pod "1fb47321-3b60-4961-abf2-87ee1d57a58a" (UID: "1fb47321-3b60-4961-abf2-87ee1d57a58a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:38:58 crc kubenswrapper[4860]: I1211 08:38:58.481315 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1fb47321-3b60-4961-abf2-87ee1d57a58a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "1fb47321-3b60-4961-abf2-87ee1d57a58a" (UID: "1fb47321-3b60-4961-abf2-87ee1d57a58a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:38:58 crc kubenswrapper[4860]: I1211 08:38:58.545127 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9lnhm\" (UniqueName: \"kubernetes.io/projected/1fb47321-3b60-4961-abf2-87ee1d57a58a-kube-api-access-9lnhm\") on node \"crc\" DevicePath \"\"" Dec 11 08:38:58 crc kubenswrapper[4860]: I1211 08:38:58.545164 4860 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1fb47321-3b60-4961-abf2-87ee1d57a58a-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 11 08:38:58 crc kubenswrapper[4860]: I1211 08:38:58.545176 4860 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1fb47321-3b60-4961-abf2-87ee1d57a58a-inventory\") on node \"crc\" DevicePath \"\"" Dec 11 08:38:58 crc kubenswrapper[4860]: I1211 08:38:58.783678 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-zvzpf" event={"ID":"1fb47321-3b60-4961-abf2-87ee1d57a58a","Type":"ContainerDied","Data":"192ca75208fce9b3e7c211f1e57dc6b3dae105916e0ca8c8823a4bcda95c826d"} Dec 11 08:38:58 crc kubenswrapper[4860]: I1211 08:38:58.783759 4860 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="192ca75208fce9b3e7c211f1e57dc6b3dae105916e0ca8c8823a4bcda95c826d" Dec 11 08:38:58 crc kubenswrapper[4860]: I1211 08:38:58.783759 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-zvzpf" Dec 11 08:38:58 crc kubenswrapper[4860]: I1211 08:38:58.861553 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-7zg5t"] Dec 11 08:38:58 crc kubenswrapper[4860]: E1211 08:38:58.862085 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1fb47321-3b60-4961-abf2-87ee1d57a58a" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 11 08:38:58 crc kubenswrapper[4860]: I1211 08:38:58.862111 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="1fb47321-3b60-4961-abf2-87ee1d57a58a" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 11 08:38:58 crc kubenswrapper[4860]: I1211 08:38:58.862378 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="1fb47321-3b60-4961-abf2-87ee1d57a58a" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 11 08:38:58 crc kubenswrapper[4860]: I1211 08:38:58.864314 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-7zg5t" Dec 11 08:38:58 crc kubenswrapper[4860]: I1211 08:38:58.866708 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 11 08:38:58 crc kubenswrapper[4860]: I1211 08:38:58.866712 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 11 08:38:58 crc kubenswrapper[4860]: I1211 08:38:58.867405 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 11 08:38:58 crc kubenswrapper[4860]: I1211 08:38:58.867402 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-7w7kh" Dec 11 08:38:58 crc kubenswrapper[4860]: I1211 08:38:58.872191 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-7zg5t"] Dec 11 08:38:59 crc kubenswrapper[4860]: I1211 08:38:59.057868 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/155181af-f5ab-41fb-9c0f-820c03b91b23-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-7zg5t\" (UID: \"155181af-f5ab-41fb-9c0f-820c03b91b23\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-7zg5t" Dec 11 08:38:59 crc kubenswrapper[4860]: I1211 08:38:59.058016 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5f6tj\" (UniqueName: \"kubernetes.io/projected/155181af-f5ab-41fb-9c0f-820c03b91b23-kube-api-access-5f6tj\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-7zg5t\" (UID: \"155181af-f5ab-41fb-9c0f-820c03b91b23\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-7zg5t" Dec 11 08:38:59 crc kubenswrapper[4860]: I1211 08:38:59.058200 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/155181af-f5ab-41fb-9c0f-820c03b91b23-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-7zg5t\" (UID: \"155181af-f5ab-41fb-9c0f-820c03b91b23\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-7zg5t" Dec 11 08:38:59 crc kubenswrapper[4860]: I1211 08:38:59.159882 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5f6tj\" (UniqueName: \"kubernetes.io/projected/155181af-f5ab-41fb-9c0f-820c03b91b23-kube-api-access-5f6tj\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-7zg5t\" (UID: \"155181af-f5ab-41fb-9c0f-820c03b91b23\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-7zg5t" Dec 11 08:38:59 crc kubenswrapper[4860]: I1211 08:38:59.160229 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/155181af-f5ab-41fb-9c0f-820c03b91b23-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-7zg5t\" (UID: \"155181af-f5ab-41fb-9c0f-820c03b91b23\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-7zg5t" Dec 11 08:38:59 crc kubenswrapper[4860]: I1211 08:38:59.160403 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/155181af-f5ab-41fb-9c0f-820c03b91b23-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-7zg5t\" (UID: \"155181af-f5ab-41fb-9c0f-820c03b91b23\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-7zg5t" Dec 11 08:38:59 crc kubenswrapper[4860]: I1211 08:38:59.165497 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/155181af-f5ab-41fb-9c0f-820c03b91b23-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-7zg5t\" (UID: \"155181af-f5ab-41fb-9c0f-820c03b91b23\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-7zg5t" Dec 11 08:38:59 crc kubenswrapper[4860]: I1211 08:38:59.166699 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/155181af-f5ab-41fb-9c0f-820c03b91b23-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-7zg5t\" (UID: \"155181af-f5ab-41fb-9c0f-820c03b91b23\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-7zg5t" Dec 11 08:38:59 crc kubenswrapper[4860]: I1211 08:38:59.176508 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5f6tj\" (UniqueName: \"kubernetes.io/projected/155181af-f5ab-41fb-9c0f-820c03b91b23-kube-api-access-5f6tj\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-7zg5t\" (UID: \"155181af-f5ab-41fb-9c0f-820c03b91b23\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-7zg5t" Dec 11 08:38:59 crc kubenswrapper[4860]: I1211 08:38:59.182528 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-7zg5t" Dec 11 08:38:59 crc kubenswrapper[4860]: I1211 08:38:59.764215 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-7zg5t"] Dec 11 08:38:59 crc kubenswrapper[4860]: W1211 08:38:59.765785 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod155181af_f5ab_41fb_9c0f_820c03b91b23.slice/crio-2656961f7fd1879a7c668b128917096b708da2db5af1cdde63710d8d770bb030 WatchSource:0}: Error finding container 2656961f7fd1879a7c668b128917096b708da2db5af1cdde63710d8d770bb030: Status 404 returned error can't find the container with id 2656961f7fd1879a7c668b128917096b708da2db5af1cdde63710d8d770bb030 Dec 11 08:38:59 crc kubenswrapper[4860]: I1211 08:38:59.769137 4860 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 11 08:38:59 crc kubenswrapper[4860]: I1211 08:38:59.794857 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-7zg5t" event={"ID":"155181af-f5ab-41fb-9c0f-820c03b91b23","Type":"ContainerStarted","Data":"2656961f7fd1879a7c668b128917096b708da2db5af1cdde63710d8d770bb030"} Dec 11 08:39:00 crc kubenswrapper[4860]: I1211 08:39:00.806480 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-7zg5t" event={"ID":"155181af-f5ab-41fb-9c0f-820c03b91b23","Type":"ContainerStarted","Data":"6154af5b0844abcaccc8d61bd88937426587c37ab32d347369bdbfd6bc3e2894"} Dec 11 08:39:00 crc kubenswrapper[4860]: I1211 08:39:00.823574 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-7zg5t" podStartSLOduration=2.142866298 podStartE2EDuration="2.823555367s" podCreationTimestamp="2025-12-11 08:38:58 +0000 UTC" firstStartedPulling="2025-12-11 08:38:59.768917871 +0000 UTC m=+1672.497436936" lastFinishedPulling="2025-12-11 08:39:00.44960695 +0000 UTC m=+1673.178126005" observedRunningTime="2025-12-11 08:39:00.821283565 +0000 UTC m=+1673.549802620" watchObservedRunningTime="2025-12-11 08:39:00.823555367 +0000 UTC m=+1673.552074422" Dec 11 08:39:08 crc kubenswrapper[4860]: I1211 08:39:08.795042 4860 patch_prober.go:28] interesting pod/machine-config-daemon-99qgp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 08:39:08 crc kubenswrapper[4860]: I1211 08:39:08.795578 4860 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 08:39:08 crc kubenswrapper[4860]: I1211 08:39:08.795629 4860 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" Dec 11 08:39:08 crc kubenswrapper[4860]: I1211 08:39:08.796393 4860 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"af76095360a11870b804fb7f16c40dad604dbf93152022249293687b1519ca1b"} pod="openshift-machine-config-operator/machine-config-daemon-99qgp" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 11 08:39:08 crc kubenswrapper[4860]: I1211 08:39:08.796437 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" containerName="machine-config-daemon" containerID="cri-o://af76095360a11870b804fb7f16c40dad604dbf93152022249293687b1519ca1b" gracePeriod=600 Dec 11 08:39:08 crc kubenswrapper[4860]: E1211 08:39:08.916567 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-99qgp_openshift-machine-config-operator(31c30642-6e60-41b4-a477-0d802424e0aa)\"" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" Dec 11 08:39:09 crc kubenswrapper[4860]: I1211 08:39:09.892226 4860 generic.go:334] "Generic (PLEG): container finished" podID="31c30642-6e60-41b4-a477-0d802424e0aa" containerID="af76095360a11870b804fb7f16c40dad604dbf93152022249293687b1519ca1b" exitCode=0 Dec 11 08:39:09 crc kubenswrapper[4860]: I1211 08:39:09.892283 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" event={"ID":"31c30642-6e60-41b4-a477-0d802424e0aa","Type":"ContainerDied","Data":"af76095360a11870b804fb7f16c40dad604dbf93152022249293687b1519ca1b"} Dec 11 08:39:09 crc kubenswrapper[4860]: I1211 08:39:09.892792 4860 scope.go:117] "RemoveContainer" containerID="2c5567d203277d4236f1f4cd1d73721ecd8809240167b07eac4086db3323fa9f" Dec 11 08:39:09 crc kubenswrapper[4860]: I1211 08:39:09.893540 4860 scope.go:117] "RemoveContainer" containerID="af76095360a11870b804fb7f16c40dad604dbf93152022249293687b1519ca1b" Dec 11 08:39:09 crc kubenswrapper[4860]: E1211 08:39:09.894076 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-99qgp_openshift-machine-config-operator(31c30642-6e60-41b4-a477-0d802424e0aa)\"" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" Dec 11 08:39:24 crc kubenswrapper[4860]: I1211 08:39:24.579702 4860 scope.go:117] "RemoveContainer" containerID="af76095360a11870b804fb7f16c40dad604dbf93152022249293687b1519ca1b" Dec 11 08:39:24 crc kubenswrapper[4860]: E1211 08:39:24.580802 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-99qgp_openshift-machine-config-operator(31c30642-6e60-41b4-a477-0d802424e0aa)\"" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" Dec 11 08:39:28 crc kubenswrapper[4860]: I1211 08:39:28.042947 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-4fsqr"] Dec 11 08:39:28 crc kubenswrapper[4860]: I1211 08:39:28.053227 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-4fsqr"] Dec 11 08:39:29 crc kubenswrapper[4860]: I1211 08:39:29.590861 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cf433d5d-8c95-4d72-864c-3b74f315dd2b" path="/var/lib/kubelet/pods/cf433d5d-8c95-4d72-864c-3b74f315dd2b/volumes" Dec 11 08:39:31 crc kubenswrapper[4860]: I1211 08:39:31.042110 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-wgk5t"] Dec 11 08:39:31 crc kubenswrapper[4860]: I1211 08:39:31.056297 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-wgk5t"] Dec 11 08:39:31 crc kubenswrapper[4860]: I1211 08:39:31.591788 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e3822ab5-d1a7-46da-9533-21b67bd240eb" path="/var/lib/kubelet/pods/e3822ab5-d1a7-46da-9533-21b67bd240eb/volumes" Dec 11 08:39:32 crc kubenswrapper[4860]: I1211 08:39:32.038685 4860 scope.go:117] "RemoveContainer" containerID="d737396b95b4dbb30ba8c85e823e3e330e74286479aa940ead93372b72ec06f3" Dec 11 08:39:32 crc kubenswrapper[4860]: I1211 08:39:32.078891 4860 scope.go:117] "RemoveContainer" containerID="cf5a7e627cf219a9c4504179ba2642a4118fe67030941daa18f9ca7063564d27" Dec 11 08:39:32 crc kubenswrapper[4860]: I1211 08:39:32.114151 4860 scope.go:117] "RemoveContainer" containerID="0b11a85c964986dfd28dd78240fb9ced04ee192ff15828c7ed754be53dcb5116" Dec 11 08:39:32 crc kubenswrapper[4860]: I1211 08:39:32.144485 4860 scope.go:117] "RemoveContainer" containerID="1d267bbbb0fe1655be49f1107ace8ca93c9627c3c5d44ca882c0d994a7f8e050" Dec 11 08:39:32 crc kubenswrapper[4860]: I1211 08:39:32.185233 4860 scope.go:117] "RemoveContainer" containerID="452366f5a3dcc497c0b8da7d4128cab636c136a87b2ae2638147a11767c43545" Dec 11 08:39:32 crc kubenswrapper[4860]: I1211 08:39:32.233613 4860 scope.go:117] "RemoveContainer" containerID="732cd87f4f33757f7f25275c3c1bafe341bdbdc36956a4339205959a3b10edbd" Dec 11 08:39:32 crc kubenswrapper[4860]: I1211 08:39:32.280720 4860 scope.go:117] "RemoveContainer" containerID="da10db9abaea4b7eac10a44bfd1ba1fec17869d0afab3a371521004f840f6199" Dec 11 08:39:32 crc kubenswrapper[4860]: I1211 08:39:32.304358 4860 scope.go:117] "RemoveContainer" containerID="650883b438680b9c1ab516bfe7c40142cef24b094b49f32d2439af5c7e7508d6" Dec 11 08:39:32 crc kubenswrapper[4860]: I1211 08:39:32.348209 4860 scope.go:117] "RemoveContainer" containerID="e5a0736b99cd2f37cf621cf0decadfdb56873bb6ff930b8c31cd4a8e299cb90f" Dec 11 08:39:36 crc kubenswrapper[4860]: I1211 08:39:36.032570 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-jqgbc"] Dec 11 08:39:36 crc kubenswrapper[4860]: I1211 08:39:36.047281 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-jqgbc"] Dec 11 08:39:37 crc kubenswrapper[4860]: I1211 08:39:37.039037 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-pfz4t"] Dec 11 08:39:37 crc kubenswrapper[4860]: I1211 08:39:37.052181 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-pfz4t"] Dec 11 08:39:37 crc kubenswrapper[4860]: I1211 08:39:37.591843 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4ff61ef1-6a67-4a8e-80e2-11f0154e54c6" path="/var/lib/kubelet/pods/4ff61ef1-6a67-4a8e-80e2-11f0154e54c6/volumes" Dec 11 08:39:37 crc kubenswrapper[4860]: I1211 08:39:37.592760 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eafaa4e1-5187-4406-b55a-9d9349e3a744" path="/var/lib/kubelet/pods/eafaa4e1-5187-4406-b55a-9d9349e3a744/volumes" Dec 11 08:39:38 crc kubenswrapper[4860]: I1211 08:39:38.028860 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-plxxc"] Dec 11 08:39:38 crc kubenswrapper[4860]: I1211 08:39:38.038533 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-plxxc"] Dec 11 08:39:39 crc kubenswrapper[4860]: I1211 08:39:39.578842 4860 scope.go:117] "RemoveContainer" containerID="af76095360a11870b804fb7f16c40dad604dbf93152022249293687b1519ca1b" Dec 11 08:39:39 crc kubenswrapper[4860]: E1211 08:39:39.579384 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-99qgp_openshift-machine-config-operator(31c30642-6e60-41b4-a477-0d802424e0aa)\"" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" Dec 11 08:39:39 crc kubenswrapper[4860]: I1211 08:39:39.591486 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fecaee75-2dc0-4d48-9d40-11fc1ec8e972" path="/var/lib/kubelet/pods/fecaee75-2dc0-4d48-9d40-11fc1ec8e972/volumes" Dec 11 08:39:50 crc kubenswrapper[4860]: I1211 08:39:50.579093 4860 scope.go:117] "RemoveContainer" containerID="af76095360a11870b804fb7f16c40dad604dbf93152022249293687b1519ca1b" Dec 11 08:39:50 crc kubenswrapper[4860]: E1211 08:39:50.580112 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-99qgp_openshift-machine-config-operator(31c30642-6e60-41b4-a477-0d802424e0aa)\"" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" Dec 11 08:39:56 crc kubenswrapper[4860]: I1211 08:39:56.050859 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-hccts"] Dec 11 08:39:56 crc kubenswrapper[4860]: I1211 08:39:56.061393 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-hccts"] Dec 11 08:39:57 crc kubenswrapper[4860]: I1211 08:39:57.595324 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d472d27e-9bc9-4d99-8fd9-697f7f44f57d" path="/var/lib/kubelet/pods/d472d27e-9bc9-4d99-8fd9-697f7f44f57d/volumes" Dec 11 08:40:04 crc kubenswrapper[4860]: I1211 08:40:04.579490 4860 scope.go:117] "RemoveContainer" containerID="af76095360a11870b804fb7f16c40dad604dbf93152022249293687b1519ca1b" Dec 11 08:40:04 crc kubenswrapper[4860]: E1211 08:40:04.581095 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-99qgp_openshift-machine-config-operator(31c30642-6e60-41b4-a477-0d802424e0aa)\"" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" Dec 11 08:40:14 crc kubenswrapper[4860]: I1211 08:40:14.532954 4860 generic.go:334] "Generic (PLEG): container finished" podID="155181af-f5ab-41fb-9c0f-820c03b91b23" containerID="6154af5b0844abcaccc8d61bd88937426587c37ab32d347369bdbfd6bc3e2894" exitCode=0 Dec 11 08:40:14 crc kubenswrapper[4860]: I1211 08:40:14.533090 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-7zg5t" event={"ID":"155181af-f5ab-41fb-9c0f-820c03b91b23","Type":"ContainerDied","Data":"6154af5b0844abcaccc8d61bd88937426587c37ab32d347369bdbfd6bc3e2894"} Dec 11 08:40:15 crc kubenswrapper[4860]: I1211 08:40:15.969020 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-7zg5t" Dec 11 08:40:16 crc kubenswrapper[4860]: I1211 08:40:16.033033 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/155181af-f5ab-41fb-9c0f-820c03b91b23-inventory\") pod \"155181af-f5ab-41fb-9c0f-820c03b91b23\" (UID: \"155181af-f5ab-41fb-9c0f-820c03b91b23\") " Dec 11 08:40:16 crc kubenswrapper[4860]: I1211 08:40:16.033146 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5f6tj\" (UniqueName: \"kubernetes.io/projected/155181af-f5ab-41fb-9c0f-820c03b91b23-kube-api-access-5f6tj\") pod \"155181af-f5ab-41fb-9c0f-820c03b91b23\" (UID: \"155181af-f5ab-41fb-9c0f-820c03b91b23\") " Dec 11 08:40:16 crc kubenswrapper[4860]: I1211 08:40:16.033167 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/155181af-f5ab-41fb-9c0f-820c03b91b23-ssh-key\") pod \"155181af-f5ab-41fb-9c0f-820c03b91b23\" (UID: \"155181af-f5ab-41fb-9c0f-820c03b91b23\") " Dec 11 08:40:16 crc kubenswrapper[4860]: I1211 08:40:16.043970 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/155181af-f5ab-41fb-9c0f-820c03b91b23-kube-api-access-5f6tj" (OuterVolumeSpecName: "kube-api-access-5f6tj") pod "155181af-f5ab-41fb-9c0f-820c03b91b23" (UID: "155181af-f5ab-41fb-9c0f-820c03b91b23"). InnerVolumeSpecName "kube-api-access-5f6tj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:40:16 crc kubenswrapper[4860]: I1211 08:40:16.063540 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/155181af-f5ab-41fb-9c0f-820c03b91b23-inventory" (OuterVolumeSpecName: "inventory") pod "155181af-f5ab-41fb-9c0f-820c03b91b23" (UID: "155181af-f5ab-41fb-9c0f-820c03b91b23"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:40:16 crc kubenswrapper[4860]: I1211 08:40:16.070826 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/155181af-f5ab-41fb-9c0f-820c03b91b23-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "155181af-f5ab-41fb-9c0f-820c03b91b23" (UID: "155181af-f5ab-41fb-9c0f-820c03b91b23"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:40:16 crc kubenswrapper[4860]: I1211 08:40:16.135897 4860 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/155181af-f5ab-41fb-9c0f-820c03b91b23-inventory\") on node \"crc\" DevicePath \"\"" Dec 11 08:40:16 crc kubenswrapper[4860]: I1211 08:40:16.136182 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5f6tj\" (UniqueName: \"kubernetes.io/projected/155181af-f5ab-41fb-9c0f-820c03b91b23-kube-api-access-5f6tj\") on node \"crc\" DevicePath \"\"" Dec 11 08:40:16 crc kubenswrapper[4860]: I1211 08:40:16.136364 4860 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/155181af-f5ab-41fb-9c0f-820c03b91b23-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 11 08:40:16 crc kubenswrapper[4860]: I1211 08:40:16.553675 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-7zg5t" event={"ID":"155181af-f5ab-41fb-9c0f-820c03b91b23","Type":"ContainerDied","Data":"2656961f7fd1879a7c668b128917096b708da2db5af1cdde63710d8d770bb030"} Dec 11 08:40:16 crc kubenswrapper[4860]: I1211 08:40:16.553746 4860 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2656961f7fd1879a7c668b128917096b708da2db5af1cdde63710d8d770bb030" Dec 11 08:40:16 crc kubenswrapper[4860]: I1211 08:40:16.553710 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-7zg5t" Dec 11 08:40:16 crc kubenswrapper[4860]: I1211 08:40:16.640979 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ljr6l"] Dec 11 08:40:16 crc kubenswrapper[4860]: E1211 08:40:16.641405 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="155181af-f5ab-41fb-9c0f-820c03b91b23" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 11 08:40:16 crc kubenswrapper[4860]: I1211 08:40:16.641421 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="155181af-f5ab-41fb-9c0f-820c03b91b23" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 11 08:40:16 crc kubenswrapper[4860]: I1211 08:40:16.641593 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="155181af-f5ab-41fb-9c0f-820c03b91b23" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 11 08:40:16 crc kubenswrapper[4860]: I1211 08:40:16.642325 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ljr6l" Dec 11 08:40:16 crc kubenswrapper[4860]: I1211 08:40:16.644912 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/27ee41d4-1afd-4700-828c-7388caf498c5-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-ljr6l\" (UID: \"27ee41d4-1afd-4700-828c-7388caf498c5\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ljr6l" Dec 11 08:40:16 crc kubenswrapper[4860]: I1211 08:40:16.644989 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-spdkq\" (UniqueName: \"kubernetes.io/projected/27ee41d4-1afd-4700-828c-7388caf498c5-kube-api-access-spdkq\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-ljr6l\" (UID: \"27ee41d4-1afd-4700-828c-7388caf498c5\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ljr6l" Dec 11 08:40:16 crc kubenswrapper[4860]: I1211 08:40:16.645025 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/27ee41d4-1afd-4700-828c-7388caf498c5-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-ljr6l\" (UID: \"27ee41d4-1afd-4700-828c-7388caf498c5\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ljr6l" Dec 11 08:40:16 crc kubenswrapper[4860]: I1211 08:40:16.646521 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 11 08:40:16 crc kubenswrapper[4860]: I1211 08:40:16.646560 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 11 08:40:16 crc kubenswrapper[4860]: I1211 08:40:16.646634 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 11 08:40:16 crc kubenswrapper[4860]: I1211 08:40:16.646999 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-7w7kh" Dec 11 08:40:16 crc kubenswrapper[4860]: I1211 08:40:16.669235 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ljr6l"] Dec 11 08:40:16 crc kubenswrapper[4860]: I1211 08:40:16.746849 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/27ee41d4-1afd-4700-828c-7388caf498c5-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-ljr6l\" (UID: \"27ee41d4-1afd-4700-828c-7388caf498c5\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ljr6l" Dec 11 08:40:16 crc kubenswrapper[4860]: I1211 08:40:16.746918 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-spdkq\" (UniqueName: \"kubernetes.io/projected/27ee41d4-1afd-4700-828c-7388caf498c5-kube-api-access-spdkq\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-ljr6l\" (UID: \"27ee41d4-1afd-4700-828c-7388caf498c5\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ljr6l" Dec 11 08:40:16 crc kubenswrapper[4860]: I1211 08:40:16.746950 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/27ee41d4-1afd-4700-828c-7388caf498c5-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-ljr6l\" (UID: \"27ee41d4-1afd-4700-828c-7388caf498c5\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ljr6l" Dec 11 08:40:16 crc kubenswrapper[4860]: I1211 08:40:16.752414 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/27ee41d4-1afd-4700-828c-7388caf498c5-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-ljr6l\" (UID: \"27ee41d4-1afd-4700-828c-7388caf498c5\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ljr6l" Dec 11 08:40:16 crc kubenswrapper[4860]: I1211 08:40:16.761141 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/27ee41d4-1afd-4700-828c-7388caf498c5-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-ljr6l\" (UID: \"27ee41d4-1afd-4700-828c-7388caf498c5\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ljr6l" Dec 11 08:40:16 crc kubenswrapper[4860]: I1211 08:40:16.770705 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-spdkq\" (UniqueName: \"kubernetes.io/projected/27ee41d4-1afd-4700-828c-7388caf498c5-kube-api-access-spdkq\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-ljr6l\" (UID: \"27ee41d4-1afd-4700-828c-7388caf498c5\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ljr6l" Dec 11 08:40:16 crc kubenswrapper[4860]: I1211 08:40:16.966262 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ljr6l" Dec 11 08:40:17 crc kubenswrapper[4860]: I1211 08:40:17.475005 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ljr6l"] Dec 11 08:40:17 crc kubenswrapper[4860]: I1211 08:40:17.562052 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ljr6l" event={"ID":"27ee41d4-1afd-4700-828c-7388caf498c5","Type":"ContainerStarted","Data":"9955a156164211309a06f84a5ed0fd0826daa338273f2e64a85c6e05ae9574f7"} Dec 11 08:40:18 crc kubenswrapper[4860]: I1211 08:40:18.572271 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ljr6l" event={"ID":"27ee41d4-1afd-4700-828c-7388caf498c5","Type":"ContainerStarted","Data":"438d66d7267ab4ac99def14cb53b079a3904d94657f3d14ca6ec3187aed4d37b"} Dec 11 08:40:18 crc kubenswrapper[4860]: I1211 08:40:18.598783 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ljr6l" podStartSLOduration=1.886408588 podStartE2EDuration="2.59875618s" podCreationTimestamp="2025-12-11 08:40:16 +0000 UTC" firstStartedPulling="2025-12-11 08:40:17.477057914 +0000 UTC m=+1750.205576969" lastFinishedPulling="2025-12-11 08:40:18.189405506 +0000 UTC m=+1750.917924561" observedRunningTime="2025-12-11 08:40:18.589502418 +0000 UTC m=+1751.318021483" watchObservedRunningTime="2025-12-11 08:40:18.59875618 +0000 UTC m=+1751.327275245" Dec 11 08:40:19 crc kubenswrapper[4860]: I1211 08:40:19.579480 4860 scope.go:117] "RemoveContainer" containerID="af76095360a11870b804fb7f16c40dad604dbf93152022249293687b1519ca1b" Dec 11 08:40:19 crc kubenswrapper[4860]: E1211 08:40:19.579952 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-99qgp_openshift-machine-config-operator(31c30642-6e60-41b4-a477-0d802424e0aa)\"" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" Dec 11 08:40:23 crc kubenswrapper[4860]: I1211 08:40:23.611912 4860 generic.go:334] "Generic (PLEG): container finished" podID="27ee41d4-1afd-4700-828c-7388caf498c5" containerID="438d66d7267ab4ac99def14cb53b079a3904d94657f3d14ca6ec3187aed4d37b" exitCode=0 Dec 11 08:40:23 crc kubenswrapper[4860]: I1211 08:40:23.612367 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ljr6l" event={"ID":"27ee41d4-1afd-4700-828c-7388caf498c5","Type":"ContainerDied","Data":"438d66d7267ab4ac99def14cb53b079a3904d94657f3d14ca6ec3187aed4d37b"} Dec 11 08:40:25 crc kubenswrapper[4860]: I1211 08:40:25.030187 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ljr6l" Dec 11 08:40:25 crc kubenswrapper[4860]: I1211 08:40:25.108015 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/27ee41d4-1afd-4700-828c-7388caf498c5-ssh-key\") pod \"27ee41d4-1afd-4700-828c-7388caf498c5\" (UID: \"27ee41d4-1afd-4700-828c-7388caf498c5\") " Dec 11 08:40:25 crc kubenswrapper[4860]: I1211 08:40:25.108135 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-spdkq\" (UniqueName: \"kubernetes.io/projected/27ee41d4-1afd-4700-828c-7388caf498c5-kube-api-access-spdkq\") pod \"27ee41d4-1afd-4700-828c-7388caf498c5\" (UID: \"27ee41d4-1afd-4700-828c-7388caf498c5\") " Dec 11 08:40:25 crc kubenswrapper[4860]: I1211 08:40:25.108165 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/27ee41d4-1afd-4700-828c-7388caf498c5-inventory\") pod \"27ee41d4-1afd-4700-828c-7388caf498c5\" (UID: \"27ee41d4-1afd-4700-828c-7388caf498c5\") " Dec 11 08:40:25 crc kubenswrapper[4860]: I1211 08:40:25.114244 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/27ee41d4-1afd-4700-828c-7388caf498c5-kube-api-access-spdkq" (OuterVolumeSpecName: "kube-api-access-spdkq") pod "27ee41d4-1afd-4700-828c-7388caf498c5" (UID: "27ee41d4-1afd-4700-828c-7388caf498c5"). InnerVolumeSpecName "kube-api-access-spdkq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:40:25 crc kubenswrapper[4860]: I1211 08:40:25.137219 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27ee41d4-1afd-4700-828c-7388caf498c5-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "27ee41d4-1afd-4700-828c-7388caf498c5" (UID: "27ee41d4-1afd-4700-828c-7388caf498c5"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:40:25 crc kubenswrapper[4860]: I1211 08:40:25.137309 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27ee41d4-1afd-4700-828c-7388caf498c5-inventory" (OuterVolumeSpecName: "inventory") pod "27ee41d4-1afd-4700-828c-7388caf498c5" (UID: "27ee41d4-1afd-4700-828c-7388caf498c5"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:40:25 crc kubenswrapper[4860]: I1211 08:40:25.210314 4860 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/27ee41d4-1afd-4700-828c-7388caf498c5-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 11 08:40:25 crc kubenswrapper[4860]: I1211 08:40:25.210348 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-spdkq\" (UniqueName: \"kubernetes.io/projected/27ee41d4-1afd-4700-828c-7388caf498c5-kube-api-access-spdkq\") on node \"crc\" DevicePath \"\"" Dec 11 08:40:25 crc kubenswrapper[4860]: I1211 08:40:25.210362 4860 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/27ee41d4-1afd-4700-828c-7388caf498c5-inventory\") on node \"crc\" DevicePath \"\"" Dec 11 08:40:25 crc kubenswrapper[4860]: I1211 08:40:25.638229 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ljr6l" event={"ID":"27ee41d4-1afd-4700-828c-7388caf498c5","Type":"ContainerDied","Data":"9955a156164211309a06f84a5ed0fd0826daa338273f2e64a85c6e05ae9574f7"} Dec 11 08:40:25 crc kubenswrapper[4860]: I1211 08:40:25.639914 4860 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9955a156164211309a06f84a5ed0fd0826daa338273f2e64a85c6e05ae9574f7" Dec 11 08:40:25 crc kubenswrapper[4860]: I1211 08:40:25.638279 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ljr6l" Dec 11 08:40:25 crc kubenswrapper[4860]: I1211 08:40:25.695984 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-j8pcd"] Dec 11 08:40:25 crc kubenswrapper[4860]: E1211 08:40:25.696394 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27ee41d4-1afd-4700-828c-7388caf498c5" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 11 08:40:25 crc kubenswrapper[4860]: I1211 08:40:25.696414 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="27ee41d4-1afd-4700-828c-7388caf498c5" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 11 08:40:25 crc kubenswrapper[4860]: I1211 08:40:25.696587 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="27ee41d4-1afd-4700-828c-7388caf498c5" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 11 08:40:25 crc kubenswrapper[4860]: I1211 08:40:25.697382 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-j8pcd" Dec 11 08:40:25 crc kubenswrapper[4860]: I1211 08:40:25.699800 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-7w7kh" Dec 11 08:40:25 crc kubenswrapper[4860]: I1211 08:40:25.699906 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 11 08:40:25 crc kubenswrapper[4860]: I1211 08:40:25.699722 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 11 08:40:25 crc kubenswrapper[4860]: I1211 08:40:25.700409 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 11 08:40:25 crc kubenswrapper[4860]: I1211 08:40:25.712314 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-j8pcd"] Dec 11 08:40:25 crc kubenswrapper[4860]: I1211 08:40:25.721980 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7dxnc\" (UniqueName: \"kubernetes.io/projected/f9f3c692-bca8-49fb-a602-cc747b566849-kube-api-access-7dxnc\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-j8pcd\" (UID: \"f9f3c692-bca8-49fb-a602-cc747b566849\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-j8pcd" Dec 11 08:40:25 crc kubenswrapper[4860]: I1211 08:40:25.722040 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f9f3c692-bca8-49fb-a602-cc747b566849-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-j8pcd\" (UID: \"f9f3c692-bca8-49fb-a602-cc747b566849\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-j8pcd" Dec 11 08:40:25 crc kubenswrapper[4860]: I1211 08:40:25.722093 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f9f3c692-bca8-49fb-a602-cc747b566849-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-j8pcd\" (UID: \"f9f3c692-bca8-49fb-a602-cc747b566849\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-j8pcd" Dec 11 08:40:25 crc kubenswrapper[4860]: I1211 08:40:25.824010 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7dxnc\" (UniqueName: \"kubernetes.io/projected/f9f3c692-bca8-49fb-a602-cc747b566849-kube-api-access-7dxnc\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-j8pcd\" (UID: \"f9f3c692-bca8-49fb-a602-cc747b566849\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-j8pcd" Dec 11 08:40:25 crc kubenswrapper[4860]: I1211 08:40:25.824128 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f9f3c692-bca8-49fb-a602-cc747b566849-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-j8pcd\" (UID: \"f9f3c692-bca8-49fb-a602-cc747b566849\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-j8pcd" Dec 11 08:40:25 crc kubenswrapper[4860]: I1211 08:40:25.824229 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f9f3c692-bca8-49fb-a602-cc747b566849-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-j8pcd\" (UID: \"f9f3c692-bca8-49fb-a602-cc747b566849\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-j8pcd" Dec 11 08:40:25 crc kubenswrapper[4860]: I1211 08:40:25.830611 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f9f3c692-bca8-49fb-a602-cc747b566849-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-j8pcd\" (UID: \"f9f3c692-bca8-49fb-a602-cc747b566849\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-j8pcd" Dec 11 08:40:25 crc kubenswrapper[4860]: I1211 08:40:25.831029 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f9f3c692-bca8-49fb-a602-cc747b566849-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-j8pcd\" (UID: \"f9f3c692-bca8-49fb-a602-cc747b566849\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-j8pcd" Dec 11 08:40:25 crc kubenswrapper[4860]: I1211 08:40:25.842296 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7dxnc\" (UniqueName: \"kubernetes.io/projected/f9f3c692-bca8-49fb-a602-cc747b566849-kube-api-access-7dxnc\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-j8pcd\" (UID: \"f9f3c692-bca8-49fb-a602-cc747b566849\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-j8pcd" Dec 11 08:40:26 crc kubenswrapper[4860]: I1211 08:40:26.023689 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-j8pcd" Dec 11 08:40:26 crc kubenswrapper[4860]: I1211 08:40:26.531193 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-j8pcd"] Dec 11 08:40:26 crc kubenswrapper[4860]: I1211 08:40:26.657501 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-j8pcd" event={"ID":"f9f3c692-bca8-49fb-a602-cc747b566849","Type":"ContainerStarted","Data":"3979c0da2b9ccd76af990c3622f49447c9b0a47ae645ae8a5e172ebb09891699"} Dec 11 08:40:27 crc kubenswrapper[4860]: I1211 08:40:27.668077 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-j8pcd" event={"ID":"f9f3c692-bca8-49fb-a602-cc747b566849","Type":"ContainerStarted","Data":"2a2c1f3353ad80dfcc56d9cc6f285d7fc75e13a39893d8cfab56a380d3aeea77"} Dec 11 08:40:27 crc kubenswrapper[4860]: I1211 08:40:27.688166 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-j8pcd" podStartSLOduration=2.001549628 podStartE2EDuration="2.688143181s" podCreationTimestamp="2025-12-11 08:40:25 +0000 UTC" firstStartedPulling="2025-12-11 08:40:26.533101217 +0000 UTC m=+1759.261620282" lastFinishedPulling="2025-12-11 08:40:27.21969478 +0000 UTC m=+1759.948213835" observedRunningTime="2025-12-11 08:40:27.685122717 +0000 UTC m=+1760.413641792" watchObservedRunningTime="2025-12-11 08:40:27.688143181 +0000 UTC m=+1760.416662236" Dec 11 08:40:31 crc kubenswrapper[4860]: I1211 08:40:31.580241 4860 scope.go:117] "RemoveContainer" containerID="af76095360a11870b804fb7f16c40dad604dbf93152022249293687b1519ca1b" Dec 11 08:40:31 crc kubenswrapper[4860]: E1211 08:40:31.581438 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-99qgp_openshift-machine-config-operator(31c30642-6e60-41b4-a477-0d802424e0aa)\"" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" Dec 11 08:40:32 crc kubenswrapper[4860]: I1211 08:40:32.537014 4860 scope.go:117] "RemoveContainer" containerID="2c222aa49386e8cb938f0cf793517fceae739a76f3c585759558399c58a07d51" Dec 11 08:40:32 crc kubenswrapper[4860]: I1211 08:40:32.563073 4860 scope.go:117] "RemoveContainer" containerID="5225f72cd5a3da2be39a5c10f3cec79dbb7979ee0b367d21f16de3045373d813" Dec 11 08:40:32 crc kubenswrapper[4860]: I1211 08:40:32.615012 4860 scope.go:117] "RemoveContainer" containerID="b5e16e2ff7b9a93d183828e46286db5391626046b43b832b915a44d625d1f621" Dec 11 08:40:32 crc kubenswrapper[4860]: I1211 08:40:32.668706 4860 scope.go:117] "RemoveContainer" containerID="3e0c4ee5f6bdd93a3d5e3c150e68872d9d1ed659b0878f63a767498c0c729c7c" Dec 11 08:40:37 crc kubenswrapper[4860]: I1211 08:40:37.042197 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-hc9mt"] Dec 11 08:40:37 crc kubenswrapper[4860]: I1211 08:40:37.051619 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-7bmlr"] Dec 11 08:40:37 crc kubenswrapper[4860]: I1211 08:40:37.059056 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-hc9mt"] Dec 11 08:40:37 crc kubenswrapper[4860]: I1211 08:40:37.066244 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-7bmlr"] Dec 11 08:40:37 crc kubenswrapper[4860]: I1211 08:40:37.589405 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b4598c5c-a584-487e-83ed-0126b5fd253e" path="/var/lib/kubelet/pods/b4598c5c-a584-487e-83ed-0126b5fd253e/volumes" Dec 11 08:40:37 crc kubenswrapper[4860]: I1211 08:40:37.589971 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e01fc44a-9c71-49bb-bc43-f4a20d44b41a" path="/var/lib/kubelet/pods/e01fc44a-9c71-49bb-bc43-f4a20d44b41a/volumes" Dec 11 08:40:38 crc kubenswrapper[4860]: I1211 08:40:38.027929 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-vzvdr"] Dec 11 08:40:38 crc kubenswrapper[4860]: I1211 08:40:38.038636 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-b7a3-account-create-update-qgft5"] Dec 11 08:40:38 crc kubenswrapper[4860]: I1211 08:40:38.048449 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-b7a3-account-create-update-qgft5"] Dec 11 08:40:38 crc kubenswrapper[4860]: I1211 08:40:38.057863 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-vzvdr"] Dec 11 08:40:39 crc kubenswrapper[4860]: I1211 08:40:39.025184 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-7093-account-create-update-4m7w4"] Dec 11 08:40:39 crc kubenswrapper[4860]: I1211 08:40:39.033014 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-caef-account-create-update-jwjlc"] Dec 11 08:40:39 crc kubenswrapper[4860]: I1211 08:40:39.041652 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-7093-account-create-update-4m7w4"] Dec 11 08:40:39 crc kubenswrapper[4860]: I1211 08:40:39.050336 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-caef-account-create-update-jwjlc"] Dec 11 08:40:39 crc kubenswrapper[4860]: I1211 08:40:39.591791 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="293660ed-0401-45cc-811e-a612da96284f" path="/var/lib/kubelet/pods/293660ed-0401-45cc-811e-a612da96284f/volumes" Dec 11 08:40:39 crc kubenswrapper[4860]: I1211 08:40:39.592396 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d35fc5-e254-4e3f-86bf-9a944f8a44af" path="/var/lib/kubelet/pods/31d35fc5-e254-4e3f-86bf-9a944f8a44af/volumes" Dec 11 08:40:39 crc kubenswrapper[4860]: I1211 08:40:39.592977 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6a7614eb-5a8b-44ca-afe2-22a8b1151065" path="/var/lib/kubelet/pods/6a7614eb-5a8b-44ca-afe2-22a8b1151065/volumes" Dec 11 08:40:39 crc kubenswrapper[4860]: I1211 08:40:39.593584 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fdae480c-a60d-4508-add1-e1e89bd3e8e3" path="/var/lib/kubelet/pods/fdae480c-a60d-4508-add1-e1e89bd3e8e3/volumes" Dec 11 08:40:46 crc kubenswrapper[4860]: I1211 08:40:46.579486 4860 scope.go:117] "RemoveContainer" containerID="af76095360a11870b804fb7f16c40dad604dbf93152022249293687b1519ca1b" Dec 11 08:40:46 crc kubenswrapper[4860]: E1211 08:40:46.580643 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-99qgp_openshift-machine-config-operator(31c30642-6e60-41b4-a477-0d802424e0aa)\"" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" Dec 11 08:40:58 crc kubenswrapper[4860]: I1211 08:40:58.578445 4860 scope.go:117] "RemoveContainer" containerID="af76095360a11870b804fb7f16c40dad604dbf93152022249293687b1519ca1b" Dec 11 08:40:58 crc kubenswrapper[4860]: E1211 08:40:58.579237 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-99qgp_openshift-machine-config-operator(31c30642-6e60-41b4-a477-0d802424e0aa)\"" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" Dec 11 08:41:01 crc kubenswrapper[4860]: I1211 08:41:01.977794 4860 generic.go:334] "Generic (PLEG): container finished" podID="f9f3c692-bca8-49fb-a602-cc747b566849" containerID="2a2c1f3353ad80dfcc56d9cc6f285d7fc75e13a39893d8cfab56a380d3aeea77" exitCode=0 Dec 11 08:41:01 crc kubenswrapper[4860]: I1211 08:41:01.977886 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-j8pcd" event={"ID":"f9f3c692-bca8-49fb-a602-cc747b566849","Type":"ContainerDied","Data":"2a2c1f3353ad80dfcc56d9cc6f285d7fc75e13a39893d8cfab56a380d3aeea77"} Dec 11 08:41:03 crc kubenswrapper[4860]: I1211 08:41:03.375075 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-j8pcd" Dec 11 08:41:03 crc kubenswrapper[4860]: I1211 08:41:03.553290 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f9f3c692-bca8-49fb-a602-cc747b566849-inventory\") pod \"f9f3c692-bca8-49fb-a602-cc747b566849\" (UID: \"f9f3c692-bca8-49fb-a602-cc747b566849\") " Dec 11 08:41:03 crc kubenswrapper[4860]: I1211 08:41:03.553987 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f9f3c692-bca8-49fb-a602-cc747b566849-ssh-key\") pod \"f9f3c692-bca8-49fb-a602-cc747b566849\" (UID: \"f9f3c692-bca8-49fb-a602-cc747b566849\") " Dec 11 08:41:03 crc kubenswrapper[4860]: I1211 08:41:03.554084 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7dxnc\" (UniqueName: \"kubernetes.io/projected/f9f3c692-bca8-49fb-a602-cc747b566849-kube-api-access-7dxnc\") pod \"f9f3c692-bca8-49fb-a602-cc747b566849\" (UID: \"f9f3c692-bca8-49fb-a602-cc747b566849\") " Dec 11 08:41:03 crc kubenswrapper[4860]: I1211 08:41:03.560283 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f9f3c692-bca8-49fb-a602-cc747b566849-kube-api-access-7dxnc" (OuterVolumeSpecName: "kube-api-access-7dxnc") pod "f9f3c692-bca8-49fb-a602-cc747b566849" (UID: "f9f3c692-bca8-49fb-a602-cc747b566849"). InnerVolumeSpecName "kube-api-access-7dxnc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:41:03 crc kubenswrapper[4860]: I1211 08:41:03.583185 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9f3c692-bca8-49fb-a602-cc747b566849-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "f9f3c692-bca8-49fb-a602-cc747b566849" (UID: "f9f3c692-bca8-49fb-a602-cc747b566849"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:41:03 crc kubenswrapper[4860]: I1211 08:41:03.595862 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9f3c692-bca8-49fb-a602-cc747b566849-inventory" (OuterVolumeSpecName: "inventory") pod "f9f3c692-bca8-49fb-a602-cc747b566849" (UID: "f9f3c692-bca8-49fb-a602-cc747b566849"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:41:03 crc kubenswrapper[4860]: I1211 08:41:03.656519 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7dxnc\" (UniqueName: \"kubernetes.io/projected/f9f3c692-bca8-49fb-a602-cc747b566849-kube-api-access-7dxnc\") on node \"crc\" DevicePath \"\"" Dec 11 08:41:03 crc kubenswrapper[4860]: I1211 08:41:03.656552 4860 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f9f3c692-bca8-49fb-a602-cc747b566849-inventory\") on node \"crc\" DevicePath \"\"" Dec 11 08:41:03 crc kubenswrapper[4860]: I1211 08:41:03.656563 4860 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f9f3c692-bca8-49fb-a602-cc747b566849-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 11 08:41:04 crc kubenswrapper[4860]: I1211 08:41:03.999966 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-j8pcd" event={"ID":"f9f3c692-bca8-49fb-a602-cc747b566849","Type":"ContainerDied","Data":"3979c0da2b9ccd76af990c3622f49447c9b0a47ae645ae8a5e172ebb09891699"} Dec 11 08:41:04 crc kubenswrapper[4860]: I1211 08:41:04.000012 4860 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3979c0da2b9ccd76af990c3622f49447c9b0a47ae645ae8a5e172ebb09891699" Dec 11 08:41:04 crc kubenswrapper[4860]: I1211 08:41:04.000127 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-j8pcd" Dec 11 08:41:04 crc kubenswrapper[4860]: I1211 08:41:04.119296 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-bv6xw"] Dec 11 08:41:04 crc kubenswrapper[4860]: E1211 08:41:04.120088 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9f3c692-bca8-49fb-a602-cc747b566849" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 11 08:41:04 crc kubenswrapper[4860]: I1211 08:41:04.120210 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9f3c692-bca8-49fb-a602-cc747b566849" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 11 08:41:04 crc kubenswrapper[4860]: I1211 08:41:04.120526 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="f9f3c692-bca8-49fb-a602-cc747b566849" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 11 08:41:04 crc kubenswrapper[4860]: I1211 08:41:04.121428 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-bv6xw" Dec 11 08:41:04 crc kubenswrapper[4860]: I1211 08:41:04.125146 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 11 08:41:04 crc kubenswrapper[4860]: I1211 08:41:04.126798 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 11 08:41:04 crc kubenswrapper[4860]: I1211 08:41:04.126816 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-7w7kh" Dec 11 08:41:04 crc kubenswrapper[4860]: I1211 08:41:04.134638 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-bv6xw"] Dec 11 08:41:04 crc kubenswrapper[4860]: I1211 08:41:04.134961 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 11 08:41:04 crc kubenswrapper[4860]: I1211 08:41:04.267260 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2651e591-78fd-410d-b8a0-bcb0da2066bb-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-bv6xw\" (UID: \"2651e591-78fd-410d-b8a0-bcb0da2066bb\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-bv6xw" Dec 11 08:41:04 crc kubenswrapper[4860]: I1211 08:41:04.267664 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2651e591-78fd-410d-b8a0-bcb0da2066bb-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-bv6xw\" (UID: \"2651e591-78fd-410d-b8a0-bcb0da2066bb\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-bv6xw" Dec 11 08:41:04 crc kubenswrapper[4860]: I1211 08:41:04.267791 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mghv9\" (UniqueName: \"kubernetes.io/projected/2651e591-78fd-410d-b8a0-bcb0da2066bb-kube-api-access-mghv9\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-bv6xw\" (UID: \"2651e591-78fd-410d-b8a0-bcb0da2066bb\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-bv6xw" Dec 11 08:41:04 crc kubenswrapper[4860]: I1211 08:41:04.369843 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mghv9\" (UniqueName: \"kubernetes.io/projected/2651e591-78fd-410d-b8a0-bcb0da2066bb-kube-api-access-mghv9\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-bv6xw\" (UID: \"2651e591-78fd-410d-b8a0-bcb0da2066bb\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-bv6xw" Dec 11 08:41:04 crc kubenswrapper[4860]: I1211 08:41:04.370035 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2651e591-78fd-410d-b8a0-bcb0da2066bb-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-bv6xw\" (UID: \"2651e591-78fd-410d-b8a0-bcb0da2066bb\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-bv6xw" Dec 11 08:41:04 crc kubenswrapper[4860]: I1211 08:41:04.370249 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2651e591-78fd-410d-b8a0-bcb0da2066bb-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-bv6xw\" (UID: \"2651e591-78fd-410d-b8a0-bcb0da2066bb\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-bv6xw" Dec 11 08:41:04 crc kubenswrapper[4860]: I1211 08:41:04.375355 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2651e591-78fd-410d-b8a0-bcb0da2066bb-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-bv6xw\" (UID: \"2651e591-78fd-410d-b8a0-bcb0da2066bb\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-bv6xw" Dec 11 08:41:04 crc kubenswrapper[4860]: I1211 08:41:04.377065 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2651e591-78fd-410d-b8a0-bcb0da2066bb-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-bv6xw\" (UID: \"2651e591-78fd-410d-b8a0-bcb0da2066bb\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-bv6xw" Dec 11 08:41:04 crc kubenswrapper[4860]: I1211 08:41:04.403838 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mghv9\" (UniqueName: \"kubernetes.io/projected/2651e591-78fd-410d-b8a0-bcb0da2066bb-kube-api-access-mghv9\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-bv6xw\" (UID: \"2651e591-78fd-410d-b8a0-bcb0da2066bb\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-bv6xw" Dec 11 08:41:04 crc kubenswrapper[4860]: I1211 08:41:04.446135 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-bv6xw" Dec 11 08:41:04 crc kubenswrapper[4860]: I1211 08:41:04.937891 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-bv6xw"] Dec 11 08:41:05 crc kubenswrapper[4860]: I1211 08:41:05.008409 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-bv6xw" event={"ID":"2651e591-78fd-410d-b8a0-bcb0da2066bb","Type":"ContainerStarted","Data":"10bf3008aec550c0a558bcf05a94f6772c899ae2d3f271b4b6e697b720e9712d"} Dec 11 08:41:06 crc kubenswrapper[4860]: I1211 08:41:06.034936 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-bv6xw" event={"ID":"2651e591-78fd-410d-b8a0-bcb0da2066bb","Type":"ContainerStarted","Data":"8abae7dd40bc66121d08c99b58e5a5201dc36f3e53a6e71712148f20eff914e7"} Dec 11 08:41:06 crc kubenswrapper[4860]: I1211 08:41:06.069491 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-kb24v"] Dec 11 08:41:06 crc kubenswrapper[4860]: I1211 08:41:06.078562 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-kb24v"] Dec 11 08:41:06 crc kubenswrapper[4860]: I1211 08:41:06.082613 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-bv6xw" podStartSLOduration=1.365399855 podStartE2EDuration="2.082600491s" podCreationTimestamp="2025-12-11 08:41:04 +0000 UTC" firstStartedPulling="2025-12-11 08:41:04.943824647 +0000 UTC m=+1797.672343702" lastFinishedPulling="2025-12-11 08:41:05.661025283 +0000 UTC m=+1798.389544338" observedRunningTime="2025-12-11 08:41:06.048319962 +0000 UTC m=+1798.776839017" watchObservedRunningTime="2025-12-11 08:41:06.082600491 +0000 UTC m=+1798.811119546" Dec 11 08:41:07 crc kubenswrapper[4860]: I1211 08:41:07.590502 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff7bf9ff-e357-4af7-b708-2eb6298b85a8" path="/var/lib/kubelet/pods/ff7bf9ff-e357-4af7-b708-2eb6298b85a8/volumes" Dec 11 08:41:11 crc kubenswrapper[4860]: I1211 08:41:11.579747 4860 scope.go:117] "RemoveContainer" containerID="af76095360a11870b804fb7f16c40dad604dbf93152022249293687b1519ca1b" Dec 11 08:41:11 crc kubenswrapper[4860]: E1211 08:41:11.580344 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-99qgp_openshift-machine-config-operator(31c30642-6e60-41b4-a477-0d802424e0aa)\"" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" Dec 11 08:41:26 crc kubenswrapper[4860]: I1211 08:41:26.578801 4860 scope.go:117] "RemoveContainer" containerID="af76095360a11870b804fb7f16c40dad604dbf93152022249293687b1519ca1b" Dec 11 08:41:26 crc kubenswrapper[4860]: E1211 08:41:26.582754 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-99qgp_openshift-machine-config-operator(31c30642-6e60-41b4-a477-0d802424e0aa)\"" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" Dec 11 08:41:29 crc kubenswrapper[4860]: I1211 08:41:29.049228 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-nvt7h"] Dec 11 08:41:29 crc kubenswrapper[4860]: I1211 08:41:29.057619 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-nvt7h"] Dec 11 08:41:29 crc kubenswrapper[4860]: I1211 08:41:29.589480 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="983cbde3-46fe-4643-b213-c74900941587" path="/var/lib/kubelet/pods/983cbde3-46fe-4643-b213-c74900941587/volumes" Dec 11 08:41:30 crc kubenswrapper[4860]: I1211 08:41:30.029197 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-bq4cn"] Dec 11 08:41:30 crc kubenswrapper[4860]: I1211 08:41:30.040430 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-bq4cn"] Dec 11 08:41:31 crc kubenswrapper[4860]: I1211 08:41:31.590013 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af3bd0b9-ab28-4e7b-8316-3206671fb97b" path="/var/lib/kubelet/pods/af3bd0b9-ab28-4e7b-8316-3206671fb97b/volumes" Dec 11 08:41:32 crc kubenswrapper[4860]: I1211 08:41:32.850494 4860 scope.go:117] "RemoveContainer" containerID="ba86ec1b211d7519b3b1d47ede401a6153052bcb53b200f1e27b5cbe89e1a916" Dec 11 08:41:32 crc kubenswrapper[4860]: I1211 08:41:32.904071 4860 scope.go:117] "RemoveContainer" containerID="4de218dffc3e8e24efb4fe4428903323ac601fe2154ece0ae358ea367ca41eb9" Dec 11 08:41:32 crc kubenswrapper[4860]: I1211 08:41:32.927975 4860 scope.go:117] "RemoveContainer" containerID="ec87f01347028dc991f26c633f3ca640d51582e9ba4678b414d94b21b0235ef4" Dec 11 08:41:32 crc kubenswrapper[4860]: I1211 08:41:32.974503 4860 scope.go:117] "RemoveContainer" containerID="25fa35ea439bdf563989e6595788235e05f9c66929274bdfa3aa467f9c6dac01" Dec 11 08:41:33 crc kubenswrapper[4860]: I1211 08:41:33.036466 4860 scope.go:117] "RemoveContainer" containerID="b217915b36edf21ab47701779f8cba774a5f626850151fb5e7bd9eeb8370d4d6" Dec 11 08:41:33 crc kubenswrapper[4860]: I1211 08:41:33.057204 4860 scope.go:117] "RemoveContainer" containerID="b29ada59900b63fcf4b78fcd929f9f91ae19df8e2b3a8506766956c7d67e078c" Dec 11 08:41:33 crc kubenswrapper[4860]: I1211 08:41:33.111298 4860 scope.go:117] "RemoveContainer" containerID="a8dfc96d64f0aed235dcb8437d52cc0a75402373d7c90da2ec678b3aecf52e4f" Dec 11 08:41:33 crc kubenswrapper[4860]: I1211 08:41:33.129275 4860 scope.go:117] "RemoveContainer" containerID="be956e7e27260afa9cb66d3f8d92a9d6ba13ffd3eb5429dc7166b50c41401383" Dec 11 08:41:33 crc kubenswrapper[4860]: I1211 08:41:33.146489 4860 scope.go:117] "RemoveContainer" containerID="d5b4720e97798804622ed48c582631bd6b457e29bca173b7f14ad4cc5860fe14" Dec 11 08:41:40 crc kubenswrapper[4860]: I1211 08:41:40.579750 4860 scope.go:117] "RemoveContainer" containerID="af76095360a11870b804fb7f16c40dad604dbf93152022249293687b1519ca1b" Dec 11 08:41:40 crc kubenswrapper[4860]: E1211 08:41:40.580525 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-99qgp_openshift-machine-config-operator(31c30642-6e60-41b4-a477-0d802424e0aa)\"" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" Dec 11 08:41:52 crc kubenswrapper[4860]: I1211 08:41:52.578790 4860 scope.go:117] "RemoveContainer" containerID="af76095360a11870b804fb7f16c40dad604dbf93152022249293687b1519ca1b" Dec 11 08:41:52 crc kubenswrapper[4860]: E1211 08:41:52.579518 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-99qgp_openshift-machine-config-operator(31c30642-6e60-41b4-a477-0d802424e0aa)\"" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" Dec 11 08:41:54 crc kubenswrapper[4860]: I1211 08:41:54.494942 4860 generic.go:334] "Generic (PLEG): container finished" podID="2651e591-78fd-410d-b8a0-bcb0da2066bb" containerID="8abae7dd40bc66121d08c99b58e5a5201dc36f3e53a6e71712148f20eff914e7" exitCode=0 Dec 11 08:41:54 crc kubenswrapper[4860]: I1211 08:41:54.495020 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-bv6xw" event={"ID":"2651e591-78fd-410d-b8a0-bcb0da2066bb","Type":"ContainerDied","Data":"8abae7dd40bc66121d08c99b58e5a5201dc36f3e53a6e71712148f20eff914e7"} Dec 11 08:41:55 crc kubenswrapper[4860]: I1211 08:41:55.906724 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-bv6xw" Dec 11 08:41:56 crc kubenswrapper[4860]: I1211 08:41:56.049443 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2651e591-78fd-410d-b8a0-bcb0da2066bb-inventory\") pod \"2651e591-78fd-410d-b8a0-bcb0da2066bb\" (UID: \"2651e591-78fd-410d-b8a0-bcb0da2066bb\") " Dec 11 08:41:56 crc kubenswrapper[4860]: I1211 08:41:56.049979 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mghv9\" (UniqueName: \"kubernetes.io/projected/2651e591-78fd-410d-b8a0-bcb0da2066bb-kube-api-access-mghv9\") pod \"2651e591-78fd-410d-b8a0-bcb0da2066bb\" (UID: \"2651e591-78fd-410d-b8a0-bcb0da2066bb\") " Dec 11 08:41:56 crc kubenswrapper[4860]: I1211 08:41:56.050049 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2651e591-78fd-410d-b8a0-bcb0da2066bb-ssh-key\") pod \"2651e591-78fd-410d-b8a0-bcb0da2066bb\" (UID: \"2651e591-78fd-410d-b8a0-bcb0da2066bb\") " Dec 11 08:41:56 crc kubenswrapper[4860]: I1211 08:41:56.055350 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2651e591-78fd-410d-b8a0-bcb0da2066bb-kube-api-access-mghv9" (OuterVolumeSpecName: "kube-api-access-mghv9") pod "2651e591-78fd-410d-b8a0-bcb0da2066bb" (UID: "2651e591-78fd-410d-b8a0-bcb0da2066bb"). InnerVolumeSpecName "kube-api-access-mghv9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:41:56 crc kubenswrapper[4860]: I1211 08:41:56.078407 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2651e591-78fd-410d-b8a0-bcb0da2066bb-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "2651e591-78fd-410d-b8a0-bcb0da2066bb" (UID: "2651e591-78fd-410d-b8a0-bcb0da2066bb"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:41:56 crc kubenswrapper[4860]: I1211 08:41:56.091813 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2651e591-78fd-410d-b8a0-bcb0da2066bb-inventory" (OuterVolumeSpecName: "inventory") pod "2651e591-78fd-410d-b8a0-bcb0da2066bb" (UID: "2651e591-78fd-410d-b8a0-bcb0da2066bb"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:41:56 crc kubenswrapper[4860]: I1211 08:41:56.152801 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mghv9\" (UniqueName: \"kubernetes.io/projected/2651e591-78fd-410d-b8a0-bcb0da2066bb-kube-api-access-mghv9\") on node \"crc\" DevicePath \"\"" Dec 11 08:41:56 crc kubenswrapper[4860]: I1211 08:41:56.153128 4860 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2651e591-78fd-410d-b8a0-bcb0da2066bb-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 11 08:41:56 crc kubenswrapper[4860]: I1211 08:41:56.153230 4860 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2651e591-78fd-410d-b8a0-bcb0da2066bb-inventory\") on node \"crc\" DevicePath \"\"" Dec 11 08:41:56 crc kubenswrapper[4860]: I1211 08:41:56.514581 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-bv6xw" event={"ID":"2651e591-78fd-410d-b8a0-bcb0da2066bb","Type":"ContainerDied","Data":"10bf3008aec550c0a558bcf05a94f6772c899ae2d3f271b4b6e697b720e9712d"} Dec 11 08:41:56 crc kubenswrapper[4860]: I1211 08:41:56.514687 4860 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="10bf3008aec550c0a558bcf05a94f6772c899ae2d3f271b4b6e697b720e9712d" Dec 11 08:41:56 crc kubenswrapper[4860]: I1211 08:41:56.514773 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-bv6xw" Dec 11 08:41:56 crc kubenswrapper[4860]: I1211 08:41:56.608436 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-rl5hl"] Dec 11 08:41:56 crc kubenswrapper[4860]: E1211 08:41:56.608855 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2651e591-78fd-410d-b8a0-bcb0da2066bb" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 11 08:41:56 crc kubenswrapper[4860]: I1211 08:41:56.608875 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="2651e591-78fd-410d-b8a0-bcb0da2066bb" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 11 08:41:56 crc kubenswrapper[4860]: I1211 08:41:56.609064 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="2651e591-78fd-410d-b8a0-bcb0da2066bb" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 11 08:41:56 crc kubenswrapper[4860]: I1211 08:41:56.609677 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-rl5hl" Dec 11 08:41:56 crc kubenswrapper[4860]: I1211 08:41:56.612455 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 11 08:41:56 crc kubenswrapper[4860]: I1211 08:41:56.612800 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 11 08:41:56 crc kubenswrapper[4860]: I1211 08:41:56.612977 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-7w7kh" Dec 11 08:41:56 crc kubenswrapper[4860]: I1211 08:41:56.613019 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 11 08:41:56 crc kubenswrapper[4860]: I1211 08:41:56.669005 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-rl5hl"] Dec 11 08:41:56 crc kubenswrapper[4860]: I1211 08:41:56.767593 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/6a0f9071-e3f7-43f2-a0ec-8ab22fd9c785-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-rl5hl\" (UID: \"6a0f9071-e3f7-43f2-a0ec-8ab22fd9c785\") " pod="openstack/ssh-known-hosts-edpm-deployment-rl5hl" Dec 11 08:41:56 crc kubenswrapper[4860]: I1211 08:41:56.767816 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/6a0f9071-e3f7-43f2-a0ec-8ab22fd9c785-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-rl5hl\" (UID: \"6a0f9071-e3f7-43f2-a0ec-8ab22fd9c785\") " pod="openstack/ssh-known-hosts-edpm-deployment-rl5hl" Dec 11 08:41:56 crc kubenswrapper[4860]: I1211 08:41:56.767857 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bm6mg\" (UniqueName: \"kubernetes.io/projected/6a0f9071-e3f7-43f2-a0ec-8ab22fd9c785-kube-api-access-bm6mg\") pod \"ssh-known-hosts-edpm-deployment-rl5hl\" (UID: \"6a0f9071-e3f7-43f2-a0ec-8ab22fd9c785\") " pod="openstack/ssh-known-hosts-edpm-deployment-rl5hl" Dec 11 08:41:56 crc kubenswrapper[4860]: I1211 08:41:56.869498 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/6a0f9071-e3f7-43f2-a0ec-8ab22fd9c785-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-rl5hl\" (UID: \"6a0f9071-e3f7-43f2-a0ec-8ab22fd9c785\") " pod="openstack/ssh-known-hosts-edpm-deployment-rl5hl" Dec 11 08:41:56 crc kubenswrapper[4860]: I1211 08:41:56.869615 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/6a0f9071-e3f7-43f2-a0ec-8ab22fd9c785-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-rl5hl\" (UID: \"6a0f9071-e3f7-43f2-a0ec-8ab22fd9c785\") " pod="openstack/ssh-known-hosts-edpm-deployment-rl5hl" Dec 11 08:41:56 crc kubenswrapper[4860]: I1211 08:41:56.869664 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bm6mg\" (UniqueName: \"kubernetes.io/projected/6a0f9071-e3f7-43f2-a0ec-8ab22fd9c785-kube-api-access-bm6mg\") pod \"ssh-known-hosts-edpm-deployment-rl5hl\" (UID: \"6a0f9071-e3f7-43f2-a0ec-8ab22fd9c785\") " pod="openstack/ssh-known-hosts-edpm-deployment-rl5hl" Dec 11 08:41:56 crc kubenswrapper[4860]: I1211 08:41:56.875081 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/6a0f9071-e3f7-43f2-a0ec-8ab22fd9c785-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-rl5hl\" (UID: \"6a0f9071-e3f7-43f2-a0ec-8ab22fd9c785\") " pod="openstack/ssh-known-hosts-edpm-deployment-rl5hl" Dec 11 08:41:56 crc kubenswrapper[4860]: I1211 08:41:56.877370 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/6a0f9071-e3f7-43f2-a0ec-8ab22fd9c785-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-rl5hl\" (UID: \"6a0f9071-e3f7-43f2-a0ec-8ab22fd9c785\") " pod="openstack/ssh-known-hosts-edpm-deployment-rl5hl" Dec 11 08:41:56 crc kubenswrapper[4860]: I1211 08:41:56.895938 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bm6mg\" (UniqueName: \"kubernetes.io/projected/6a0f9071-e3f7-43f2-a0ec-8ab22fd9c785-kube-api-access-bm6mg\") pod \"ssh-known-hosts-edpm-deployment-rl5hl\" (UID: \"6a0f9071-e3f7-43f2-a0ec-8ab22fd9c785\") " pod="openstack/ssh-known-hosts-edpm-deployment-rl5hl" Dec 11 08:41:56 crc kubenswrapper[4860]: I1211 08:41:56.984400 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-rl5hl" Dec 11 08:41:57 crc kubenswrapper[4860]: I1211 08:41:57.529797 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-rl5hl"] Dec 11 08:41:58 crc kubenswrapper[4860]: I1211 08:41:58.531190 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-rl5hl" event={"ID":"6a0f9071-e3f7-43f2-a0ec-8ab22fd9c785","Type":"ContainerStarted","Data":"7ad58632a5c57cad8398252ac4d14597525de932102049a0d46d1146166b7ae1"} Dec 11 08:41:59 crc kubenswrapper[4860]: I1211 08:41:59.545628 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-rl5hl" event={"ID":"6a0f9071-e3f7-43f2-a0ec-8ab22fd9c785","Type":"ContainerStarted","Data":"e1248a2d1096a37e18633d487f5b0562791dc8971f5a535515a6460ae1ecc7d7"} Dec 11 08:41:59 crc kubenswrapper[4860]: I1211 08:41:59.561048 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-rl5hl" podStartSLOduration=2.526455968 podStartE2EDuration="3.561031205s" podCreationTimestamp="2025-12-11 08:41:56 +0000 UTC" firstStartedPulling="2025-12-11 08:41:57.537698733 +0000 UTC m=+1850.266217788" lastFinishedPulling="2025-12-11 08:41:58.57227397 +0000 UTC m=+1851.300793025" observedRunningTime="2025-12-11 08:41:59.558363821 +0000 UTC m=+1852.286882886" watchObservedRunningTime="2025-12-11 08:41:59.561031205 +0000 UTC m=+1852.289550260" Dec 11 08:42:05 crc kubenswrapper[4860]: I1211 08:42:05.579502 4860 scope.go:117] "RemoveContainer" containerID="af76095360a11870b804fb7f16c40dad604dbf93152022249293687b1519ca1b" Dec 11 08:42:05 crc kubenswrapper[4860]: E1211 08:42:05.580419 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-99qgp_openshift-machine-config-operator(31c30642-6e60-41b4-a477-0d802424e0aa)\"" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" Dec 11 08:42:05 crc kubenswrapper[4860]: I1211 08:42:05.597522 4860 generic.go:334] "Generic (PLEG): container finished" podID="6a0f9071-e3f7-43f2-a0ec-8ab22fd9c785" containerID="e1248a2d1096a37e18633d487f5b0562791dc8971f5a535515a6460ae1ecc7d7" exitCode=0 Dec 11 08:42:05 crc kubenswrapper[4860]: I1211 08:42:05.598201 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-rl5hl" event={"ID":"6a0f9071-e3f7-43f2-a0ec-8ab22fd9c785","Type":"ContainerDied","Data":"e1248a2d1096a37e18633d487f5b0562791dc8971f5a535515a6460ae1ecc7d7"} Dec 11 08:42:07 crc kubenswrapper[4860]: I1211 08:42:07.062555 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-rl5hl" Dec 11 08:42:07 crc kubenswrapper[4860]: I1211 08:42:07.126449 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/6a0f9071-e3f7-43f2-a0ec-8ab22fd9c785-inventory-0\") pod \"6a0f9071-e3f7-43f2-a0ec-8ab22fd9c785\" (UID: \"6a0f9071-e3f7-43f2-a0ec-8ab22fd9c785\") " Dec 11 08:42:07 crc kubenswrapper[4860]: I1211 08:42:07.126557 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/6a0f9071-e3f7-43f2-a0ec-8ab22fd9c785-ssh-key-openstack-edpm-ipam\") pod \"6a0f9071-e3f7-43f2-a0ec-8ab22fd9c785\" (UID: \"6a0f9071-e3f7-43f2-a0ec-8ab22fd9c785\") " Dec 11 08:42:07 crc kubenswrapper[4860]: I1211 08:42:07.126603 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bm6mg\" (UniqueName: \"kubernetes.io/projected/6a0f9071-e3f7-43f2-a0ec-8ab22fd9c785-kube-api-access-bm6mg\") pod \"6a0f9071-e3f7-43f2-a0ec-8ab22fd9c785\" (UID: \"6a0f9071-e3f7-43f2-a0ec-8ab22fd9c785\") " Dec 11 08:42:07 crc kubenswrapper[4860]: I1211 08:42:07.132608 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a0f9071-e3f7-43f2-a0ec-8ab22fd9c785-kube-api-access-bm6mg" (OuterVolumeSpecName: "kube-api-access-bm6mg") pod "6a0f9071-e3f7-43f2-a0ec-8ab22fd9c785" (UID: "6a0f9071-e3f7-43f2-a0ec-8ab22fd9c785"). InnerVolumeSpecName "kube-api-access-bm6mg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:42:07 crc kubenswrapper[4860]: I1211 08:42:07.153114 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a0f9071-e3f7-43f2-a0ec-8ab22fd9c785-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "6a0f9071-e3f7-43f2-a0ec-8ab22fd9c785" (UID: "6a0f9071-e3f7-43f2-a0ec-8ab22fd9c785"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:42:07 crc kubenswrapper[4860]: I1211 08:42:07.153692 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a0f9071-e3f7-43f2-a0ec-8ab22fd9c785-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "6a0f9071-e3f7-43f2-a0ec-8ab22fd9c785" (UID: "6a0f9071-e3f7-43f2-a0ec-8ab22fd9c785"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:42:07 crc kubenswrapper[4860]: I1211 08:42:07.229310 4860 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/6a0f9071-e3f7-43f2-a0ec-8ab22fd9c785-inventory-0\") on node \"crc\" DevicePath \"\"" Dec 11 08:42:07 crc kubenswrapper[4860]: I1211 08:42:07.229343 4860 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/6a0f9071-e3f7-43f2-a0ec-8ab22fd9c785-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 11 08:42:07 crc kubenswrapper[4860]: I1211 08:42:07.229356 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bm6mg\" (UniqueName: \"kubernetes.io/projected/6a0f9071-e3f7-43f2-a0ec-8ab22fd9c785-kube-api-access-bm6mg\") on node \"crc\" DevicePath \"\"" Dec 11 08:42:07 crc kubenswrapper[4860]: I1211 08:42:07.619185 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-rl5hl" event={"ID":"6a0f9071-e3f7-43f2-a0ec-8ab22fd9c785","Type":"ContainerDied","Data":"7ad58632a5c57cad8398252ac4d14597525de932102049a0d46d1146166b7ae1"} Dec 11 08:42:07 crc kubenswrapper[4860]: I1211 08:42:07.619550 4860 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7ad58632a5c57cad8398252ac4d14597525de932102049a0d46d1146166b7ae1" Dec 11 08:42:07 crc kubenswrapper[4860]: I1211 08:42:07.619269 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-rl5hl" Dec 11 08:42:07 crc kubenswrapper[4860]: I1211 08:42:07.723608 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-c6n8q"] Dec 11 08:42:07 crc kubenswrapper[4860]: E1211 08:42:07.724077 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a0f9071-e3f7-43f2-a0ec-8ab22fd9c785" containerName="ssh-known-hosts-edpm-deployment" Dec 11 08:42:07 crc kubenswrapper[4860]: I1211 08:42:07.724099 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a0f9071-e3f7-43f2-a0ec-8ab22fd9c785" containerName="ssh-known-hosts-edpm-deployment" Dec 11 08:42:07 crc kubenswrapper[4860]: I1211 08:42:07.724357 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a0f9071-e3f7-43f2-a0ec-8ab22fd9c785" containerName="ssh-known-hosts-edpm-deployment" Dec 11 08:42:07 crc kubenswrapper[4860]: I1211 08:42:07.725192 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-c6n8q" Dec 11 08:42:07 crc kubenswrapper[4860]: I1211 08:42:07.727395 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 11 08:42:07 crc kubenswrapper[4860]: I1211 08:42:07.729816 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 11 08:42:07 crc kubenswrapper[4860]: I1211 08:42:07.729996 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-7w7kh" Dec 11 08:42:07 crc kubenswrapper[4860]: I1211 08:42:07.730134 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 11 08:42:07 crc kubenswrapper[4860]: I1211 08:42:07.743355 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-c6n8q"] Dec 11 08:42:07 crc kubenswrapper[4860]: I1211 08:42:07.838765 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h7qt7\" (UniqueName: \"kubernetes.io/projected/bc86a63a-f36d-4893-a55e-df029aa586ea-kube-api-access-h7qt7\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-c6n8q\" (UID: \"bc86a63a-f36d-4893-a55e-df029aa586ea\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-c6n8q" Dec 11 08:42:07 crc kubenswrapper[4860]: I1211 08:42:07.838860 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bc86a63a-f36d-4893-a55e-df029aa586ea-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-c6n8q\" (UID: \"bc86a63a-f36d-4893-a55e-df029aa586ea\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-c6n8q" Dec 11 08:42:07 crc kubenswrapper[4860]: I1211 08:42:07.838914 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bc86a63a-f36d-4893-a55e-df029aa586ea-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-c6n8q\" (UID: \"bc86a63a-f36d-4893-a55e-df029aa586ea\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-c6n8q" Dec 11 08:42:07 crc kubenswrapper[4860]: I1211 08:42:07.940817 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bc86a63a-f36d-4893-a55e-df029aa586ea-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-c6n8q\" (UID: \"bc86a63a-f36d-4893-a55e-df029aa586ea\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-c6n8q" Dec 11 08:42:07 crc kubenswrapper[4860]: I1211 08:42:07.940905 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bc86a63a-f36d-4893-a55e-df029aa586ea-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-c6n8q\" (UID: \"bc86a63a-f36d-4893-a55e-df029aa586ea\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-c6n8q" Dec 11 08:42:07 crc kubenswrapper[4860]: I1211 08:42:07.941034 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h7qt7\" (UniqueName: \"kubernetes.io/projected/bc86a63a-f36d-4893-a55e-df029aa586ea-kube-api-access-h7qt7\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-c6n8q\" (UID: \"bc86a63a-f36d-4893-a55e-df029aa586ea\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-c6n8q" Dec 11 08:42:07 crc kubenswrapper[4860]: I1211 08:42:07.945329 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bc86a63a-f36d-4893-a55e-df029aa586ea-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-c6n8q\" (UID: \"bc86a63a-f36d-4893-a55e-df029aa586ea\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-c6n8q" Dec 11 08:42:07 crc kubenswrapper[4860]: I1211 08:42:07.945555 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bc86a63a-f36d-4893-a55e-df029aa586ea-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-c6n8q\" (UID: \"bc86a63a-f36d-4893-a55e-df029aa586ea\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-c6n8q" Dec 11 08:42:07 crc kubenswrapper[4860]: I1211 08:42:07.970198 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h7qt7\" (UniqueName: \"kubernetes.io/projected/bc86a63a-f36d-4893-a55e-df029aa586ea-kube-api-access-h7qt7\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-c6n8q\" (UID: \"bc86a63a-f36d-4893-a55e-df029aa586ea\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-c6n8q" Dec 11 08:42:08 crc kubenswrapper[4860]: I1211 08:42:08.049022 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-c6n8q" Dec 11 08:42:08 crc kubenswrapper[4860]: I1211 08:42:08.557882 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-c6n8q"] Dec 11 08:42:08 crc kubenswrapper[4860]: W1211 08:42:08.563891 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbc86a63a_f36d_4893_a55e_df029aa586ea.slice/crio-2a7b8a477fa757959432f48da8a647c148f4c2d004b711f28aaf7f31e6f487b4 WatchSource:0}: Error finding container 2a7b8a477fa757959432f48da8a647c148f4c2d004b711f28aaf7f31e6f487b4: Status 404 returned error can't find the container with id 2a7b8a477fa757959432f48da8a647c148f4c2d004b711f28aaf7f31e6f487b4 Dec 11 08:42:08 crc kubenswrapper[4860]: I1211 08:42:08.629147 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-c6n8q" event={"ID":"bc86a63a-f36d-4893-a55e-df029aa586ea","Type":"ContainerStarted","Data":"2a7b8a477fa757959432f48da8a647c148f4c2d004b711f28aaf7f31e6f487b4"} Dec 11 08:42:10 crc kubenswrapper[4860]: I1211 08:42:10.656154 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-c6n8q" event={"ID":"bc86a63a-f36d-4893-a55e-df029aa586ea","Type":"ContainerStarted","Data":"9bd30b2f5585220c76e5039076dcdfad2cae3ae5542b0cc730d70fc16a2062f0"} Dec 11 08:42:10 crc kubenswrapper[4860]: I1211 08:42:10.682936 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-c6n8q" podStartSLOduration=2.771378451 podStartE2EDuration="3.682910388s" podCreationTimestamp="2025-12-11 08:42:07 +0000 UTC" firstStartedPulling="2025-12-11 08:42:08.567025955 +0000 UTC m=+1861.295545010" lastFinishedPulling="2025-12-11 08:42:09.478557892 +0000 UTC m=+1862.207076947" observedRunningTime="2025-12-11 08:42:10.671805898 +0000 UTC m=+1863.400324953" watchObservedRunningTime="2025-12-11 08:42:10.682910388 +0000 UTC m=+1863.411429443" Dec 11 08:42:14 crc kubenswrapper[4860]: I1211 08:42:14.044615 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-7nvzj"] Dec 11 08:42:14 crc kubenswrapper[4860]: I1211 08:42:14.054844 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-7nvzj"] Dec 11 08:42:15 crc kubenswrapper[4860]: I1211 08:42:15.590440 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d07f5ac6-54a5-4490-a6b4-09fe7325e58a" path="/var/lib/kubelet/pods/d07f5ac6-54a5-4490-a6b4-09fe7325e58a/volumes" Dec 11 08:42:17 crc kubenswrapper[4860]: I1211 08:42:17.585495 4860 scope.go:117] "RemoveContainer" containerID="af76095360a11870b804fb7f16c40dad604dbf93152022249293687b1519ca1b" Dec 11 08:42:17 crc kubenswrapper[4860]: E1211 08:42:17.585981 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-99qgp_openshift-machine-config-operator(31c30642-6e60-41b4-a477-0d802424e0aa)\"" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" Dec 11 08:42:17 crc kubenswrapper[4860]: I1211 08:42:17.714402 4860 generic.go:334] "Generic (PLEG): container finished" podID="bc86a63a-f36d-4893-a55e-df029aa586ea" containerID="9bd30b2f5585220c76e5039076dcdfad2cae3ae5542b0cc730d70fc16a2062f0" exitCode=0 Dec 11 08:42:17 crc kubenswrapper[4860]: I1211 08:42:17.714452 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-c6n8q" event={"ID":"bc86a63a-f36d-4893-a55e-df029aa586ea","Type":"ContainerDied","Data":"9bd30b2f5585220c76e5039076dcdfad2cae3ae5542b0cc730d70fc16a2062f0"} Dec 11 08:42:19 crc kubenswrapper[4860]: I1211 08:42:19.126042 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-c6n8q" Dec 11 08:42:19 crc kubenswrapper[4860]: I1211 08:42:19.183894 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bc86a63a-f36d-4893-a55e-df029aa586ea-inventory\") pod \"bc86a63a-f36d-4893-a55e-df029aa586ea\" (UID: \"bc86a63a-f36d-4893-a55e-df029aa586ea\") " Dec 11 08:42:19 crc kubenswrapper[4860]: I1211 08:42:19.184082 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h7qt7\" (UniqueName: \"kubernetes.io/projected/bc86a63a-f36d-4893-a55e-df029aa586ea-kube-api-access-h7qt7\") pod \"bc86a63a-f36d-4893-a55e-df029aa586ea\" (UID: \"bc86a63a-f36d-4893-a55e-df029aa586ea\") " Dec 11 08:42:19 crc kubenswrapper[4860]: I1211 08:42:19.184170 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bc86a63a-f36d-4893-a55e-df029aa586ea-ssh-key\") pod \"bc86a63a-f36d-4893-a55e-df029aa586ea\" (UID: \"bc86a63a-f36d-4893-a55e-df029aa586ea\") " Dec 11 08:42:19 crc kubenswrapper[4860]: I1211 08:42:19.189924 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc86a63a-f36d-4893-a55e-df029aa586ea-kube-api-access-h7qt7" (OuterVolumeSpecName: "kube-api-access-h7qt7") pod "bc86a63a-f36d-4893-a55e-df029aa586ea" (UID: "bc86a63a-f36d-4893-a55e-df029aa586ea"). InnerVolumeSpecName "kube-api-access-h7qt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:42:19 crc kubenswrapper[4860]: I1211 08:42:19.210727 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc86a63a-f36d-4893-a55e-df029aa586ea-inventory" (OuterVolumeSpecName: "inventory") pod "bc86a63a-f36d-4893-a55e-df029aa586ea" (UID: "bc86a63a-f36d-4893-a55e-df029aa586ea"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:42:19 crc kubenswrapper[4860]: I1211 08:42:19.211244 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc86a63a-f36d-4893-a55e-df029aa586ea-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "bc86a63a-f36d-4893-a55e-df029aa586ea" (UID: "bc86a63a-f36d-4893-a55e-df029aa586ea"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:42:19 crc kubenswrapper[4860]: I1211 08:42:19.286440 4860 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bc86a63a-f36d-4893-a55e-df029aa586ea-inventory\") on node \"crc\" DevicePath \"\"" Dec 11 08:42:19 crc kubenswrapper[4860]: I1211 08:42:19.286477 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h7qt7\" (UniqueName: \"kubernetes.io/projected/bc86a63a-f36d-4893-a55e-df029aa586ea-kube-api-access-h7qt7\") on node \"crc\" DevicePath \"\"" Dec 11 08:42:19 crc kubenswrapper[4860]: I1211 08:42:19.286490 4860 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bc86a63a-f36d-4893-a55e-df029aa586ea-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 11 08:42:19 crc kubenswrapper[4860]: I1211 08:42:19.734043 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-c6n8q" event={"ID":"bc86a63a-f36d-4893-a55e-df029aa586ea","Type":"ContainerDied","Data":"2a7b8a477fa757959432f48da8a647c148f4c2d004b711f28aaf7f31e6f487b4"} Dec 11 08:42:19 crc kubenswrapper[4860]: I1211 08:42:19.734082 4860 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2a7b8a477fa757959432f48da8a647c148f4c2d004b711f28aaf7f31e6f487b4" Dec 11 08:42:19 crc kubenswrapper[4860]: I1211 08:42:19.734162 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-c6n8q" Dec 11 08:42:19 crc kubenswrapper[4860]: I1211 08:42:19.816852 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-g9pmk"] Dec 11 08:42:19 crc kubenswrapper[4860]: E1211 08:42:19.817406 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc86a63a-f36d-4893-a55e-df029aa586ea" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 11 08:42:19 crc kubenswrapper[4860]: I1211 08:42:19.817434 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc86a63a-f36d-4893-a55e-df029aa586ea" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 11 08:42:19 crc kubenswrapper[4860]: I1211 08:42:19.817777 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc86a63a-f36d-4893-a55e-df029aa586ea" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 11 08:42:19 crc kubenswrapper[4860]: I1211 08:42:19.818675 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-g9pmk" Dec 11 08:42:19 crc kubenswrapper[4860]: I1211 08:42:19.821913 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 11 08:42:19 crc kubenswrapper[4860]: I1211 08:42:19.829353 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-g9pmk"] Dec 11 08:42:19 crc kubenswrapper[4860]: I1211 08:42:19.834927 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 11 08:42:19 crc kubenswrapper[4860]: I1211 08:42:19.835003 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 11 08:42:19 crc kubenswrapper[4860]: I1211 08:42:19.835120 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-7w7kh" Dec 11 08:42:19 crc kubenswrapper[4860]: I1211 08:42:19.902835 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/07a4b9f9-a1a7-416d-89a5-e8a57c2484cc-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-g9pmk\" (UID: \"07a4b9f9-a1a7-416d-89a5-e8a57c2484cc\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-g9pmk" Dec 11 08:42:19 crc kubenswrapper[4860]: I1211 08:42:19.902933 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/07a4b9f9-a1a7-416d-89a5-e8a57c2484cc-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-g9pmk\" (UID: \"07a4b9f9-a1a7-416d-89a5-e8a57c2484cc\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-g9pmk" Dec 11 08:42:19 crc kubenswrapper[4860]: I1211 08:42:19.903010 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-prt8h\" (UniqueName: \"kubernetes.io/projected/07a4b9f9-a1a7-416d-89a5-e8a57c2484cc-kube-api-access-prt8h\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-g9pmk\" (UID: \"07a4b9f9-a1a7-416d-89a5-e8a57c2484cc\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-g9pmk" Dec 11 08:42:20 crc kubenswrapper[4860]: I1211 08:42:20.005403 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/07a4b9f9-a1a7-416d-89a5-e8a57c2484cc-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-g9pmk\" (UID: \"07a4b9f9-a1a7-416d-89a5-e8a57c2484cc\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-g9pmk" Dec 11 08:42:20 crc kubenswrapper[4860]: I1211 08:42:20.005534 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-prt8h\" (UniqueName: \"kubernetes.io/projected/07a4b9f9-a1a7-416d-89a5-e8a57c2484cc-kube-api-access-prt8h\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-g9pmk\" (UID: \"07a4b9f9-a1a7-416d-89a5-e8a57c2484cc\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-g9pmk" Dec 11 08:42:20 crc kubenswrapper[4860]: I1211 08:42:20.005657 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/07a4b9f9-a1a7-416d-89a5-e8a57c2484cc-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-g9pmk\" (UID: \"07a4b9f9-a1a7-416d-89a5-e8a57c2484cc\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-g9pmk" Dec 11 08:42:20 crc kubenswrapper[4860]: I1211 08:42:20.010904 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/07a4b9f9-a1a7-416d-89a5-e8a57c2484cc-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-g9pmk\" (UID: \"07a4b9f9-a1a7-416d-89a5-e8a57c2484cc\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-g9pmk" Dec 11 08:42:20 crc kubenswrapper[4860]: I1211 08:42:20.011129 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/07a4b9f9-a1a7-416d-89a5-e8a57c2484cc-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-g9pmk\" (UID: \"07a4b9f9-a1a7-416d-89a5-e8a57c2484cc\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-g9pmk" Dec 11 08:42:20 crc kubenswrapper[4860]: I1211 08:42:20.023031 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-prt8h\" (UniqueName: \"kubernetes.io/projected/07a4b9f9-a1a7-416d-89a5-e8a57c2484cc-kube-api-access-prt8h\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-g9pmk\" (UID: \"07a4b9f9-a1a7-416d-89a5-e8a57c2484cc\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-g9pmk" Dec 11 08:42:20 crc kubenswrapper[4860]: I1211 08:42:20.140745 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-g9pmk" Dec 11 08:42:20 crc kubenswrapper[4860]: I1211 08:42:20.676558 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-g9pmk"] Dec 11 08:42:20 crc kubenswrapper[4860]: I1211 08:42:20.749051 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-g9pmk" event={"ID":"07a4b9f9-a1a7-416d-89a5-e8a57c2484cc","Type":"ContainerStarted","Data":"80e3fdbcc9c49e4684e5d089d51070c0066428744a45d135d195451c40ff3ca9"} Dec 11 08:42:21 crc kubenswrapper[4860]: I1211 08:42:21.765498 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-g9pmk" event={"ID":"07a4b9f9-a1a7-416d-89a5-e8a57c2484cc","Type":"ContainerStarted","Data":"e126c02909dc19d5450e1a3c931d144c34713adf1b387de0fa1254cc98e4e216"} Dec 11 08:42:21 crc kubenswrapper[4860]: I1211 08:42:21.806594 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-g9pmk" podStartSLOduration=2.382631133 podStartE2EDuration="2.806572357s" podCreationTimestamp="2025-12-11 08:42:19 +0000 UTC" firstStartedPulling="2025-12-11 08:42:20.683038352 +0000 UTC m=+1873.411557417" lastFinishedPulling="2025-12-11 08:42:21.106979586 +0000 UTC m=+1873.835498641" observedRunningTime="2025-12-11 08:42:21.803073347 +0000 UTC m=+1874.531592412" watchObservedRunningTime="2025-12-11 08:42:21.806572357 +0000 UTC m=+1874.535091432" Dec 11 08:42:30 crc kubenswrapper[4860]: I1211 08:42:30.579305 4860 scope.go:117] "RemoveContainer" containerID="af76095360a11870b804fb7f16c40dad604dbf93152022249293687b1519ca1b" Dec 11 08:42:30 crc kubenswrapper[4860]: E1211 08:42:30.580458 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-99qgp_openshift-machine-config-operator(31c30642-6e60-41b4-a477-0d802424e0aa)\"" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" Dec 11 08:42:30 crc kubenswrapper[4860]: I1211 08:42:30.855660 4860 generic.go:334] "Generic (PLEG): container finished" podID="07a4b9f9-a1a7-416d-89a5-e8a57c2484cc" containerID="e126c02909dc19d5450e1a3c931d144c34713adf1b387de0fa1254cc98e4e216" exitCode=0 Dec 11 08:42:30 crc kubenswrapper[4860]: I1211 08:42:30.855707 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-g9pmk" event={"ID":"07a4b9f9-a1a7-416d-89a5-e8a57c2484cc","Type":"ContainerDied","Data":"e126c02909dc19d5450e1a3c931d144c34713adf1b387de0fa1254cc98e4e216"} Dec 11 08:42:32 crc kubenswrapper[4860]: I1211 08:42:32.250677 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-g9pmk" Dec 11 08:42:32 crc kubenswrapper[4860]: I1211 08:42:32.342408 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-prt8h\" (UniqueName: \"kubernetes.io/projected/07a4b9f9-a1a7-416d-89a5-e8a57c2484cc-kube-api-access-prt8h\") pod \"07a4b9f9-a1a7-416d-89a5-e8a57c2484cc\" (UID: \"07a4b9f9-a1a7-416d-89a5-e8a57c2484cc\") " Dec 11 08:42:32 crc kubenswrapper[4860]: I1211 08:42:32.342563 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/07a4b9f9-a1a7-416d-89a5-e8a57c2484cc-ssh-key\") pod \"07a4b9f9-a1a7-416d-89a5-e8a57c2484cc\" (UID: \"07a4b9f9-a1a7-416d-89a5-e8a57c2484cc\") " Dec 11 08:42:32 crc kubenswrapper[4860]: I1211 08:42:32.342587 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/07a4b9f9-a1a7-416d-89a5-e8a57c2484cc-inventory\") pod \"07a4b9f9-a1a7-416d-89a5-e8a57c2484cc\" (UID: \"07a4b9f9-a1a7-416d-89a5-e8a57c2484cc\") " Dec 11 08:42:32 crc kubenswrapper[4860]: I1211 08:42:32.348779 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/07a4b9f9-a1a7-416d-89a5-e8a57c2484cc-kube-api-access-prt8h" (OuterVolumeSpecName: "kube-api-access-prt8h") pod "07a4b9f9-a1a7-416d-89a5-e8a57c2484cc" (UID: "07a4b9f9-a1a7-416d-89a5-e8a57c2484cc"). InnerVolumeSpecName "kube-api-access-prt8h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:42:32 crc kubenswrapper[4860]: I1211 08:42:32.370214 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/07a4b9f9-a1a7-416d-89a5-e8a57c2484cc-inventory" (OuterVolumeSpecName: "inventory") pod "07a4b9f9-a1a7-416d-89a5-e8a57c2484cc" (UID: "07a4b9f9-a1a7-416d-89a5-e8a57c2484cc"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:42:32 crc kubenswrapper[4860]: I1211 08:42:32.376174 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/07a4b9f9-a1a7-416d-89a5-e8a57c2484cc-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "07a4b9f9-a1a7-416d-89a5-e8a57c2484cc" (UID: "07a4b9f9-a1a7-416d-89a5-e8a57c2484cc"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:42:32 crc kubenswrapper[4860]: I1211 08:42:32.445615 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-prt8h\" (UniqueName: \"kubernetes.io/projected/07a4b9f9-a1a7-416d-89a5-e8a57c2484cc-kube-api-access-prt8h\") on node \"crc\" DevicePath \"\"" Dec 11 08:42:32 crc kubenswrapper[4860]: I1211 08:42:32.445668 4860 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/07a4b9f9-a1a7-416d-89a5-e8a57c2484cc-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 11 08:42:32 crc kubenswrapper[4860]: I1211 08:42:32.445679 4860 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/07a4b9f9-a1a7-416d-89a5-e8a57c2484cc-inventory\") on node \"crc\" DevicePath \"\"" Dec 11 08:42:32 crc kubenswrapper[4860]: I1211 08:42:32.874275 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-g9pmk" event={"ID":"07a4b9f9-a1a7-416d-89a5-e8a57c2484cc","Type":"ContainerDied","Data":"80e3fdbcc9c49e4684e5d089d51070c0066428744a45d135d195451c40ff3ca9"} Dec 11 08:42:32 crc kubenswrapper[4860]: I1211 08:42:32.874315 4860 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="80e3fdbcc9c49e4684e5d089d51070c0066428744a45d135d195451c40ff3ca9" Dec 11 08:42:32 crc kubenswrapper[4860]: I1211 08:42:32.874380 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-g9pmk" Dec 11 08:42:32 crc kubenswrapper[4860]: I1211 08:42:32.954056 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lwc56"] Dec 11 08:42:32 crc kubenswrapper[4860]: E1211 08:42:32.954456 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07a4b9f9-a1a7-416d-89a5-e8a57c2484cc" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 11 08:42:32 crc kubenswrapper[4860]: I1211 08:42:32.954476 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="07a4b9f9-a1a7-416d-89a5-e8a57c2484cc" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 11 08:42:32 crc kubenswrapper[4860]: I1211 08:42:32.954702 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="07a4b9f9-a1a7-416d-89a5-e8a57c2484cc" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 11 08:42:32 crc kubenswrapper[4860]: I1211 08:42:32.955330 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lwc56" Dec 11 08:42:32 crc kubenswrapper[4860]: I1211 08:42:32.958129 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 11 08:42:32 crc kubenswrapper[4860]: I1211 08:42:32.959004 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Dec 11 08:42:32 crc kubenswrapper[4860]: I1211 08:42:32.959302 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 11 08:42:32 crc kubenswrapper[4860]: I1211 08:42:32.959543 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Dec 11 08:42:32 crc kubenswrapper[4860]: I1211 08:42:32.959812 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 11 08:42:32 crc kubenswrapper[4860]: I1211 08:42:32.960053 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Dec 11 08:42:32 crc kubenswrapper[4860]: I1211 08:42:32.960269 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-7w7kh" Dec 11 08:42:32 crc kubenswrapper[4860]: I1211 08:42:32.960845 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Dec 11 08:42:32 crc kubenswrapper[4860]: I1211 08:42:32.975350 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lwc56"] Dec 11 08:42:33 crc kubenswrapper[4860]: I1211 08:42:33.056150 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddaa83e7-d6db-4169-abdb-226ec9f50204-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lwc56\" (UID: \"ddaa83e7-d6db-4169-abdb-226ec9f50204\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lwc56" Dec 11 08:42:33 crc kubenswrapper[4860]: I1211 08:42:33.056193 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/ddaa83e7-d6db-4169-abdb-226ec9f50204-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lwc56\" (UID: \"ddaa83e7-d6db-4169-abdb-226ec9f50204\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lwc56" Dec 11 08:42:33 crc kubenswrapper[4860]: I1211 08:42:33.056235 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddaa83e7-d6db-4169-abdb-226ec9f50204-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lwc56\" (UID: \"ddaa83e7-d6db-4169-abdb-226ec9f50204\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lwc56" Dec 11 08:42:33 crc kubenswrapper[4860]: I1211 08:42:33.056259 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddaa83e7-d6db-4169-abdb-226ec9f50204-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lwc56\" (UID: \"ddaa83e7-d6db-4169-abdb-226ec9f50204\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lwc56" Dec 11 08:42:33 crc kubenswrapper[4860]: I1211 08:42:33.056342 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ddaa83e7-d6db-4169-abdb-226ec9f50204-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lwc56\" (UID: \"ddaa83e7-d6db-4169-abdb-226ec9f50204\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lwc56" Dec 11 08:42:33 crc kubenswrapper[4860]: I1211 08:42:33.056399 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/ddaa83e7-d6db-4169-abdb-226ec9f50204-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lwc56\" (UID: \"ddaa83e7-d6db-4169-abdb-226ec9f50204\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lwc56" Dec 11 08:42:33 crc kubenswrapper[4860]: I1211 08:42:33.056471 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4bf7b\" (UniqueName: \"kubernetes.io/projected/ddaa83e7-d6db-4169-abdb-226ec9f50204-kube-api-access-4bf7b\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lwc56\" (UID: \"ddaa83e7-d6db-4169-abdb-226ec9f50204\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lwc56" Dec 11 08:42:33 crc kubenswrapper[4860]: I1211 08:42:33.056507 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/ddaa83e7-d6db-4169-abdb-226ec9f50204-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lwc56\" (UID: \"ddaa83e7-d6db-4169-abdb-226ec9f50204\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lwc56" Dec 11 08:42:33 crc kubenswrapper[4860]: I1211 08:42:33.056683 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddaa83e7-d6db-4169-abdb-226ec9f50204-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lwc56\" (UID: \"ddaa83e7-d6db-4169-abdb-226ec9f50204\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lwc56" Dec 11 08:42:33 crc kubenswrapper[4860]: I1211 08:42:33.056844 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddaa83e7-d6db-4169-abdb-226ec9f50204-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lwc56\" (UID: \"ddaa83e7-d6db-4169-abdb-226ec9f50204\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lwc56" Dec 11 08:42:33 crc kubenswrapper[4860]: I1211 08:42:33.056920 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ddaa83e7-d6db-4169-abdb-226ec9f50204-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lwc56\" (UID: \"ddaa83e7-d6db-4169-abdb-226ec9f50204\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lwc56" Dec 11 08:42:33 crc kubenswrapper[4860]: I1211 08:42:33.056997 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/ddaa83e7-d6db-4169-abdb-226ec9f50204-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lwc56\" (UID: \"ddaa83e7-d6db-4169-abdb-226ec9f50204\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lwc56" Dec 11 08:42:33 crc kubenswrapper[4860]: I1211 08:42:33.057031 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddaa83e7-d6db-4169-abdb-226ec9f50204-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lwc56\" (UID: \"ddaa83e7-d6db-4169-abdb-226ec9f50204\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lwc56" Dec 11 08:42:33 crc kubenswrapper[4860]: I1211 08:42:33.057054 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddaa83e7-d6db-4169-abdb-226ec9f50204-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lwc56\" (UID: \"ddaa83e7-d6db-4169-abdb-226ec9f50204\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lwc56" Dec 11 08:42:33 crc kubenswrapper[4860]: I1211 08:42:33.158426 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ddaa83e7-d6db-4169-abdb-226ec9f50204-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lwc56\" (UID: \"ddaa83e7-d6db-4169-abdb-226ec9f50204\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lwc56" Dec 11 08:42:33 crc kubenswrapper[4860]: I1211 08:42:33.158503 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/ddaa83e7-d6db-4169-abdb-226ec9f50204-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lwc56\" (UID: \"ddaa83e7-d6db-4169-abdb-226ec9f50204\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lwc56" Dec 11 08:42:33 crc kubenswrapper[4860]: I1211 08:42:33.158534 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddaa83e7-d6db-4169-abdb-226ec9f50204-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lwc56\" (UID: \"ddaa83e7-d6db-4169-abdb-226ec9f50204\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lwc56" Dec 11 08:42:33 crc kubenswrapper[4860]: I1211 08:42:33.158568 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddaa83e7-d6db-4169-abdb-226ec9f50204-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lwc56\" (UID: \"ddaa83e7-d6db-4169-abdb-226ec9f50204\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lwc56" Dec 11 08:42:33 crc kubenswrapper[4860]: I1211 08:42:33.158601 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddaa83e7-d6db-4169-abdb-226ec9f50204-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lwc56\" (UID: \"ddaa83e7-d6db-4169-abdb-226ec9f50204\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lwc56" Dec 11 08:42:33 crc kubenswrapper[4860]: I1211 08:42:33.158633 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/ddaa83e7-d6db-4169-abdb-226ec9f50204-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lwc56\" (UID: \"ddaa83e7-d6db-4169-abdb-226ec9f50204\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lwc56" Dec 11 08:42:33 crc kubenswrapper[4860]: I1211 08:42:33.158704 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddaa83e7-d6db-4169-abdb-226ec9f50204-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lwc56\" (UID: \"ddaa83e7-d6db-4169-abdb-226ec9f50204\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lwc56" Dec 11 08:42:33 crc kubenswrapper[4860]: I1211 08:42:33.158732 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddaa83e7-d6db-4169-abdb-226ec9f50204-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lwc56\" (UID: \"ddaa83e7-d6db-4169-abdb-226ec9f50204\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lwc56" Dec 11 08:42:33 crc kubenswrapper[4860]: I1211 08:42:33.158756 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ddaa83e7-d6db-4169-abdb-226ec9f50204-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lwc56\" (UID: \"ddaa83e7-d6db-4169-abdb-226ec9f50204\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lwc56" Dec 11 08:42:33 crc kubenswrapper[4860]: I1211 08:42:33.158787 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/ddaa83e7-d6db-4169-abdb-226ec9f50204-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lwc56\" (UID: \"ddaa83e7-d6db-4169-abdb-226ec9f50204\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lwc56" Dec 11 08:42:33 crc kubenswrapper[4860]: I1211 08:42:33.158864 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4bf7b\" (UniqueName: \"kubernetes.io/projected/ddaa83e7-d6db-4169-abdb-226ec9f50204-kube-api-access-4bf7b\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lwc56\" (UID: \"ddaa83e7-d6db-4169-abdb-226ec9f50204\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lwc56" Dec 11 08:42:33 crc kubenswrapper[4860]: I1211 08:42:33.158893 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/ddaa83e7-d6db-4169-abdb-226ec9f50204-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lwc56\" (UID: \"ddaa83e7-d6db-4169-abdb-226ec9f50204\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lwc56" Dec 11 08:42:33 crc kubenswrapper[4860]: I1211 08:42:33.158937 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddaa83e7-d6db-4169-abdb-226ec9f50204-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lwc56\" (UID: \"ddaa83e7-d6db-4169-abdb-226ec9f50204\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lwc56" Dec 11 08:42:33 crc kubenswrapper[4860]: I1211 08:42:33.158992 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddaa83e7-d6db-4169-abdb-226ec9f50204-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lwc56\" (UID: \"ddaa83e7-d6db-4169-abdb-226ec9f50204\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lwc56" Dec 11 08:42:33 crc kubenswrapper[4860]: I1211 08:42:33.164765 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddaa83e7-d6db-4169-abdb-226ec9f50204-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lwc56\" (UID: \"ddaa83e7-d6db-4169-abdb-226ec9f50204\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lwc56" Dec 11 08:42:33 crc kubenswrapper[4860]: I1211 08:42:33.164993 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddaa83e7-d6db-4169-abdb-226ec9f50204-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lwc56\" (UID: \"ddaa83e7-d6db-4169-abdb-226ec9f50204\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lwc56" Dec 11 08:42:33 crc kubenswrapper[4860]: I1211 08:42:33.165258 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/ddaa83e7-d6db-4169-abdb-226ec9f50204-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lwc56\" (UID: \"ddaa83e7-d6db-4169-abdb-226ec9f50204\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lwc56" Dec 11 08:42:33 crc kubenswrapper[4860]: I1211 08:42:33.165683 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddaa83e7-d6db-4169-abdb-226ec9f50204-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lwc56\" (UID: \"ddaa83e7-d6db-4169-abdb-226ec9f50204\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lwc56" Dec 11 08:42:33 crc kubenswrapper[4860]: I1211 08:42:33.166404 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddaa83e7-d6db-4169-abdb-226ec9f50204-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lwc56\" (UID: \"ddaa83e7-d6db-4169-abdb-226ec9f50204\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lwc56" Dec 11 08:42:33 crc kubenswrapper[4860]: I1211 08:42:33.167518 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddaa83e7-d6db-4169-abdb-226ec9f50204-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lwc56\" (UID: \"ddaa83e7-d6db-4169-abdb-226ec9f50204\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lwc56" Dec 11 08:42:33 crc kubenswrapper[4860]: I1211 08:42:33.167789 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ddaa83e7-d6db-4169-abdb-226ec9f50204-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lwc56\" (UID: \"ddaa83e7-d6db-4169-abdb-226ec9f50204\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lwc56" Dec 11 08:42:33 crc kubenswrapper[4860]: I1211 08:42:33.168962 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddaa83e7-d6db-4169-abdb-226ec9f50204-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lwc56\" (UID: \"ddaa83e7-d6db-4169-abdb-226ec9f50204\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lwc56" Dec 11 08:42:33 crc kubenswrapper[4860]: I1211 08:42:33.169395 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/ddaa83e7-d6db-4169-abdb-226ec9f50204-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lwc56\" (UID: \"ddaa83e7-d6db-4169-abdb-226ec9f50204\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lwc56" Dec 11 08:42:33 crc kubenswrapper[4860]: I1211 08:42:33.169946 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/ddaa83e7-d6db-4169-abdb-226ec9f50204-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lwc56\" (UID: \"ddaa83e7-d6db-4169-abdb-226ec9f50204\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lwc56" Dec 11 08:42:33 crc kubenswrapper[4860]: I1211 08:42:33.169994 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddaa83e7-d6db-4169-abdb-226ec9f50204-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lwc56\" (UID: \"ddaa83e7-d6db-4169-abdb-226ec9f50204\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lwc56" Dec 11 08:42:33 crc kubenswrapper[4860]: I1211 08:42:33.173251 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/ddaa83e7-d6db-4169-abdb-226ec9f50204-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lwc56\" (UID: \"ddaa83e7-d6db-4169-abdb-226ec9f50204\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lwc56" Dec 11 08:42:33 crc kubenswrapper[4860]: I1211 08:42:33.176001 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ddaa83e7-d6db-4169-abdb-226ec9f50204-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lwc56\" (UID: \"ddaa83e7-d6db-4169-abdb-226ec9f50204\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lwc56" Dec 11 08:42:33 crc kubenswrapper[4860]: I1211 08:42:33.180025 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4bf7b\" (UniqueName: \"kubernetes.io/projected/ddaa83e7-d6db-4169-abdb-226ec9f50204-kube-api-access-4bf7b\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lwc56\" (UID: \"ddaa83e7-d6db-4169-abdb-226ec9f50204\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lwc56" Dec 11 08:42:33 crc kubenswrapper[4860]: I1211 08:42:33.278075 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lwc56" Dec 11 08:42:33 crc kubenswrapper[4860]: I1211 08:42:33.305424 4860 scope.go:117] "RemoveContainer" containerID="33f54983234fbd4f8d8f95d07373fa5a12a595642d29f4c19fb280a7e225d5ab" Dec 11 08:42:33 crc kubenswrapper[4860]: I1211 08:42:33.833613 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lwc56"] Dec 11 08:42:33 crc kubenswrapper[4860]: I1211 08:42:33.898770 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lwc56" event={"ID":"ddaa83e7-d6db-4169-abdb-226ec9f50204","Type":"ContainerStarted","Data":"1c9734d626d729246e8686500208dac2b1a878f71cce3c929fcd55883d1be9dd"} Dec 11 08:42:34 crc kubenswrapper[4860]: I1211 08:42:34.908377 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lwc56" event={"ID":"ddaa83e7-d6db-4169-abdb-226ec9f50204","Type":"ContainerStarted","Data":"b3b7401f4de64349c32d09f85e57aa88721269e5f47b8b35ef7dfe016753aad3"} Dec 11 08:42:34 crc kubenswrapper[4860]: I1211 08:42:34.941507 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lwc56" podStartSLOduration=2.390020503 podStartE2EDuration="2.941478915s" podCreationTimestamp="2025-12-11 08:42:32 +0000 UTC" firstStartedPulling="2025-12-11 08:42:33.838909531 +0000 UTC m=+1886.567428586" lastFinishedPulling="2025-12-11 08:42:34.390367933 +0000 UTC m=+1887.118886998" observedRunningTime="2025-12-11 08:42:34.928428695 +0000 UTC m=+1887.656947790" watchObservedRunningTime="2025-12-11 08:42:34.941478915 +0000 UTC m=+1887.669998010" Dec 11 08:42:45 crc kubenswrapper[4860]: I1211 08:42:45.579705 4860 scope.go:117] "RemoveContainer" containerID="af76095360a11870b804fb7f16c40dad604dbf93152022249293687b1519ca1b" Dec 11 08:42:45 crc kubenswrapper[4860]: E1211 08:42:45.580510 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-99qgp_openshift-machine-config-operator(31c30642-6e60-41b4-a477-0d802424e0aa)\"" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" Dec 11 08:43:00 crc kubenswrapper[4860]: I1211 08:43:00.578929 4860 scope.go:117] "RemoveContainer" containerID="af76095360a11870b804fb7f16c40dad604dbf93152022249293687b1519ca1b" Dec 11 08:43:00 crc kubenswrapper[4860]: E1211 08:43:00.579667 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-99qgp_openshift-machine-config-operator(31c30642-6e60-41b4-a477-0d802424e0aa)\"" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" Dec 11 08:43:11 crc kubenswrapper[4860]: I1211 08:43:11.227656 4860 generic.go:334] "Generic (PLEG): container finished" podID="ddaa83e7-d6db-4169-abdb-226ec9f50204" containerID="b3b7401f4de64349c32d09f85e57aa88721269e5f47b8b35ef7dfe016753aad3" exitCode=0 Dec 11 08:43:11 crc kubenswrapper[4860]: I1211 08:43:11.228154 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lwc56" event={"ID":"ddaa83e7-d6db-4169-abdb-226ec9f50204","Type":"ContainerDied","Data":"b3b7401f4de64349c32d09f85e57aa88721269e5f47b8b35ef7dfe016753aad3"} Dec 11 08:43:12 crc kubenswrapper[4860]: I1211 08:43:12.643588 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lwc56" Dec 11 08:43:12 crc kubenswrapper[4860]: I1211 08:43:12.786189 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/ddaa83e7-d6db-4169-abdb-226ec9f50204-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"ddaa83e7-d6db-4169-abdb-226ec9f50204\" (UID: \"ddaa83e7-d6db-4169-abdb-226ec9f50204\") " Dec 11 08:43:12 crc kubenswrapper[4860]: I1211 08:43:12.786285 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddaa83e7-d6db-4169-abdb-226ec9f50204-repo-setup-combined-ca-bundle\") pod \"ddaa83e7-d6db-4169-abdb-226ec9f50204\" (UID: \"ddaa83e7-d6db-4169-abdb-226ec9f50204\") " Dec 11 08:43:12 crc kubenswrapper[4860]: I1211 08:43:12.786359 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/ddaa83e7-d6db-4169-abdb-226ec9f50204-openstack-edpm-ipam-ovn-default-certs-0\") pod \"ddaa83e7-d6db-4169-abdb-226ec9f50204\" (UID: \"ddaa83e7-d6db-4169-abdb-226ec9f50204\") " Dec 11 08:43:12 crc kubenswrapper[4860]: I1211 08:43:12.786407 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddaa83e7-d6db-4169-abdb-226ec9f50204-bootstrap-combined-ca-bundle\") pod \"ddaa83e7-d6db-4169-abdb-226ec9f50204\" (UID: \"ddaa83e7-d6db-4169-abdb-226ec9f50204\") " Dec 11 08:43:12 crc kubenswrapper[4860]: I1211 08:43:12.786430 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ddaa83e7-d6db-4169-abdb-226ec9f50204-ssh-key\") pod \"ddaa83e7-d6db-4169-abdb-226ec9f50204\" (UID: \"ddaa83e7-d6db-4169-abdb-226ec9f50204\") " Dec 11 08:43:12 crc kubenswrapper[4860]: I1211 08:43:12.786478 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddaa83e7-d6db-4169-abdb-226ec9f50204-ovn-combined-ca-bundle\") pod \"ddaa83e7-d6db-4169-abdb-226ec9f50204\" (UID: \"ddaa83e7-d6db-4169-abdb-226ec9f50204\") " Dec 11 08:43:12 crc kubenswrapper[4860]: I1211 08:43:12.786504 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddaa83e7-d6db-4169-abdb-226ec9f50204-nova-combined-ca-bundle\") pod \"ddaa83e7-d6db-4169-abdb-226ec9f50204\" (UID: \"ddaa83e7-d6db-4169-abdb-226ec9f50204\") " Dec 11 08:43:12 crc kubenswrapper[4860]: I1211 08:43:12.786565 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/ddaa83e7-d6db-4169-abdb-226ec9f50204-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"ddaa83e7-d6db-4169-abdb-226ec9f50204\" (UID: \"ddaa83e7-d6db-4169-abdb-226ec9f50204\") " Dec 11 08:43:12 crc kubenswrapper[4860]: I1211 08:43:12.786583 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ddaa83e7-d6db-4169-abdb-226ec9f50204-inventory\") pod \"ddaa83e7-d6db-4169-abdb-226ec9f50204\" (UID: \"ddaa83e7-d6db-4169-abdb-226ec9f50204\") " Dec 11 08:43:12 crc kubenswrapper[4860]: I1211 08:43:12.786601 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddaa83e7-d6db-4169-abdb-226ec9f50204-telemetry-combined-ca-bundle\") pod \"ddaa83e7-d6db-4169-abdb-226ec9f50204\" (UID: \"ddaa83e7-d6db-4169-abdb-226ec9f50204\") " Dec 11 08:43:12 crc kubenswrapper[4860]: I1211 08:43:12.786629 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4bf7b\" (UniqueName: \"kubernetes.io/projected/ddaa83e7-d6db-4169-abdb-226ec9f50204-kube-api-access-4bf7b\") pod \"ddaa83e7-d6db-4169-abdb-226ec9f50204\" (UID: \"ddaa83e7-d6db-4169-abdb-226ec9f50204\") " Dec 11 08:43:12 crc kubenswrapper[4860]: I1211 08:43:12.786682 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/ddaa83e7-d6db-4169-abdb-226ec9f50204-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"ddaa83e7-d6db-4169-abdb-226ec9f50204\" (UID: \"ddaa83e7-d6db-4169-abdb-226ec9f50204\") " Dec 11 08:43:12 crc kubenswrapper[4860]: I1211 08:43:12.786712 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddaa83e7-d6db-4169-abdb-226ec9f50204-neutron-metadata-combined-ca-bundle\") pod \"ddaa83e7-d6db-4169-abdb-226ec9f50204\" (UID: \"ddaa83e7-d6db-4169-abdb-226ec9f50204\") " Dec 11 08:43:12 crc kubenswrapper[4860]: I1211 08:43:12.786739 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddaa83e7-d6db-4169-abdb-226ec9f50204-libvirt-combined-ca-bundle\") pod \"ddaa83e7-d6db-4169-abdb-226ec9f50204\" (UID: \"ddaa83e7-d6db-4169-abdb-226ec9f50204\") " Dec 11 08:43:12 crc kubenswrapper[4860]: I1211 08:43:12.792859 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ddaa83e7-d6db-4169-abdb-226ec9f50204-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "ddaa83e7-d6db-4169-abdb-226ec9f50204" (UID: "ddaa83e7-d6db-4169-abdb-226ec9f50204"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:43:12 crc kubenswrapper[4860]: I1211 08:43:12.793224 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ddaa83e7-d6db-4169-abdb-226ec9f50204-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "ddaa83e7-d6db-4169-abdb-226ec9f50204" (UID: "ddaa83e7-d6db-4169-abdb-226ec9f50204"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:43:12 crc kubenswrapper[4860]: I1211 08:43:12.794169 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ddaa83e7-d6db-4169-abdb-226ec9f50204-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "ddaa83e7-d6db-4169-abdb-226ec9f50204" (UID: "ddaa83e7-d6db-4169-abdb-226ec9f50204"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:43:12 crc kubenswrapper[4860]: I1211 08:43:12.794250 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ddaa83e7-d6db-4169-abdb-226ec9f50204-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "ddaa83e7-d6db-4169-abdb-226ec9f50204" (UID: "ddaa83e7-d6db-4169-abdb-226ec9f50204"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:43:12 crc kubenswrapper[4860]: I1211 08:43:12.794389 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ddaa83e7-d6db-4169-abdb-226ec9f50204-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "ddaa83e7-d6db-4169-abdb-226ec9f50204" (UID: "ddaa83e7-d6db-4169-abdb-226ec9f50204"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:43:12 crc kubenswrapper[4860]: I1211 08:43:12.794747 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ddaa83e7-d6db-4169-abdb-226ec9f50204-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "ddaa83e7-d6db-4169-abdb-226ec9f50204" (UID: "ddaa83e7-d6db-4169-abdb-226ec9f50204"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:43:12 crc kubenswrapper[4860]: I1211 08:43:12.794822 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ddaa83e7-d6db-4169-abdb-226ec9f50204-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "ddaa83e7-d6db-4169-abdb-226ec9f50204" (UID: "ddaa83e7-d6db-4169-abdb-226ec9f50204"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:43:12 crc kubenswrapper[4860]: I1211 08:43:12.795012 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ddaa83e7-d6db-4169-abdb-226ec9f50204-kube-api-access-4bf7b" (OuterVolumeSpecName: "kube-api-access-4bf7b") pod "ddaa83e7-d6db-4169-abdb-226ec9f50204" (UID: "ddaa83e7-d6db-4169-abdb-226ec9f50204"). InnerVolumeSpecName "kube-api-access-4bf7b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:43:12 crc kubenswrapper[4860]: I1211 08:43:12.796003 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ddaa83e7-d6db-4169-abdb-226ec9f50204-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "ddaa83e7-d6db-4169-abdb-226ec9f50204" (UID: "ddaa83e7-d6db-4169-abdb-226ec9f50204"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:43:12 crc kubenswrapper[4860]: I1211 08:43:12.796319 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ddaa83e7-d6db-4169-abdb-226ec9f50204-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "ddaa83e7-d6db-4169-abdb-226ec9f50204" (UID: "ddaa83e7-d6db-4169-abdb-226ec9f50204"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:43:12 crc kubenswrapper[4860]: I1211 08:43:12.798831 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ddaa83e7-d6db-4169-abdb-226ec9f50204-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "ddaa83e7-d6db-4169-abdb-226ec9f50204" (UID: "ddaa83e7-d6db-4169-abdb-226ec9f50204"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:43:12 crc kubenswrapper[4860]: I1211 08:43:12.799383 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ddaa83e7-d6db-4169-abdb-226ec9f50204-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "ddaa83e7-d6db-4169-abdb-226ec9f50204" (UID: "ddaa83e7-d6db-4169-abdb-226ec9f50204"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:43:12 crc kubenswrapper[4860]: I1211 08:43:12.824300 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ddaa83e7-d6db-4169-abdb-226ec9f50204-inventory" (OuterVolumeSpecName: "inventory") pod "ddaa83e7-d6db-4169-abdb-226ec9f50204" (UID: "ddaa83e7-d6db-4169-abdb-226ec9f50204"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:43:12 crc kubenswrapper[4860]: I1211 08:43:12.828142 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ddaa83e7-d6db-4169-abdb-226ec9f50204-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ddaa83e7-d6db-4169-abdb-226ec9f50204" (UID: "ddaa83e7-d6db-4169-abdb-226ec9f50204"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:43:12 crc kubenswrapper[4860]: I1211 08:43:12.888815 4860 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/ddaa83e7-d6db-4169-abdb-226ec9f50204-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 11 08:43:12 crc kubenswrapper[4860]: I1211 08:43:12.888855 4860 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddaa83e7-d6db-4169-abdb-226ec9f50204-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 08:43:12 crc kubenswrapper[4860]: I1211 08:43:12.888868 4860 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ddaa83e7-d6db-4169-abdb-226ec9f50204-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 11 08:43:12 crc kubenswrapper[4860]: I1211 08:43:12.888879 4860 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddaa83e7-d6db-4169-abdb-226ec9f50204-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 08:43:12 crc kubenswrapper[4860]: I1211 08:43:12.888887 4860 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddaa83e7-d6db-4169-abdb-226ec9f50204-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 08:43:12 crc kubenswrapper[4860]: I1211 08:43:12.888896 4860 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/ddaa83e7-d6db-4169-abdb-226ec9f50204-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 11 08:43:12 crc kubenswrapper[4860]: I1211 08:43:12.888905 4860 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ddaa83e7-d6db-4169-abdb-226ec9f50204-inventory\") on node \"crc\" DevicePath \"\"" Dec 11 08:43:12 crc kubenswrapper[4860]: I1211 08:43:12.888913 4860 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddaa83e7-d6db-4169-abdb-226ec9f50204-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 08:43:12 crc kubenswrapper[4860]: I1211 08:43:12.888924 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4bf7b\" (UniqueName: \"kubernetes.io/projected/ddaa83e7-d6db-4169-abdb-226ec9f50204-kube-api-access-4bf7b\") on node \"crc\" DevicePath \"\"" Dec 11 08:43:12 crc kubenswrapper[4860]: I1211 08:43:12.888934 4860 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/ddaa83e7-d6db-4169-abdb-226ec9f50204-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 11 08:43:12 crc kubenswrapper[4860]: I1211 08:43:12.888942 4860 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddaa83e7-d6db-4169-abdb-226ec9f50204-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 08:43:12 crc kubenswrapper[4860]: I1211 08:43:12.888951 4860 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddaa83e7-d6db-4169-abdb-226ec9f50204-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 08:43:12 crc kubenswrapper[4860]: I1211 08:43:12.888959 4860 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/ddaa83e7-d6db-4169-abdb-226ec9f50204-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 11 08:43:12 crc kubenswrapper[4860]: I1211 08:43:12.888970 4860 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddaa83e7-d6db-4169-abdb-226ec9f50204-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 08:43:13 crc kubenswrapper[4860]: I1211 08:43:13.246984 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lwc56" event={"ID":"ddaa83e7-d6db-4169-abdb-226ec9f50204","Type":"ContainerDied","Data":"1c9734d626d729246e8686500208dac2b1a878f71cce3c929fcd55883d1be9dd"} Dec 11 08:43:13 crc kubenswrapper[4860]: I1211 08:43:13.247039 4860 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1c9734d626d729246e8686500208dac2b1a878f71cce3c929fcd55883d1be9dd" Dec 11 08:43:13 crc kubenswrapper[4860]: I1211 08:43:13.247053 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lwc56" Dec 11 08:43:13 crc kubenswrapper[4860]: I1211 08:43:13.344712 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-8ptks"] Dec 11 08:43:13 crc kubenswrapper[4860]: E1211 08:43:13.345275 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ddaa83e7-d6db-4169-abdb-226ec9f50204" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 11 08:43:13 crc kubenswrapper[4860]: I1211 08:43:13.345298 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="ddaa83e7-d6db-4169-abdb-226ec9f50204" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 11 08:43:13 crc kubenswrapper[4860]: I1211 08:43:13.345626 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="ddaa83e7-d6db-4169-abdb-226ec9f50204" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 11 08:43:13 crc kubenswrapper[4860]: I1211 08:43:13.346440 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8ptks" Dec 11 08:43:13 crc kubenswrapper[4860]: I1211 08:43:13.348044 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 11 08:43:13 crc kubenswrapper[4860]: I1211 08:43:13.350058 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 11 08:43:13 crc kubenswrapper[4860]: I1211 08:43:13.350241 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Dec 11 08:43:13 crc kubenswrapper[4860]: I1211 08:43:13.350399 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-7w7kh" Dec 11 08:43:13 crc kubenswrapper[4860]: I1211 08:43:13.350514 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 11 08:43:13 crc kubenswrapper[4860]: I1211 08:43:13.355040 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-8ptks"] Dec 11 08:43:13 crc kubenswrapper[4860]: I1211 08:43:13.500991 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9jkl9\" (UniqueName: \"kubernetes.io/projected/c2dda62c-b6df-481f-a388-a6bfc3b02c9e-kube-api-access-9jkl9\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-8ptks\" (UID: \"c2dda62c-b6df-481f-a388-a6bfc3b02c9e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8ptks" Dec 11 08:43:13 crc kubenswrapper[4860]: I1211 08:43:13.501420 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2dda62c-b6df-481f-a388-a6bfc3b02c9e-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-8ptks\" (UID: \"c2dda62c-b6df-481f-a388-a6bfc3b02c9e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8ptks" Dec 11 08:43:13 crc kubenswrapper[4860]: I1211 08:43:13.501470 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/c2dda62c-b6df-481f-a388-a6bfc3b02c9e-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-8ptks\" (UID: \"c2dda62c-b6df-481f-a388-a6bfc3b02c9e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8ptks" Dec 11 08:43:13 crc kubenswrapper[4860]: I1211 08:43:13.501498 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c2dda62c-b6df-481f-a388-a6bfc3b02c9e-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-8ptks\" (UID: \"c2dda62c-b6df-481f-a388-a6bfc3b02c9e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8ptks" Dec 11 08:43:13 crc kubenswrapper[4860]: I1211 08:43:13.501705 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c2dda62c-b6df-481f-a388-a6bfc3b02c9e-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-8ptks\" (UID: \"c2dda62c-b6df-481f-a388-a6bfc3b02c9e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8ptks" Dec 11 08:43:13 crc kubenswrapper[4860]: I1211 08:43:13.581193 4860 scope.go:117] "RemoveContainer" containerID="af76095360a11870b804fb7f16c40dad604dbf93152022249293687b1519ca1b" Dec 11 08:43:13 crc kubenswrapper[4860]: E1211 08:43:13.581751 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-99qgp_openshift-machine-config-operator(31c30642-6e60-41b4-a477-0d802424e0aa)\"" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" Dec 11 08:43:13 crc kubenswrapper[4860]: I1211 08:43:13.603495 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c2dda62c-b6df-481f-a388-a6bfc3b02c9e-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-8ptks\" (UID: \"c2dda62c-b6df-481f-a388-a6bfc3b02c9e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8ptks" Dec 11 08:43:13 crc kubenswrapper[4860]: I1211 08:43:13.603687 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9jkl9\" (UniqueName: \"kubernetes.io/projected/c2dda62c-b6df-481f-a388-a6bfc3b02c9e-kube-api-access-9jkl9\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-8ptks\" (UID: \"c2dda62c-b6df-481f-a388-a6bfc3b02c9e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8ptks" Dec 11 08:43:13 crc kubenswrapper[4860]: I1211 08:43:13.603787 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2dda62c-b6df-481f-a388-a6bfc3b02c9e-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-8ptks\" (UID: \"c2dda62c-b6df-481f-a388-a6bfc3b02c9e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8ptks" Dec 11 08:43:13 crc kubenswrapper[4860]: I1211 08:43:13.603861 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/c2dda62c-b6df-481f-a388-a6bfc3b02c9e-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-8ptks\" (UID: \"c2dda62c-b6df-481f-a388-a6bfc3b02c9e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8ptks" Dec 11 08:43:13 crc kubenswrapper[4860]: I1211 08:43:13.603903 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c2dda62c-b6df-481f-a388-a6bfc3b02c9e-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-8ptks\" (UID: \"c2dda62c-b6df-481f-a388-a6bfc3b02c9e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8ptks" Dec 11 08:43:13 crc kubenswrapper[4860]: I1211 08:43:13.606086 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/c2dda62c-b6df-481f-a388-a6bfc3b02c9e-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-8ptks\" (UID: \"c2dda62c-b6df-481f-a388-a6bfc3b02c9e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8ptks" Dec 11 08:43:13 crc kubenswrapper[4860]: I1211 08:43:13.612144 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c2dda62c-b6df-481f-a388-a6bfc3b02c9e-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-8ptks\" (UID: \"c2dda62c-b6df-481f-a388-a6bfc3b02c9e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8ptks" Dec 11 08:43:13 crc kubenswrapper[4860]: I1211 08:43:13.612211 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2dda62c-b6df-481f-a388-a6bfc3b02c9e-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-8ptks\" (UID: \"c2dda62c-b6df-481f-a388-a6bfc3b02c9e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8ptks" Dec 11 08:43:13 crc kubenswrapper[4860]: I1211 08:43:13.612535 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c2dda62c-b6df-481f-a388-a6bfc3b02c9e-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-8ptks\" (UID: \"c2dda62c-b6df-481f-a388-a6bfc3b02c9e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8ptks" Dec 11 08:43:13 crc kubenswrapper[4860]: I1211 08:43:13.627794 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9jkl9\" (UniqueName: \"kubernetes.io/projected/c2dda62c-b6df-481f-a388-a6bfc3b02c9e-kube-api-access-9jkl9\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-8ptks\" (UID: \"c2dda62c-b6df-481f-a388-a6bfc3b02c9e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8ptks" Dec 11 08:43:13 crc kubenswrapper[4860]: I1211 08:43:13.685866 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8ptks" Dec 11 08:43:14 crc kubenswrapper[4860]: I1211 08:43:14.212944 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-8ptks"] Dec 11 08:43:14 crc kubenswrapper[4860]: I1211 08:43:14.256930 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8ptks" event={"ID":"c2dda62c-b6df-481f-a388-a6bfc3b02c9e","Type":"ContainerStarted","Data":"b2d07f0f659856caab257bb44bbd39ed58fba2f6278078e1b038b288c59d2a63"} Dec 11 08:43:16 crc kubenswrapper[4860]: I1211 08:43:16.277551 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8ptks" event={"ID":"c2dda62c-b6df-481f-a388-a6bfc3b02c9e","Type":"ContainerStarted","Data":"d99ed587652cc6cd5ae66ffb6d94936cdd900fa7764c06adda63f7fcb5f1aca3"} Dec 11 08:43:16 crc kubenswrapper[4860]: I1211 08:43:16.307401 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8ptks" podStartSLOduration=2.431481715 podStartE2EDuration="3.307383051s" podCreationTimestamp="2025-12-11 08:43:13 +0000 UTC" firstStartedPulling="2025-12-11 08:43:14.21825963 +0000 UTC m=+1926.946778685" lastFinishedPulling="2025-12-11 08:43:15.094160966 +0000 UTC m=+1927.822680021" observedRunningTime="2025-12-11 08:43:16.297714228 +0000 UTC m=+1929.026233293" watchObservedRunningTime="2025-12-11 08:43:16.307383051 +0000 UTC m=+1929.035902106" Dec 11 08:43:26 crc kubenswrapper[4860]: I1211 08:43:26.579738 4860 scope.go:117] "RemoveContainer" containerID="af76095360a11870b804fb7f16c40dad604dbf93152022249293687b1519ca1b" Dec 11 08:43:26 crc kubenswrapper[4860]: E1211 08:43:26.580559 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-99qgp_openshift-machine-config-operator(31c30642-6e60-41b4-a477-0d802424e0aa)\"" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" Dec 11 08:43:38 crc kubenswrapper[4860]: I1211 08:43:38.579343 4860 scope.go:117] "RemoveContainer" containerID="af76095360a11870b804fb7f16c40dad604dbf93152022249293687b1519ca1b" Dec 11 08:43:38 crc kubenswrapper[4860]: E1211 08:43:38.580053 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-99qgp_openshift-machine-config-operator(31c30642-6e60-41b4-a477-0d802424e0aa)\"" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" Dec 11 08:43:52 crc kubenswrapper[4860]: I1211 08:43:52.578949 4860 scope.go:117] "RemoveContainer" containerID="af76095360a11870b804fb7f16c40dad604dbf93152022249293687b1519ca1b" Dec 11 08:43:52 crc kubenswrapper[4860]: E1211 08:43:52.579779 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-99qgp_openshift-machine-config-operator(31c30642-6e60-41b4-a477-0d802424e0aa)\"" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" Dec 11 08:44:06 crc kubenswrapper[4860]: I1211 08:44:06.579546 4860 scope.go:117] "RemoveContainer" containerID="af76095360a11870b804fb7f16c40dad604dbf93152022249293687b1519ca1b" Dec 11 08:44:06 crc kubenswrapper[4860]: E1211 08:44:06.580212 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-99qgp_openshift-machine-config-operator(31c30642-6e60-41b4-a477-0d802424e0aa)\"" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" Dec 11 08:44:13 crc kubenswrapper[4860]: I1211 08:44:13.818993 4860 generic.go:334] "Generic (PLEG): container finished" podID="c2dda62c-b6df-481f-a388-a6bfc3b02c9e" containerID="d99ed587652cc6cd5ae66ffb6d94936cdd900fa7764c06adda63f7fcb5f1aca3" exitCode=0 Dec 11 08:44:13 crc kubenswrapper[4860]: I1211 08:44:13.819089 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8ptks" event={"ID":"c2dda62c-b6df-481f-a388-a6bfc3b02c9e","Type":"ContainerDied","Data":"d99ed587652cc6cd5ae66ffb6d94936cdd900fa7764c06adda63f7fcb5f1aca3"} Dec 11 08:44:15 crc kubenswrapper[4860]: I1211 08:44:15.215630 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8ptks" Dec 11 08:44:15 crc kubenswrapper[4860]: I1211 08:44:15.306388 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2dda62c-b6df-481f-a388-a6bfc3b02c9e-ovn-combined-ca-bundle\") pod \"c2dda62c-b6df-481f-a388-a6bfc3b02c9e\" (UID: \"c2dda62c-b6df-481f-a388-a6bfc3b02c9e\") " Dec 11 08:44:15 crc kubenswrapper[4860]: I1211 08:44:15.306436 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/c2dda62c-b6df-481f-a388-a6bfc3b02c9e-ovncontroller-config-0\") pod \"c2dda62c-b6df-481f-a388-a6bfc3b02c9e\" (UID: \"c2dda62c-b6df-481f-a388-a6bfc3b02c9e\") " Dec 11 08:44:15 crc kubenswrapper[4860]: I1211 08:44:15.306609 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9jkl9\" (UniqueName: \"kubernetes.io/projected/c2dda62c-b6df-481f-a388-a6bfc3b02c9e-kube-api-access-9jkl9\") pod \"c2dda62c-b6df-481f-a388-a6bfc3b02c9e\" (UID: \"c2dda62c-b6df-481f-a388-a6bfc3b02c9e\") " Dec 11 08:44:15 crc kubenswrapper[4860]: I1211 08:44:15.306706 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c2dda62c-b6df-481f-a388-a6bfc3b02c9e-inventory\") pod \"c2dda62c-b6df-481f-a388-a6bfc3b02c9e\" (UID: \"c2dda62c-b6df-481f-a388-a6bfc3b02c9e\") " Dec 11 08:44:15 crc kubenswrapper[4860]: I1211 08:44:15.314476 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2dda62c-b6df-481f-a388-a6bfc3b02c9e-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "c2dda62c-b6df-481f-a388-a6bfc3b02c9e" (UID: "c2dda62c-b6df-481f-a388-a6bfc3b02c9e"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:44:15 crc kubenswrapper[4860]: I1211 08:44:15.314869 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2dda62c-b6df-481f-a388-a6bfc3b02c9e-kube-api-access-9jkl9" (OuterVolumeSpecName: "kube-api-access-9jkl9") pod "c2dda62c-b6df-481f-a388-a6bfc3b02c9e" (UID: "c2dda62c-b6df-481f-a388-a6bfc3b02c9e"). InnerVolumeSpecName "kube-api-access-9jkl9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:44:15 crc kubenswrapper[4860]: I1211 08:44:15.336299 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c2dda62c-b6df-481f-a388-a6bfc3b02c9e-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "c2dda62c-b6df-481f-a388-a6bfc3b02c9e" (UID: "c2dda62c-b6df-481f-a388-a6bfc3b02c9e"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:44:15 crc kubenswrapper[4860]: I1211 08:44:15.339335 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2dda62c-b6df-481f-a388-a6bfc3b02c9e-inventory" (OuterVolumeSpecName: "inventory") pod "c2dda62c-b6df-481f-a388-a6bfc3b02c9e" (UID: "c2dda62c-b6df-481f-a388-a6bfc3b02c9e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:44:15 crc kubenswrapper[4860]: I1211 08:44:15.409159 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c2dda62c-b6df-481f-a388-a6bfc3b02c9e-ssh-key\") pod \"c2dda62c-b6df-481f-a388-a6bfc3b02c9e\" (UID: \"c2dda62c-b6df-481f-a388-a6bfc3b02c9e\") " Dec 11 08:44:15 crc kubenswrapper[4860]: I1211 08:44:15.410444 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9jkl9\" (UniqueName: \"kubernetes.io/projected/c2dda62c-b6df-481f-a388-a6bfc3b02c9e-kube-api-access-9jkl9\") on node \"crc\" DevicePath \"\"" Dec 11 08:44:15 crc kubenswrapper[4860]: I1211 08:44:15.410467 4860 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c2dda62c-b6df-481f-a388-a6bfc3b02c9e-inventory\") on node \"crc\" DevicePath \"\"" Dec 11 08:44:15 crc kubenswrapper[4860]: I1211 08:44:15.410478 4860 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2dda62c-b6df-481f-a388-a6bfc3b02c9e-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 08:44:15 crc kubenswrapper[4860]: I1211 08:44:15.410488 4860 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/c2dda62c-b6df-481f-a388-a6bfc3b02c9e-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Dec 11 08:44:15 crc kubenswrapper[4860]: I1211 08:44:15.433512 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2dda62c-b6df-481f-a388-a6bfc3b02c9e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "c2dda62c-b6df-481f-a388-a6bfc3b02c9e" (UID: "c2dda62c-b6df-481f-a388-a6bfc3b02c9e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:44:15 crc kubenswrapper[4860]: I1211 08:44:15.511215 4860 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c2dda62c-b6df-481f-a388-a6bfc3b02c9e-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 11 08:44:15 crc kubenswrapper[4860]: I1211 08:44:15.845392 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8ptks" event={"ID":"c2dda62c-b6df-481f-a388-a6bfc3b02c9e","Type":"ContainerDied","Data":"b2d07f0f659856caab257bb44bbd39ed58fba2f6278078e1b038b288c59d2a63"} Dec 11 08:44:15 crc kubenswrapper[4860]: I1211 08:44:15.845439 4860 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b2d07f0f659856caab257bb44bbd39ed58fba2f6278078e1b038b288c59d2a63" Dec 11 08:44:15 crc kubenswrapper[4860]: I1211 08:44:15.845515 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8ptks" Dec 11 08:44:15 crc kubenswrapper[4860]: I1211 08:44:15.943200 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gfrj6"] Dec 11 08:44:15 crc kubenswrapper[4860]: E1211 08:44:15.943747 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2dda62c-b6df-481f-a388-a6bfc3b02c9e" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 11 08:44:15 crc kubenswrapper[4860]: I1211 08:44:15.943770 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2dda62c-b6df-481f-a388-a6bfc3b02c9e" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 11 08:44:15 crc kubenswrapper[4860]: I1211 08:44:15.944016 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2dda62c-b6df-481f-a388-a6bfc3b02c9e" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 11 08:44:15 crc kubenswrapper[4860]: I1211 08:44:15.944875 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gfrj6" Dec 11 08:44:15 crc kubenswrapper[4860]: I1211 08:44:15.947116 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-7w7kh" Dec 11 08:44:15 crc kubenswrapper[4860]: I1211 08:44:15.947149 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Dec 11 08:44:15 crc kubenswrapper[4860]: I1211 08:44:15.947154 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 11 08:44:15 crc kubenswrapper[4860]: I1211 08:44:15.947116 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 11 08:44:15 crc kubenswrapper[4860]: I1211 08:44:15.947387 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 11 08:44:15 crc kubenswrapper[4860]: I1211 08:44:15.947475 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Dec 11 08:44:15 crc kubenswrapper[4860]: I1211 08:44:15.953192 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gfrj6"] Dec 11 08:44:16 crc kubenswrapper[4860]: I1211 08:44:16.023093 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p7wvc\" (UniqueName: \"kubernetes.io/projected/70691461-ffb3-4a01-aa11-bbe665e77b35-kube-api-access-p7wvc\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-gfrj6\" (UID: \"70691461-ffb3-4a01-aa11-bbe665e77b35\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gfrj6" Dec 11 08:44:16 crc kubenswrapper[4860]: I1211 08:44:16.023255 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/70691461-ffb3-4a01-aa11-bbe665e77b35-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-gfrj6\" (UID: \"70691461-ffb3-4a01-aa11-bbe665e77b35\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gfrj6" Dec 11 08:44:16 crc kubenswrapper[4860]: I1211 08:44:16.023409 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/70691461-ffb3-4a01-aa11-bbe665e77b35-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-gfrj6\" (UID: \"70691461-ffb3-4a01-aa11-bbe665e77b35\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gfrj6" Dec 11 08:44:16 crc kubenswrapper[4860]: I1211 08:44:16.023633 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/70691461-ffb3-4a01-aa11-bbe665e77b35-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-gfrj6\" (UID: \"70691461-ffb3-4a01-aa11-bbe665e77b35\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gfrj6" Dec 11 08:44:16 crc kubenswrapper[4860]: I1211 08:44:16.023765 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70691461-ffb3-4a01-aa11-bbe665e77b35-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-gfrj6\" (UID: \"70691461-ffb3-4a01-aa11-bbe665e77b35\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gfrj6" Dec 11 08:44:16 crc kubenswrapper[4860]: I1211 08:44:16.023886 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/70691461-ffb3-4a01-aa11-bbe665e77b35-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-gfrj6\" (UID: \"70691461-ffb3-4a01-aa11-bbe665e77b35\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gfrj6" Dec 11 08:44:16 crc kubenswrapper[4860]: I1211 08:44:16.125499 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/70691461-ffb3-4a01-aa11-bbe665e77b35-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-gfrj6\" (UID: \"70691461-ffb3-4a01-aa11-bbe665e77b35\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gfrj6" Dec 11 08:44:16 crc kubenswrapper[4860]: I1211 08:44:16.125924 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/70691461-ffb3-4a01-aa11-bbe665e77b35-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-gfrj6\" (UID: \"70691461-ffb3-4a01-aa11-bbe665e77b35\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gfrj6" Dec 11 08:44:16 crc kubenswrapper[4860]: I1211 08:44:16.126048 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70691461-ffb3-4a01-aa11-bbe665e77b35-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-gfrj6\" (UID: \"70691461-ffb3-4a01-aa11-bbe665e77b35\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gfrj6" Dec 11 08:44:16 crc kubenswrapper[4860]: I1211 08:44:16.126192 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/70691461-ffb3-4a01-aa11-bbe665e77b35-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-gfrj6\" (UID: \"70691461-ffb3-4a01-aa11-bbe665e77b35\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gfrj6" Dec 11 08:44:16 crc kubenswrapper[4860]: I1211 08:44:16.126324 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p7wvc\" (UniqueName: \"kubernetes.io/projected/70691461-ffb3-4a01-aa11-bbe665e77b35-kube-api-access-p7wvc\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-gfrj6\" (UID: \"70691461-ffb3-4a01-aa11-bbe665e77b35\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gfrj6" Dec 11 08:44:16 crc kubenswrapper[4860]: I1211 08:44:16.126448 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/70691461-ffb3-4a01-aa11-bbe665e77b35-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-gfrj6\" (UID: \"70691461-ffb3-4a01-aa11-bbe665e77b35\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gfrj6" Dec 11 08:44:16 crc kubenswrapper[4860]: I1211 08:44:16.131615 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/70691461-ffb3-4a01-aa11-bbe665e77b35-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-gfrj6\" (UID: \"70691461-ffb3-4a01-aa11-bbe665e77b35\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gfrj6" Dec 11 08:44:16 crc kubenswrapper[4860]: I1211 08:44:16.131624 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70691461-ffb3-4a01-aa11-bbe665e77b35-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-gfrj6\" (UID: \"70691461-ffb3-4a01-aa11-bbe665e77b35\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gfrj6" Dec 11 08:44:16 crc kubenswrapper[4860]: I1211 08:44:16.133341 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/70691461-ffb3-4a01-aa11-bbe665e77b35-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-gfrj6\" (UID: \"70691461-ffb3-4a01-aa11-bbe665e77b35\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gfrj6" Dec 11 08:44:16 crc kubenswrapper[4860]: I1211 08:44:16.134190 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/70691461-ffb3-4a01-aa11-bbe665e77b35-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-gfrj6\" (UID: \"70691461-ffb3-4a01-aa11-bbe665e77b35\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gfrj6" Dec 11 08:44:16 crc kubenswrapper[4860]: I1211 08:44:16.134447 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/70691461-ffb3-4a01-aa11-bbe665e77b35-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-gfrj6\" (UID: \"70691461-ffb3-4a01-aa11-bbe665e77b35\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gfrj6" Dec 11 08:44:16 crc kubenswrapper[4860]: I1211 08:44:16.143866 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p7wvc\" (UniqueName: \"kubernetes.io/projected/70691461-ffb3-4a01-aa11-bbe665e77b35-kube-api-access-p7wvc\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-gfrj6\" (UID: \"70691461-ffb3-4a01-aa11-bbe665e77b35\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gfrj6" Dec 11 08:44:16 crc kubenswrapper[4860]: I1211 08:44:16.270145 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gfrj6" Dec 11 08:44:16 crc kubenswrapper[4860]: I1211 08:44:16.783660 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gfrj6"] Dec 11 08:44:16 crc kubenswrapper[4860]: I1211 08:44:16.788268 4860 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 11 08:44:16 crc kubenswrapper[4860]: I1211 08:44:16.854470 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gfrj6" event={"ID":"70691461-ffb3-4a01-aa11-bbe665e77b35","Type":"ContainerStarted","Data":"b001f2819d7f21aa2d866adee7bd25b8f72a21796fda291c8556f7b6a7355af9"} Dec 11 08:44:17 crc kubenswrapper[4860]: I1211 08:44:17.602316 4860 scope.go:117] "RemoveContainer" containerID="af76095360a11870b804fb7f16c40dad604dbf93152022249293687b1519ca1b" Dec 11 08:44:17 crc kubenswrapper[4860]: I1211 08:44:17.863231 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gfrj6" event={"ID":"70691461-ffb3-4a01-aa11-bbe665e77b35","Type":"ContainerStarted","Data":"611b0cf7398e975c353daea5044d4858d090f004f59de70af1d974c2a1e443c2"} Dec 11 08:44:17 crc kubenswrapper[4860]: I1211 08:44:17.865686 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" event={"ID":"31c30642-6e60-41b4-a477-0d802424e0aa","Type":"ContainerStarted","Data":"360fac818f30e5d5bec8360d7df78a5f152a69c33b8db5723f4b538448a80e00"} Dec 11 08:44:17 crc kubenswrapper[4860]: I1211 08:44:17.880545 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gfrj6" podStartSLOduration=2.232318869 podStartE2EDuration="2.880526529s" podCreationTimestamp="2025-12-11 08:44:15 +0000 UTC" firstStartedPulling="2025-12-11 08:44:16.788094624 +0000 UTC m=+1989.516613679" lastFinishedPulling="2025-12-11 08:44:17.436302274 +0000 UTC m=+1990.164821339" observedRunningTime="2025-12-11 08:44:17.877185974 +0000 UTC m=+1990.605705029" watchObservedRunningTime="2025-12-11 08:44:17.880526529 +0000 UTC m=+1990.609045584" Dec 11 08:45:00 crc kubenswrapper[4860]: I1211 08:45:00.137751 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-l6c22"] Dec 11 08:45:00 crc kubenswrapper[4860]: I1211 08:45:00.141513 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-l6c22" Dec 11 08:45:00 crc kubenswrapper[4860]: I1211 08:45:00.157411 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-l6c22"] Dec 11 08:45:00 crc kubenswrapper[4860]: I1211 08:45:00.247094 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29424045-lt8h8"] Dec 11 08:45:00 crc kubenswrapper[4860]: I1211 08:45:00.248558 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29424045-lt8h8" Dec 11 08:45:00 crc kubenswrapper[4860]: I1211 08:45:00.250555 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 11 08:45:00 crc kubenswrapper[4860]: I1211 08:45:00.251060 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 11 08:45:00 crc kubenswrapper[4860]: I1211 08:45:00.255591 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29424045-lt8h8"] Dec 11 08:45:00 crc kubenswrapper[4860]: I1211 08:45:00.290038 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pzjq6\" (UniqueName: \"kubernetes.io/projected/81ffd57d-3106-4052-b985-5a181330c8ec-kube-api-access-pzjq6\") pod \"community-operators-l6c22\" (UID: \"81ffd57d-3106-4052-b985-5a181330c8ec\") " pod="openshift-marketplace/community-operators-l6c22" Dec 11 08:45:00 crc kubenswrapper[4860]: I1211 08:45:00.290404 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/81ffd57d-3106-4052-b985-5a181330c8ec-catalog-content\") pod \"community-operators-l6c22\" (UID: \"81ffd57d-3106-4052-b985-5a181330c8ec\") " pod="openshift-marketplace/community-operators-l6c22" Dec 11 08:45:00 crc kubenswrapper[4860]: I1211 08:45:00.290939 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/81ffd57d-3106-4052-b985-5a181330c8ec-utilities\") pod \"community-operators-l6c22\" (UID: \"81ffd57d-3106-4052-b985-5a181330c8ec\") " pod="openshift-marketplace/community-operators-l6c22" Dec 11 08:45:00 crc kubenswrapper[4860]: I1211 08:45:00.393326 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a7bead9d-e09d-4e34-81f5-da7de325c114-secret-volume\") pod \"collect-profiles-29424045-lt8h8\" (UID: \"a7bead9d-e09d-4e34-81f5-da7de325c114\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424045-lt8h8" Dec 11 08:45:00 crc kubenswrapper[4860]: I1211 08:45:00.393430 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a7bead9d-e09d-4e34-81f5-da7de325c114-config-volume\") pod \"collect-profiles-29424045-lt8h8\" (UID: \"a7bead9d-e09d-4e34-81f5-da7de325c114\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424045-lt8h8" Dec 11 08:45:00 crc kubenswrapper[4860]: I1211 08:45:00.393478 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/81ffd57d-3106-4052-b985-5a181330c8ec-utilities\") pod \"community-operators-l6c22\" (UID: \"81ffd57d-3106-4052-b985-5a181330c8ec\") " pod="openshift-marketplace/community-operators-l6c22" Dec 11 08:45:00 crc kubenswrapper[4860]: I1211 08:45:00.393569 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f5nfx\" (UniqueName: \"kubernetes.io/projected/a7bead9d-e09d-4e34-81f5-da7de325c114-kube-api-access-f5nfx\") pod \"collect-profiles-29424045-lt8h8\" (UID: \"a7bead9d-e09d-4e34-81f5-da7de325c114\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424045-lt8h8" Dec 11 08:45:00 crc kubenswrapper[4860]: I1211 08:45:00.393594 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pzjq6\" (UniqueName: \"kubernetes.io/projected/81ffd57d-3106-4052-b985-5a181330c8ec-kube-api-access-pzjq6\") pod \"community-operators-l6c22\" (UID: \"81ffd57d-3106-4052-b985-5a181330c8ec\") " pod="openshift-marketplace/community-operators-l6c22" Dec 11 08:45:00 crc kubenswrapper[4860]: I1211 08:45:00.393642 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/81ffd57d-3106-4052-b985-5a181330c8ec-catalog-content\") pod \"community-operators-l6c22\" (UID: \"81ffd57d-3106-4052-b985-5a181330c8ec\") " pod="openshift-marketplace/community-operators-l6c22" Dec 11 08:45:00 crc kubenswrapper[4860]: I1211 08:45:00.394303 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/81ffd57d-3106-4052-b985-5a181330c8ec-catalog-content\") pod \"community-operators-l6c22\" (UID: \"81ffd57d-3106-4052-b985-5a181330c8ec\") " pod="openshift-marketplace/community-operators-l6c22" Dec 11 08:45:00 crc kubenswrapper[4860]: I1211 08:45:00.394687 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/81ffd57d-3106-4052-b985-5a181330c8ec-utilities\") pod \"community-operators-l6c22\" (UID: \"81ffd57d-3106-4052-b985-5a181330c8ec\") " pod="openshift-marketplace/community-operators-l6c22" Dec 11 08:45:00 crc kubenswrapper[4860]: I1211 08:45:00.418122 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pzjq6\" (UniqueName: \"kubernetes.io/projected/81ffd57d-3106-4052-b985-5a181330c8ec-kube-api-access-pzjq6\") pod \"community-operators-l6c22\" (UID: \"81ffd57d-3106-4052-b985-5a181330c8ec\") " pod="openshift-marketplace/community-operators-l6c22" Dec 11 08:45:00 crc kubenswrapper[4860]: I1211 08:45:00.472277 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-l6c22" Dec 11 08:45:00 crc kubenswrapper[4860]: I1211 08:45:00.495611 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f5nfx\" (UniqueName: \"kubernetes.io/projected/a7bead9d-e09d-4e34-81f5-da7de325c114-kube-api-access-f5nfx\") pod \"collect-profiles-29424045-lt8h8\" (UID: \"a7bead9d-e09d-4e34-81f5-da7de325c114\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424045-lt8h8" Dec 11 08:45:00 crc kubenswrapper[4860]: I1211 08:45:00.495793 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a7bead9d-e09d-4e34-81f5-da7de325c114-secret-volume\") pod \"collect-profiles-29424045-lt8h8\" (UID: \"a7bead9d-e09d-4e34-81f5-da7de325c114\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424045-lt8h8" Dec 11 08:45:00 crc kubenswrapper[4860]: I1211 08:45:00.495858 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a7bead9d-e09d-4e34-81f5-da7de325c114-config-volume\") pod \"collect-profiles-29424045-lt8h8\" (UID: \"a7bead9d-e09d-4e34-81f5-da7de325c114\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424045-lt8h8" Dec 11 08:45:00 crc kubenswrapper[4860]: I1211 08:45:00.496807 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a7bead9d-e09d-4e34-81f5-da7de325c114-config-volume\") pod \"collect-profiles-29424045-lt8h8\" (UID: \"a7bead9d-e09d-4e34-81f5-da7de325c114\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424045-lt8h8" Dec 11 08:45:00 crc kubenswrapper[4860]: I1211 08:45:00.502696 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a7bead9d-e09d-4e34-81f5-da7de325c114-secret-volume\") pod \"collect-profiles-29424045-lt8h8\" (UID: \"a7bead9d-e09d-4e34-81f5-da7de325c114\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424045-lt8h8" Dec 11 08:45:00 crc kubenswrapper[4860]: I1211 08:45:00.520458 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f5nfx\" (UniqueName: \"kubernetes.io/projected/a7bead9d-e09d-4e34-81f5-da7de325c114-kube-api-access-f5nfx\") pod \"collect-profiles-29424045-lt8h8\" (UID: \"a7bead9d-e09d-4e34-81f5-da7de325c114\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424045-lt8h8" Dec 11 08:45:00 crc kubenswrapper[4860]: I1211 08:45:00.576309 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29424045-lt8h8" Dec 11 08:45:01 crc kubenswrapper[4860]: I1211 08:45:01.031190 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-l6c22"] Dec 11 08:45:01 crc kubenswrapper[4860]: W1211 08:45:01.035268 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod81ffd57d_3106_4052_b985_5a181330c8ec.slice/crio-3a3b1755907ba5dd6ea064b7f02fe6e6ef3c78b180c14f259c0c86f650572af9 WatchSource:0}: Error finding container 3a3b1755907ba5dd6ea064b7f02fe6e6ef3c78b180c14f259c0c86f650572af9: Status 404 returned error can't find the container with id 3a3b1755907ba5dd6ea064b7f02fe6e6ef3c78b180c14f259c0c86f650572af9 Dec 11 08:45:01 crc kubenswrapper[4860]: I1211 08:45:01.206269 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29424045-lt8h8"] Dec 11 08:45:01 crc kubenswrapper[4860]: W1211 08:45:01.215017 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda7bead9d_e09d_4e34_81f5_da7de325c114.slice/crio-9d0355bf3f9742d3c1777045f9ac7081e470ea4da0d3a5b3fdd37d5faac2ce0f WatchSource:0}: Error finding container 9d0355bf3f9742d3c1777045f9ac7081e470ea4da0d3a5b3fdd37d5faac2ce0f: Status 404 returned error can't find the container with id 9d0355bf3f9742d3c1777045f9ac7081e470ea4da0d3a5b3fdd37d5faac2ce0f Dec 11 08:45:01 crc kubenswrapper[4860]: I1211 08:45:01.235342 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l6c22" event={"ID":"81ffd57d-3106-4052-b985-5a181330c8ec","Type":"ContainerStarted","Data":"cbcb517dcca6956b63dd1ec30c8af4fdc83fa2b5a5391311a6aa5f2a56f7eeba"} Dec 11 08:45:01 crc kubenswrapper[4860]: I1211 08:45:01.235387 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l6c22" event={"ID":"81ffd57d-3106-4052-b985-5a181330c8ec","Type":"ContainerStarted","Data":"3a3b1755907ba5dd6ea064b7f02fe6e6ef3c78b180c14f259c0c86f650572af9"} Dec 11 08:45:01 crc kubenswrapper[4860]: I1211 08:45:01.236886 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29424045-lt8h8" event={"ID":"a7bead9d-e09d-4e34-81f5-da7de325c114","Type":"ContainerStarted","Data":"9d0355bf3f9742d3c1777045f9ac7081e470ea4da0d3a5b3fdd37d5faac2ce0f"} Dec 11 08:45:02 crc kubenswrapper[4860]: I1211 08:45:02.246912 4860 generic.go:334] "Generic (PLEG): container finished" podID="81ffd57d-3106-4052-b985-5a181330c8ec" containerID="cbcb517dcca6956b63dd1ec30c8af4fdc83fa2b5a5391311a6aa5f2a56f7eeba" exitCode=0 Dec 11 08:45:02 crc kubenswrapper[4860]: I1211 08:45:02.246986 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l6c22" event={"ID":"81ffd57d-3106-4052-b985-5a181330c8ec","Type":"ContainerDied","Data":"cbcb517dcca6956b63dd1ec30c8af4fdc83fa2b5a5391311a6aa5f2a56f7eeba"} Dec 11 08:45:02 crc kubenswrapper[4860]: I1211 08:45:02.249215 4860 generic.go:334] "Generic (PLEG): container finished" podID="70691461-ffb3-4a01-aa11-bbe665e77b35" containerID="611b0cf7398e975c353daea5044d4858d090f004f59de70af1d974c2a1e443c2" exitCode=0 Dec 11 08:45:02 crc kubenswrapper[4860]: I1211 08:45:02.249281 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gfrj6" event={"ID":"70691461-ffb3-4a01-aa11-bbe665e77b35","Type":"ContainerDied","Data":"611b0cf7398e975c353daea5044d4858d090f004f59de70af1d974c2a1e443c2"} Dec 11 08:45:02 crc kubenswrapper[4860]: I1211 08:45:02.252284 4860 generic.go:334] "Generic (PLEG): container finished" podID="a7bead9d-e09d-4e34-81f5-da7de325c114" containerID="52830a50c25656d35e1812e90ddd9ff9b2523eaa9b7742c242d8622a55de2950" exitCode=0 Dec 11 08:45:02 crc kubenswrapper[4860]: I1211 08:45:02.252322 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29424045-lt8h8" event={"ID":"a7bead9d-e09d-4e34-81f5-da7de325c114","Type":"ContainerDied","Data":"52830a50c25656d35e1812e90ddd9ff9b2523eaa9b7742c242d8622a55de2950"} Dec 11 08:45:03 crc kubenswrapper[4860]: I1211 08:45:03.682619 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29424045-lt8h8" Dec 11 08:45:03 crc kubenswrapper[4860]: I1211 08:45:03.690300 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gfrj6" Dec 11 08:45:03 crc kubenswrapper[4860]: I1211 08:45:03.711321 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/70691461-ffb3-4a01-aa11-bbe665e77b35-neutron-ovn-metadata-agent-neutron-config-0\") pod \"70691461-ffb3-4a01-aa11-bbe665e77b35\" (UID: \"70691461-ffb3-4a01-aa11-bbe665e77b35\") " Dec 11 08:45:03 crc kubenswrapper[4860]: I1211 08:45:03.711375 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70691461-ffb3-4a01-aa11-bbe665e77b35-neutron-metadata-combined-ca-bundle\") pod \"70691461-ffb3-4a01-aa11-bbe665e77b35\" (UID: \"70691461-ffb3-4a01-aa11-bbe665e77b35\") " Dec 11 08:45:03 crc kubenswrapper[4860]: I1211 08:45:03.711404 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a7bead9d-e09d-4e34-81f5-da7de325c114-secret-volume\") pod \"a7bead9d-e09d-4e34-81f5-da7de325c114\" (UID: \"a7bead9d-e09d-4e34-81f5-da7de325c114\") " Dec 11 08:45:03 crc kubenswrapper[4860]: I1211 08:45:03.711439 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p7wvc\" (UniqueName: \"kubernetes.io/projected/70691461-ffb3-4a01-aa11-bbe665e77b35-kube-api-access-p7wvc\") pod \"70691461-ffb3-4a01-aa11-bbe665e77b35\" (UID: \"70691461-ffb3-4a01-aa11-bbe665e77b35\") " Dec 11 08:45:03 crc kubenswrapper[4860]: I1211 08:45:03.711464 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/70691461-ffb3-4a01-aa11-bbe665e77b35-nova-metadata-neutron-config-0\") pod \"70691461-ffb3-4a01-aa11-bbe665e77b35\" (UID: \"70691461-ffb3-4a01-aa11-bbe665e77b35\") " Dec 11 08:45:03 crc kubenswrapper[4860]: I1211 08:45:03.711496 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a7bead9d-e09d-4e34-81f5-da7de325c114-config-volume\") pod \"a7bead9d-e09d-4e34-81f5-da7de325c114\" (UID: \"a7bead9d-e09d-4e34-81f5-da7de325c114\") " Dec 11 08:45:03 crc kubenswrapper[4860]: I1211 08:45:03.711534 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/70691461-ffb3-4a01-aa11-bbe665e77b35-ssh-key\") pod \"70691461-ffb3-4a01-aa11-bbe665e77b35\" (UID: \"70691461-ffb3-4a01-aa11-bbe665e77b35\") " Dec 11 08:45:03 crc kubenswrapper[4860]: I1211 08:45:03.711560 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f5nfx\" (UniqueName: \"kubernetes.io/projected/a7bead9d-e09d-4e34-81f5-da7de325c114-kube-api-access-f5nfx\") pod \"a7bead9d-e09d-4e34-81f5-da7de325c114\" (UID: \"a7bead9d-e09d-4e34-81f5-da7de325c114\") " Dec 11 08:45:03 crc kubenswrapper[4860]: I1211 08:45:03.711588 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/70691461-ffb3-4a01-aa11-bbe665e77b35-inventory\") pod \"70691461-ffb3-4a01-aa11-bbe665e77b35\" (UID: \"70691461-ffb3-4a01-aa11-bbe665e77b35\") " Dec 11 08:45:03 crc kubenswrapper[4860]: I1211 08:45:03.718999 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a7bead9d-e09d-4e34-81f5-da7de325c114-config-volume" (OuterVolumeSpecName: "config-volume") pod "a7bead9d-e09d-4e34-81f5-da7de325c114" (UID: "a7bead9d-e09d-4e34-81f5-da7de325c114"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:45:03 crc kubenswrapper[4860]: I1211 08:45:03.723584 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7bead9d-e09d-4e34-81f5-da7de325c114-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "a7bead9d-e09d-4e34-81f5-da7de325c114" (UID: "a7bead9d-e09d-4e34-81f5-da7de325c114"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:45:03 crc kubenswrapper[4860]: I1211 08:45:03.728400 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7bead9d-e09d-4e34-81f5-da7de325c114-kube-api-access-f5nfx" (OuterVolumeSpecName: "kube-api-access-f5nfx") pod "a7bead9d-e09d-4e34-81f5-da7de325c114" (UID: "a7bead9d-e09d-4e34-81f5-da7de325c114"). InnerVolumeSpecName "kube-api-access-f5nfx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:45:03 crc kubenswrapper[4860]: I1211 08:45:03.731481 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/70691461-ffb3-4a01-aa11-bbe665e77b35-kube-api-access-p7wvc" (OuterVolumeSpecName: "kube-api-access-p7wvc") pod "70691461-ffb3-4a01-aa11-bbe665e77b35" (UID: "70691461-ffb3-4a01-aa11-bbe665e77b35"). InnerVolumeSpecName "kube-api-access-p7wvc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:45:03 crc kubenswrapper[4860]: I1211 08:45:03.740097 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70691461-ffb3-4a01-aa11-bbe665e77b35-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "70691461-ffb3-4a01-aa11-bbe665e77b35" (UID: "70691461-ffb3-4a01-aa11-bbe665e77b35"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:45:03 crc kubenswrapper[4860]: I1211 08:45:03.780996 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70691461-ffb3-4a01-aa11-bbe665e77b35-inventory" (OuterVolumeSpecName: "inventory") pod "70691461-ffb3-4a01-aa11-bbe665e77b35" (UID: "70691461-ffb3-4a01-aa11-bbe665e77b35"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:45:03 crc kubenswrapper[4860]: I1211 08:45:03.781099 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70691461-ffb3-4a01-aa11-bbe665e77b35-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "70691461-ffb3-4a01-aa11-bbe665e77b35" (UID: "70691461-ffb3-4a01-aa11-bbe665e77b35"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:45:03 crc kubenswrapper[4860]: I1211 08:45:03.785810 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70691461-ffb3-4a01-aa11-bbe665e77b35-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "70691461-ffb3-4a01-aa11-bbe665e77b35" (UID: "70691461-ffb3-4a01-aa11-bbe665e77b35"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:45:03 crc kubenswrapper[4860]: I1211 08:45:03.790412 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70691461-ffb3-4a01-aa11-bbe665e77b35-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "70691461-ffb3-4a01-aa11-bbe665e77b35" (UID: "70691461-ffb3-4a01-aa11-bbe665e77b35"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:45:03 crc kubenswrapper[4860]: I1211 08:45:03.813340 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p7wvc\" (UniqueName: \"kubernetes.io/projected/70691461-ffb3-4a01-aa11-bbe665e77b35-kube-api-access-p7wvc\") on node \"crc\" DevicePath \"\"" Dec 11 08:45:03 crc kubenswrapper[4860]: I1211 08:45:03.813376 4860 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/70691461-ffb3-4a01-aa11-bbe665e77b35-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 11 08:45:03 crc kubenswrapper[4860]: I1211 08:45:03.813390 4860 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a7bead9d-e09d-4e34-81f5-da7de325c114-config-volume\") on node \"crc\" DevicePath \"\"" Dec 11 08:45:03 crc kubenswrapper[4860]: I1211 08:45:03.813400 4860 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/70691461-ffb3-4a01-aa11-bbe665e77b35-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 11 08:45:03 crc kubenswrapper[4860]: I1211 08:45:03.813409 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f5nfx\" (UniqueName: \"kubernetes.io/projected/a7bead9d-e09d-4e34-81f5-da7de325c114-kube-api-access-f5nfx\") on node \"crc\" DevicePath \"\"" Dec 11 08:45:03 crc kubenswrapper[4860]: I1211 08:45:03.813420 4860 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/70691461-ffb3-4a01-aa11-bbe665e77b35-inventory\") on node \"crc\" DevicePath \"\"" Dec 11 08:45:03 crc kubenswrapper[4860]: I1211 08:45:03.813428 4860 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/70691461-ffb3-4a01-aa11-bbe665e77b35-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 11 08:45:03 crc kubenswrapper[4860]: I1211 08:45:03.813436 4860 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70691461-ffb3-4a01-aa11-bbe665e77b35-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 08:45:03 crc kubenswrapper[4860]: I1211 08:45:03.813446 4860 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a7bead9d-e09d-4e34-81f5-da7de325c114-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 11 08:45:04 crc kubenswrapper[4860]: I1211 08:45:04.271202 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gfrj6" event={"ID":"70691461-ffb3-4a01-aa11-bbe665e77b35","Type":"ContainerDied","Data":"b001f2819d7f21aa2d866adee7bd25b8f72a21796fda291c8556f7b6a7355af9"} Dec 11 08:45:04 crc kubenswrapper[4860]: I1211 08:45:04.271550 4860 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b001f2819d7f21aa2d866adee7bd25b8f72a21796fda291c8556f7b6a7355af9" Dec 11 08:45:04 crc kubenswrapper[4860]: I1211 08:45:04.271304 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gfrj6" Dec 11 08:45:04 crc kubenswrapper[4860]: I1211 08:45:04.275197 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29424045-lt8h8" Dec 11 08:45:04 crc kubenswrapper[4860]: I1211 08:45:04.275205 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29424045-lt8h8" event={"ID":"a7bead9d-e09d-4e34-81f5-da7de325c114","Type":"ContainerDied","Data":"9d0355bf3f9742d3c1777045f9ac7081e470ea4da0d3a5b3fdd37d5faac2ce0f"} Dec 11 08:45:04 crc kubenswrapper[4860]: I1211 08:45:04.275250 4860 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9d0355bf3f9742d3c1777045f9ac7081e470ea4da0d3a5b3fdd37d5faac2ce0f" Dec 11 08:45:04 crc kubenswrapper[4860]: I1211 08:45:04.277567 4860 generic.go:334] "Generic (PLEG): container finished" podID="81ffd57d-3106-4052-b985-5a181330c8ec" containerID="2c2b907fd49217f073e205d1b82a78dbbac052dd2e097195abb0e5dc2debcca1" exitCode=0 Dec 11 08:45:04 crc kubenswrapper[4860]: I1211 08:45:04.277603 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l6c22" event={"ID":"81ffd57d-3106-4052-b985-5a181330c8ec","Type":"ContainerDied","Data":"2c2b907fd49217f073e205d1b82a78dbbac052dd2e097195abb0e5dc2debcca1"} Dec 11 08:45:04 crc kubenswrapper[4860]: I1211 08:45:04.455594 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5z2qw"] Dec 11 08:45:04 crc kubenswrapper[4860]: E1211 08:45:04.456104 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7bead9d-e09d-4e34-81f5-da7de325c114" containerName="collect-profiles" Dec 11 08:45:04 crc kubenswrapper[4860]: I1211 08:45:04.456125 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7bead9d-e09d-4e34-81f5-da7de325c114" containerName="collect-profiles" Dec 11 08:45:04 crc kubenswrapper[4860]: E1211 08:45:04.456140 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70691461-ffb3-4a01-aa11-bbe665e77b35" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 11 08:45:04 crc kubenswrapper[4860]: I1211 08:45:04.456151 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="70691461-ffb3-4a01-aa11-bbe665e77b35" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 11 08:45:04 crc kubenswrapper[4860]: I1211 08:45:04.456353 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="70691461-ffb3-4a01-aa11-bbe665e77b35" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 11 08:45:04 crc kubenswrapper[4860]: I1211 08:45:04.456383 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7bead9d-e09d-4e34-81f5-da7de325c114" containerName="collect-profiles" Dec 11 08:45:04 crc kubenswrapper[4860]: I1211 08:45:04.457023 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5z2qw" Dec 11 08:45:04 crc kubenswrapper[4860]: I1211 08:45:04.459209 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 11 08:45:04 crc kubenswrapper[4860]: I1211 08:45:04.459275 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Dec 11 08:45:04 crc kubenswrapper[4860]: I1211 08:45:04.459304 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-7w7kh" Dec 11 08:45:04 crc kubenswrapper[4860]: I1211 08:45:04.459355 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 11 08:45:04 crc kubenswrapper[4860]: I1211 08:45:04.465226 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 11 08:45:04 crc kubenswrapper[4860]: I1211 08:45:04.466890 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5z2qw"] Dec 11 08:45:04 crc kubenswrapper[4860]: I1211 08:45:04.528319 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kfh9v\" (UniqueName: \"kubernetes.io/projected/b0eed689-a827-47b6-81ba-91e76db6f35b-kube-api-access-kfh9v\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-5z2qw\" (UID: \"b0eed689-a827-47b6-81ba-91e76db6f35b\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5z2qw" Dec 11 08:45:04 crc kubenswrapper[4860]: I1211 08:45:04.528393 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b0eed689-a827-47b6-81ba-91e76db6f35b-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-5z2qw\" (UID: \"b0eed689-a827-47b6-81ba-91e76db6f35b\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5z2qw" Dec 11 08:45:04 crc kubenswrapper[4860]: I1211 08:45:04.528443 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0eed689-a827-47b6-81ba-91e76db6f35b-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-5z2qw\" (UID: \"b0eed689-a827-47b6-81ba-91e76db6f35b\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5z2qw" Dec 11 08:45:04 crc kubenswrapper[4860]: I1211 08:45:04.528507 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b0eed689-a827-47b6-81ba-91e76db6f35b-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-5z2qw\" (UID: \"b0eed689-a827-47b6-81ba-91e76db6f35b\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5z2qw" Dec 11 08:45:04 crc kubenswrapper[4860]: I1211 08:45:04.528553 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/b0eed689-a827-47b6-81ba-91e76db6f35b-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-5z2qw\" (UID: \"b0eed689-a827-47b6-81ba-91e76db6f35b\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5z2qw" Dec 11 08:45:04 crc kubenswrapper[4860]: I1211 08:45:04.629752 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0eed689-a827-47b6-81ba-91e76db6f35b-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-5z2qw\" (UID: \"b0eed689-a827-47b6-81ba-91e76db6f35b\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5z2qw" Dec 11 08:45:04 crc kubenswrapper[4860]: I1211 08:45:04.629814 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b0eed689-a827-47b6-81ba-91e76db6f35b-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-5z2qw\" (UID: \"b0eed689-a827-47b6-81ba-91e76db6f35b\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5z2qw" Dec 11 08:45:04 crc kubenswrapper[4860]: I1211 08:45:04.630174 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/b0eed689-a827-47b6-81ba-91e76db6f35b-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-5z2qw\" (UID: \"b0eed689-a827-47b6-81ba-91e76db6f35b\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5z2qw" Dec 11 08:45:04 crc kubenswrapper[4860]: I1211 08:45:04.630368 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kfh9v\" (UniqueName: \"kubernetes.io/projected/b0eed689-a827-47b6-81ba-91e76db6f35b-kube-api-access-kfh9v\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-5z2qw\" (UID: \"b0eed689-a827-47b6-81ba-91e76db6f35b\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5z2qw" Dec 11 08:45:04 crc kubenswrapper[4860]: I1211 08:45:04.630411 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b0eed689-a827-47b6-81ba-91e76db6f35b-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-5z2qw\" (UID: \"b0eed689-a827-47b6-81ba-91e76db6f35b\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5z2qw" Dec 11 08:45:04 crc kubenswrapper[4860]: I1211 08:45:04.636064 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b0eed689-a827-47b6-81ba-91e76db6f35b-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-5z2qw\" (UID: \"b0eed689-a827-47b6-81ba-91e76db6f35b\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5z2qw" Dec 11 08:45:04 crc kubenswrapper[4860]: I1211 08:45:04.636195 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/b0eed689-a827-47b6-81ba-91e76db6f35b-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-5z2qw\" (UID: \"b0eed689-a827-47b6-81ba-91e76db6f35b\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5z2qw" Dec 11 08:45:04 crc kubenswrapper[4860]: I1211 08:45:04.637274 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b0eed689-a827-47b6-81ba-91e76db6f35b-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-5z2qw\" (UID: \"b0eed689-a827-47b6-81ba-91e76db6f35b\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5z2qw" Dec 11 08:45:04 crc kubenswrapper[4860]: I1211 08:45:04.641407 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0eed689-a827-47b6-81ba-91e76db6f35b-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-5z2qw\" (UID: \"b0eed689-a827-47b6-81ba-91e76db6f35b\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5z2qw" Dec 11 08:45:04 crc kubenswrapper[4860]: I1211 08:45:04.656882 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kfh9v\" (UniqueName: \"kubernetes.io/projected/b0eed689-a827-47b6-81ba-91e76db6f35b-kube-api-access-kfh9v\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-5z2qw\" (UID: \"b0eed689-a827-47b6-81ba-91e76db6f35b\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5z2qw" Dec 11 08:45:04 crc kubenswrapper[4860]: I1211 08:45:04.750863 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29424000-nf7zl"] Dec 11 08:45:04 crc kubenswrapper[4860]: I1211 08:45:04.784599 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29424000-nf7zl"] Dec 11 08:45:04 crc kubenswrapper[4860]: I1211 08:45:04.790424 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5z2qw" Dec 11 08:45:05 crc kubenswrapper[4860]: I1211 08:45:05.290997 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l6c22" event={"ID":"81ffd57d-3106-4052-b985-5a181330c8ec","Type":"ContainerStarted","Data":"2d5133970fa366d202727c5f5fc290958773d99326eb9eda8b411b6547d0431e"} Dec 11 08:45:05 crc kubenswrapper[4860]: I1211 08:45:05.310125 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-l6c22" podStartSLOduration=2.708158891 podStartE2EDuration="5.310103417s" podCreationTimestamp="2025-12-11 08:45:00 +0000 UTC" firstStartedPulling="2025-12-11 08:45:02.249687527 +0000 UTC m=+2034.978206572" lastFinishedPulling="2025-12-11 08:45:04.851632043 +0000 UTC m=+2037.580151098" observedRunningTime="2025-12-11 08:45:05.306937587 +0000 UTC m=+2038.035456652" watchObservedRunningTime="2025-12-11 08:45:05.310103417 +0000 UTC m=+2038.038622482" Dec 11 08:45:05 crc kubenswrapper[4860]: I1211 08:45:05.343934 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5z2qw"] Dec 11 08:45:05 crc kubenswrapper[4860]: W1211 08:45:05.346832 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb0eed689_a827_47b6_81ba_91e76db6f35b.slice/crio-aef7abc9548de978941adae34fe118c396e036943bb572847ba4ceabf84d0761 WatchSource:0}: Error finding container aef7abc9548de978941adae34fe118c396e036943bb572847ba4ceabf84d0761: Status 404 returned error can't find the container with id aef7abc9548de978941adae34fe118c396e036943bb572847ba4ceabf84d0761 Dec 11 08:45:05 crc kubenswrapper[4860]: I1211 08:45:05.589970 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0f6c4496-7751-49db-adc2-18099589c708" path="/var/lib/kubelet/pods/0f6c4496-7751-49db-adc2-18099589c708/volumes" Dec 11 08:45:06 crc kubenswrapper[4860]: I1211 08:45:06.299781 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5z2qw" event={"ID":"b0eed689-a827-47b6-81ba-91e76db6f35b","Type":"ContainerStarted","Data":"3d85f48b0ba72df4ee804d19f66b272e661e9956e61cf14b2692c6cda570fbe3"} Dec 11 08:45:06 crc kubenswrapper[4860]: I1211 08:45:06.300296 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5z2qw" event={"ID":"b0eed689-a827-47b6-81ba-91e76db6f35b","Type":"ContainerStarted","Data":"aef7abc9548de978941adae34fe118c396e036943bb572847ba4ceabf84d0761"} Dec 11 08:45:06 crc kubenswrapper[4860]: I1211 08:45:06.323808 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5z2qw" podStartSLOduration=1.8605285569999999 podStartE2EDuration="2.32377604s" podCreationTimestamp="2025-12-11 08:45:04 +0000 UTC" firstStartedPulling="2025-12-11 08:45:05.349455062 +0000 UTC m=+2038.077974117" lastFinishedPulling="2025-12-11 08:45:05.812702535 +0000 UTC m=+2038.541221600" observedRunningTime="2025-12-11 08:45:06.313429345 +0000 UTC m=+2039.041948390" watchObservedRunningTime="2025-12-11 08:45:06.32377604 +0000 UTC m=+2039.052295115" Dec 11 08:45:06 crc kubenswrapper[4860]: I1211 08:45:06.516726 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-gs2nr"] Dec 11 08:45:06 crc kubenswrapper[4860]: I1211 08:45:06.519566 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gs2nr" Dec 11 08:45:06 crc kubenswrapper[4860]: I1211 08:45:06.536026 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-gs2nr"] Dec 11 08:45:06 crc kubenswrapper[4860]: I1211 08:45:06.586174 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3847111f-a08e-4089-8ee7-74648c6f43f6-catalog-content\") pod \"redhat-operators-gs2nr\" (UID: \"3847111f-a08e-4089-8ee7-74648c6f43f6\") " pod="openshift-marketplace/redhat-operators-gs2nr" Dec 11 08:45:06 crc kubenswrapper[4860]: I1211 08:45:06.586493 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xk6tm\" (UniqueName: \"kubernetes.io/projected/3847111f-a08e-4089-8ee7-74648c6f43f6-kube-api-access-xk6tm\") pod \"redhat-operators-gs2nr\" (UID: \"3847111f-a08e-4089-8ee7-74648c6f43f6\") " pod="openshift-marketplace/redhat-operators-gs2nr" Dec 11 08:45:06 crc kubenswrapper[4860]: I1211 08:45:06.586576 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3847111f-a08e-4089-8ee7-74648c6f43f6-utilities\") pod \"redhat-operators-gs2nr\" (UID: \"3847111f-a08e-4089-8ee7-74648c6f43f6\") " pod="openshift-marketplace/redhat-operators-gs2nr" Dec 11 08:45:06 crc kubenswrapper[4860]: I1211 08:45:06.688986 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3847111f-a08e-4089-8ee7-74648c6f43f6-catalog-content\") pod \"redhat-operators-gs2nr\" (UID: \"3847111f-a08e-4089-8ee7-74648c6f43f6\") " pod="openshift-marketplace/redhat-operators-gs2nr" Dec 11 08:45:06 crc kubenswrapper[4860]: I1211 08:45:06.689038 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xk6tm\" (UniqueName: \"kubernetes.io/projected/3847111f-a08e-4089-8ee7-74648c6f43f6-kube-api-access-xk6tm\") pod \"redhat-operators-gs2nr\" (UID: \"3847111f-a08e-4089-8ee7-74648c6f43f6\") " pod="openshift-marketplace/redhat-operators-gs2nr" Dec 11 08:45:06 crc kubenswrapper[4860]: I1211 08:45:06.689130 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3847111f-a08e-4089-8ee7-74648c6f43f6-utilities\") pod \"redhat-operators-gs2nr\" (UID: \"3847111f-a08e-4089-8ee7-74648c6f43f6\") " pod="openshift-marketplace/redhat-operators-gs2nr" Dec 11 08:45:06 crc kubenswrapper[4860]: I1211 08:45:06.690132 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3847111f-a08e-4089-8ee7-74648c6f43f6-catalog-content\") pod \"redhat-operators-gs2nr\" (UID: \"3847111f-a08e-4089-8ee7-74648c6f43f6\") " pod="openshift-marketplace/redhat-operators-gs2nr" Dec 11 08:45:06 crc kubenswrapper[4860]: I1211 08:45:06.690221 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3847111f-a08e-4089-8ee7-74648c6f43f6-utilities\") pod \"redhat-operators-gs2nr\" (UID: \"3847111f-a08e-4089-8ee7-74648c6f43f6\") " pod="openshift-marketplace/redhat-operators-gs2nr" Dec 11 08:45:06 crc kubenswrapper[4860]: I1211 08:45:06.711412 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xk6tm\" (UniqueName: \"kubernetes.io/projected/3847111f-a08e-4089-8ee7-74648c6f43f6-kube-api-access-xk6tm\") pod \"redhat-operators-gs2nr\" (UID: \"3847111f-a08e-4089-8ee7-74648c6f43f6\") " pod="openshift-marketplace/redhat-operators-gs2nr" Dec 11 08:45:06 crc kubenswrapper[4860]: I1211 08:45:06.837573 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gs2nr" Dec 11 08:45:07 crc kubenswrapper[4860]: I1211 08:45:07.292424 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-gs2nr"] Dec 11 08:45:07 crc kubenswrapper[4860]: W1211 08:45:07.308160 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3847111f_a08e_4089_8ee7_74648c6f43f6.slice/crio-a632794a228c434be8560a4be8d2cd0eb733f8fbb677427d4e252b95094e7271 WatchSource:0}: Error finding container a632794a228c434be8560a4be8d2cd0eb733f8fbb677427d4e252b95094e7271: Status 404 returned error can't find the container with id a632794a228c434be8560a4be8d2cd0eb733f8fbb677427d4e252b95094e7271 Dec 11 08:45:08 crc kubenswrapper[4860]: I1211 08:45:08.325135 4860 generic.go:334] "Generic (PLEG): container finished" podID="3847111f-a08e-4089-8ee7-74648c6f43f6" containerID="968584d44e01eacc893a8b322c54f726010ab703d15f4485b97327f52fa355a1" exitCode=0 Dec 11 08:45:08 crc kubenswrapper[4860]: I1211 08:45:08.325226 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gs2nr" event={"ID":"3847111f-a08e-4089-8ee7-74648c6f43f6","Type":"ContainerDied","Data":"968584d44e01eacc893a8b322c54f726010ab703d15f4485b97327f52fa355a1"} Dec 11 08:45:08 crc kubenswrapper[4860]: I1211 08:45:08.325784 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gs2nr" event={"ID":"3847111f-a08e-4089-8ee7-74648c6f43f6","Type":"ContainerStarted","Data":"a632794a228c434be8560a4be8d2cd0eb733f8fbb677427d4e252b95094e7271"} Dec 11 08:45:10 crc kubenswrapper[4860]: I1211 08:45:10.346094 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gs2nr" event={"ID":"3847111f-a08e-4089-8ee7-74648c6f43f6","Type":"ContainerStarted","Data":"4cba2cf73c11517188c1be2665dab18ac13ee318240e134f06d866bf3f909730"} Dec 11 08:45:10 crc kubenswrapper[4860]: I1211 08:45:10.472962 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-l6c22" Dec 11 08:45:10 crc kubenswrapper[4860]: I1211 08:45:10.473008 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-l6c22" Dec 11 08:45:10 crc kubenswrapper[4860]: I1211 08:45:10.520591 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-l6c22" Dec 11 08:45:11 crc kubenswrapper[4860]: I1211 08:45:11.409505 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-l6c22" Dec 11 08:45:11 crc kubenswrapper[4860]: I1211 08:45:11.708552 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-l6c22"] Dec 11 08:45:12 crc kubenswrapper[4860]: I1211 08:45:12.364221 4860 generic.go:334] "Generic (PLEG): container finished" podID="3847111f-a08e-4089-8ee7-74648c6f43f6" containerID="4cba2cf73c11517188c1be2665dab18ac13ee318240e134f06d866bf3f909730" exitCode=0 Dec 11 08:45:12 crc kubenswrapper[4860]: I1211 08:45:12.364330 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gs2nr" event={"ID":"3847111f-a08e-4089-8ee7-74648c6f43f6","Type":"ContainerDied","Data":"4cba2cf73c11517188c1be2665dab18ac13ee318240e134f06d866bf3f909730"} Dec 11 08:45:13 crc kubenswrapper[4860]: I1211 08:45:13.376413 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-l6c22" podUID="81ffd57d-3106-4052-b985-5a181330c8ec" containerName="registry-server" containerID="cri-o://2d5133970fa366d202727c5f5fc290958773d99326eb9eda8b411b6547d0431e" gracePeriod=2 Dec 11 08:45:13 crc kubenswrapper[4860]: I1211 08:45:13.852711 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-l6c22" Dec 11 08:45:13 crc kubenswrapper[4860]: I1211 08:45:13.945757 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/81ffd57d-3106-4052-b985-5a181330c8ec-catalog-content\") pod \"81ffd57d-3106-4052-b985-5a181330c8ec\" (UID: \"81ffd57d-3106-4052-b985-5a181330c8ec\") " Dec 11 08:45:13 crc kubenswrapper[4860]: I1211 08:45:13.945828 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/81ffd57d-3106-4052-b985-5a181330c8ec-utilities\") pod \"81ffd57d-3106-4052-b985-5a181330c8ec\" (UID: \"81ffd57d-3106-4052-b985-5a181330c8ec\") " Dec 11 08:45:13 crc kubenswrapper[4860]: I1211 08:45:13.945896 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pzjq6\" (UniqueName: \"kubernetes.io/projected/81ffd57d-3106-4052-b985-5a181330c8ec-kube-api-access-pzjq6\") pod \"81ffd57d-3106-4052-b985-5a181330c8ec\" (UID: \"81ffd57d-3106-4052-b985-5a181330c8ec\") " Dec 11 08:45:13 crc kubenswrapper[4860]: I1211 08:45:13.946828 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/81ffd57d-3106-4052-b985-5a181330c8ec-utilities" (OuterVolumeSpecName: "utilities") pod "81ffd57d-3106-4052-b985-5a181330c8ec" (UID: "81ffd57d-3106-4052-b985-5a181330c8ec"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:45:13 crc kubenswrapper[4860]: I1211 08:45:13.951198 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/81ffd57d-3106-4052-b985-5a181330c8ec-kube-api-access-pzjq6" (OuterVolumeSpecName: "kube-api-access-pzjq6") pod "81ffd57d-3106-4052-b985-5a181330c8ec" (UID: "81ffd57d-3106-4052-b985-5a181330c8ec"). InnerVolumeSpecName "kube-api-access-pzjq6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:45:13 crc kubenswrapper[4860]: I1211 08:45:13.995521 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/81ffd57d-3106-4052-b985-5a181330c8ec-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "81ffd57d-3106-4052-b985-5a181330c8ec" (UID: "81ffd57d-3106-4052-b985-5a181330c8ec"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:45:14 crc kubenswrapper[4860]: I1211 08:45:14.048328 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pzjq6\" (UniqueName: \"kubernetes.io/projected/81ffd57d-3106-4052-b985-5a181330c8ec-kube-api-access-pzjq6\") on node \"crc\" DevicePath \"\"" Dec 11 08:45:14 crc kubenswrapper[4860]: I1211 08:45:14.048371 4860 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/81ffd57d-3106-4052-b985-5a181330c8ec-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 08:45:14 crc kubenswrapper[4860]: I1211 08:45:14.048385 4860 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/81ffd57d-3106-4052-b985-5a181330c8ec-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 08:45:14 crc kubenswrapper[4860]: I1211 08:45:14.390918 4860 generic.go:334] "Generic (PLEG): container finished" podID="81ffd57d-3106-4052-b985-5a181330c8ec" containerID="2d5133970fa366d202727c5f5fc290958773d99326eb9eda8b411b6547d0431e" exitCode=0 Dec 11 08:45:14 crc kubenswrapper[4860]: I1211 08:45:14.390956 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-l6c22" Dec 11 08:45:14 crc kubenswrapper[4860]: I1211 08:45:14.391030 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l6c22" event={"ID":"81ffd57d-3106-4052-b985-5a181330c8ec","Type":"ContainerDied","Data":"2d5133970fa366d202727c5f5fc290958773d99326eb9eda8b411b6547d0431e"} Dec 11 08:45:14 crc kubenswrapper[4860]: I1211 08:45:14.391066 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l6c22" event={"ID":"81ffd57d-3106-4052-b985-5a181330c8ec","Type":"ContainerDied","Data":"3a3b1755907ba5dd6ea064b7f02fe6e6ef3c78b180c14f259c0c86f650572af9"} Dec 11 08:45:14 crc kubenswrapper[4860]: I1211 08:45:14.391094 4860 scope.go:117] "RemoveContainer" containerID="2d5133970fa366d202727c5f5fc290958773d99326eb9eda8b411b6547d0431e" Dec 11 08:45:14 crc kubenswrapper[4860]: I1211 08:45:14.395326 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gs2nr" event={"ID":"3847111f-a08e-4089-8ee7-74648c6f43f6","Type":"ContainerStarted","Data":"52cf54e402797f20d4d332616630197db1d45d282e4aae64b5d8c290bc706166"} Dec 11 08:45:14 crc kubenswrapper[4860]: I1211 08:45:14.412300 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-gs2nr" podStartSLOduration=3.251591097 podStartE2EDuration="8.412261873s" podCreationTimestamp="2025-12-11 08:45:06 +0000 UTC" firstStartedPulling="2025-12-11 08:45:08.327552178 +0000 UTC m=+2041.056071233" lastFinishedPulling="2025-12-11 08:45:13.488222954 +0000 UTC m=+2046.216742009" observedRunningTime="2025-12-11 08:45:14.411359616 +0000 UTC m=+2047.139878701" watchObservedRunningTime="2025-12-11 08:45:14.412261873 +0000 UTC m=+2047.140780948" Dec 11 08:45:14 crc kubenswrapper[4860]: I1211 08:45:14.414801 4860 scope.go:117] "RemoveContainer" containerID="2c2b907fd49217f073e205d1b82a78dbbac052dd2e097195abb0e5dc2debcca1" Dec 11 08:45:14 crc kubenswrapper[4860]: I1211 08:45:14.443512 4860 scope.go:117] "RemoveContainer" containerID="cbcb517dcca6956b63dd1ec30c8af4fdc83fa2b5a5391311a6aa5f2a56f7eeba" Dec 11 08:45:14 crc kubenswrapper[4860]: I1211 08:45:14.448800 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-l6c22"] Dec 11 08:45:14 crc kubenswrapper[4860]: I1211 08:45:14.459777 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-l6c22"] Dec 11 08:45:14 crc kubenswrapper[4860]: I1211 08:45:14.465456 4860 scope.go:117] "RemoveContainer" containerID="2d5133970fa366d202727c5f5fc290958773d99326eb9eda8b411b6547d0431e" Dec 11 08:45:14 crc kubenswrapper[4860]: E1211 08:45:14.465838 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2d5133970fa366d202727c5f5fc290958773d99326eb9eda8b411b6547d0431e\": container with ID starting with 2d5133970fa366d202727c5f5fc290958773d99326eb9eda8b411b6547d0431e not found: ID does not exist" containerID="2d5133970fa366d202727c5f5fc290958773d99326eb9eda8b411b6547d0431e" Dec 11 08:45:14 crc kubenswrapper[4860]: I1211 08:45:14.465872 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d5133970fa366d202727c5f5fc290958773d99326eb9eda8b411b6547d0431e"} err="failed to get container status \"2d5133970fa366d202727c5f5fc290958773d99326eb9eda8b411b6547d0431e\": rpc error: code = NotFound desc = could not find container \"2d5133970fa366d202727c5f5fc290958773d99326eb9eda8b411b6547d0431e\": container with ID starting with 2d5133970fa366d202727c5f5fc290958773d99326eb9eda8b411b6547d0431e not found: ID does not exist" Dec 11 08:45:14 crc kubenswrapper[4860]: I1211 08:45:14.465896 4860 scope.go:117] "RemoveContainer" containerID="2c2b907fd49217f073e205d1b82a78dbbac052dd2e097195abb0e5dc2debcca1" Dec 11 08:45:14 crc kubenswrapper[4860]: E1211 08:45:14.466225 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2c2b907fd49217f073e205d1b82a78dbbac052dd2e097195abb0e5dc2debcca1\": container with ID starting with 2c2b907fd49217f073e205d1b82a78dbbac052dd2e097195abb0e5dc2debcca1 not found: ID does not exist" containerID="2c2b907fd49217f073e205d1b82a78dbbac052dd2e097195abb0e5dc2debcca1" Dec 11 08:45:14 crc kubenswrapper[4860]: I1211 08:45:14.466258 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2c2b907fd49217f073e205d1b82a78dbbac052dd2e097195abb0e5dc2debcca1"} err="failed to get container status \"2c2b907fd49217f073e205d1b82a78dbbac052dd2e097195abb0e5dc2debcca1\": rpc error: code = NotFound desc = could not find container \"2c2b907fd49217f073e205d1b82a78dbbac052dd2e097195abb0e5dc2debcca1\": container with ID starting with 2c2b907fd49217f073e205d1b82a78dbbac052dd2e097195abb0e5dc2debcca1 not found: ID does not exist" Dec 11 08:45:14 crc kubenswrapper[4860]: I1211 08:45:14.466276 4860 scope.go:117] "RemoveContainer" containerID="cbcb517dcca6956b63dd1ec30c8af4fdc83fa2b5a5391311a6aa5f2a56f7eeba" Dec 11 08:45:14 crc kubenswrapper[4860]: E1211 08:45:14.467164 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cbcb517dcca6956b63dd1ec30c8af4fdc83fa2b5a5391311a6aa5f2a56f7eeba\": container with ID starting with cbcb517dcca6956b63dd1ec30c8af4fdc83fa2b5a5391311a6aa5f2a56f7eeba not found: ID does not exist" containerID="cbcb517dcca6956b63dd1ec30c8af4fdc83fa2b5a5391311a6aa5f2a56f7eeba" Dec 11 08:45:14 crc kubenswrapper[4860]: I1211 08:45:14.467192 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cbcb517dcca6956b63dd1ec30c8af4fdc83fa2b5a5391311a6aa5f2a56f7eeba"} err="failed to get container status \"cbcb517dcca6956b63dd1ec30c8af4fdc83fa2b5a5391311a6aa5f2a56f7eeba\": rpc error: code = NotFound desc = could not find container \"cbcb517dcca6956b63dd1ec30c8af4fdc83fa2b5a5391311a6aa5f2a56f7eeba\": container with ID starting with cbcb517dcca6956b63dd1ec30c8af4fdc83fa2b5a5391311a6aa5f2a56f7eeba not found: ID does not exist" Dec 11 08:45:15 crc kubenswrapper[4860]: I1211 08:45:15.590159 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="81ffd57d-3106-4052-b985-5a181330c8ec" path="/var/lib/kubelet/pods/81ffd57d-3106-4052-b985-5a181330c8ec/volumes" Dec 11 08:45:16 crc kubenswrapper[4860]: I1211 08:45:16.838261 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-gs2nr" Dec 11 08:45:16 crc kubenswrapper[4860]: I1211 08:45:16.838602 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-gs2nr" Dec 11 08:45:17 crc kubenswrapper[4860]: I1211 08:45:17.885438 4860 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-gs2nr" podUID="3847111f-a08e-4089-8ee7-74648c6f43f6" containerName="registry-server" probeResult="failure" output=< Dec 11 08:45:17 crc kubenswrapper[4860]: timeout: failed to connect service ":50051" within 1s Dec 11 08:45:17 crc kubenswrapper[4860]: > Dec 11 08:45:26 crc kubenswrapper[4860]: I1211 08:45:26.894624 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-gs2nr" Dec 11 08:45:26 crc kubenswrapper[4860]: I1211 08:45:26.942413 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-gs2nr" Dec 11 08:45:27 crc kubenswrapper[4860]: I1211 08:45:27.127456 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-gs2nr"] Dec 11 08:45:28 crc kubenswrapper[4860]: I1211 08:45:28.536848 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-gs2nr" podUID="3847111f-a08e-4089-8ee7-74648c6f43f6" containerName="registry-server" containerID="cri-o://52cf54e402797f20d4d332616630197db1d45d282e4aae64b5d8c290bc706166" gracePeriod=2 Dec 11 08:45:29 crc kubenswrapper[4860]: I1211 08:45:29.010546 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gs2nr" Dec 11 08:45:29 crc kubenswrapper[4860]: I1211 08:45:29.033957 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3847111f-a08e-4089-8ee7-74648c6f43f6-catalog-content\") pod \"3847111f-a08e-4089-8ee7-74648c6f43f6\" (UID: \"3847111f-a08e-4089-8ee7-74648c6f43f6\") " Dec 11 08:45:29 crc kubenswrapper[4860]: I1211 08:45:29.034040 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3847111f-a08e-4089-8ee7-74648c6f43f6-utilities\") pod \"3847111f-a08e-4089-8ee7-74648c6f43f6\" (UID: \"3847111f-a08e-4089-8ee7-74648c6f43f6\") " Dec 11 08:45:29 crc kubenswrapper[4860]: I1211 08:45:29.034157 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xk6tm\" (UniqueName: \"kubernetes.io/projected/3847111f-a08e-4089-8ee7-74648c6f43f6-kube-api-access-xk6tm\") pod \"3847111f-a08e-4089-8ee7-74648c6f43f6\" (UID: \"3847111f-a08e-4089-8ee7-74648c6f43f6\") " Dec 11 08:45:29 crc kubenswrapper[4860]: I1211 08:45:29.036685 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3847111f-a08e-4089-8ee7-74648c6f43f6-utilities" (OuterVolumeSpecName: "utilities") pod "3847111f-a08e-4089-8ee7-74648c6f43f6" (UID: "3847111f-a08e-4089-8ee7-74648c6f43f6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:45:29 crc kubenswrapper[4860]: I1211 08:45:29.040942 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3847111f-a08e-4089-8ee7-74648c6f43f6-kube-api-access-xk6tm" (OuterVolumeSpecName: "kube-api-access-xk6tm") pod "3847111f-a08e-4089-8ee7-74648c6f43f6" (UID: "3847111f-a08e-4089-8ee7-74648c6f43f6"). InnerVolumeSpecName "kube-api-access-xk6tm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:45:29 crc kubenswrapper[4860]: I1211 08:45:29.137448 4860 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3847111f-a08e-4089-8ee7-74648c6f43f6-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 08:45:29 crc kubenswrapper[4860]: I1211 08:45:29.142735 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xk6tm\" (UniqueName: \"kubernetes.io/projected/3847111f-a08e-4089-8ee7-74648c6f43f6-kube-api-access-xk6tm\") on node \"crc\" DevicePath \"\"" Dec 11 08:45:29 crc kubenswrapper[4860]: I1211 08:45:29.165393 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3847111f-a08e-4089-8ee7-74648c6f43f6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3847111f-a08e-4089-8ee7-74648c6f43f6" (UID: "3847111f-a08e-4089-8ee7-74648c6f43f6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:45:29 crc kubenswrapper[4860]: I1211 08:45:29.245392 4860 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3847111f-a08e-4089-8ee7-74648c6f43f6-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 08:45:29 crc kubenswrapper[4860]: I1211 08:45:29.552518 4860 generic.go:334] "Generic (PLEG): container finished" podID="3847111f-a08e-4089-8ee7-74648c6f43f6" containerID="52cf54e402797f20d4d332616630197db1d45d282e4aae64b5d8c290bc706166" exitCode=0 Dec 11 08:45:29 crc kubenswrapper[4860]: I1211 08:45:29.552613 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gs2nr" Dec 11 08:45:29 crc kubenswrapper[4860]: I1211 08:45:29.552616 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gs2nr" event={"ID":"3847111f-a08e-4089-8ee7-74648c6f43f6","Type":"ContainerDied","Data":"52cf54e402797f20d4d332616630197db1d45d282e4aae64b5d8c290bc706166"} Dec 11 08:45:29 crc kubenswrapper[4860]: I1211 08:45:29.554022 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gs2nr" event={"ID":"3847111f-a08e-4089-8ee7-74648c6f43f6","Type":"ContainerDied","Data":"a632794a228c434be8560a4be8d2cd0eb733f8fbb677427d4e252b95094e7271"} Dec 11 08:45:29 crc kubenswrapper[4860]: I1211 08:45:29.554102 4860 scope.go:117] "RemoveContainer" containerID="52cf54e402797f20d4d332616630197db1d45d282e4aae64b5d8c290bc706166" Dec 11 08:45:29 crc kubenswrapper[4860]: I1211 08:45:29.580963 4860 scope.go:117] "RemoveContainer" containerID="4cba2cf73c11517188c1be2665dab18ac13ee318240e134f06d866bf3f909730" Dec 11 08:45:29 crc kubenswrapper[4860]: I1211 08:45:29.598758 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-gs2nr"] Dec 11 08:45:29 crc kubenswrapper[4860]: I1211 08:45:29.598795 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-gs2nr"] Dec 11 08:45:29 crc kubenswrapper[4860]: I1211 08:45:29.608438 4860 scope.go:117] "RemoveContainer" containerID="968584d44e01eacc893a8b322c54f726010ab703d15f4485b97327f52fa355a1" Dec 11 08:45:29 crc kubenswrapper[4860]: I1211 08:45:29.659970 4860 scope.go:117] "RemoveContainer" containerID="52cf54e402797f20d4d332616630197db1d45d282e4aae64b5d8c290bc706166" Dec 11 08:45:29 crc kubenswrapper[4860]: E1211 08:45:29.660623 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"52cf54e402797f20d4d332616630197db1d45d282e4aae64b5d8c290bc706166\": container with ID starting with 52cf54e402797f20d4d332616630197db1d45d282e4aae64b5d8c290bc706166 not found: ID does not exist" containerID="52cf54e402797f20d4d332616630197db1d45d282e4aae64b5d8c290bc706166" Dec 11 08:45:29 crc kubenswrapper[4860]: I1211 08:45:29.660694 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"52cf54e402797f20d4d332616630197db1d45d282e4aae64b5d8c290bc706166"} err="failed to get container status \"52cf54e402797f20d4d332616630197db1d45d282e4aae64b5d8c290bc706166\": rpc error: code = NotFound desc = could not find container \"52cf54e402797f20d4d332616630197db1d45d282e4aae64b5d8c290bc706166\": container with ID starting with 52cf54e402797f20d4d332616630197db1d45d282e4aae64b5d8c290bc706166 not found: ID does not exist" Dec 11 08:45:29 crc kubenswrapper[4860]: I1211 08:45:29.660725 4860 scope.go:117] "RemoveContainer" containerID="4cba2cf73c11517188c1be2665dab18ac13ee318240e134f06d866bf3f909730" Dec 11 08:45:29 crc kubenswrapper[4860]: E1211 08:45:29.661179 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4cba2cf73c11517188c1be2665dab18ac13ee318240e134f06d866bf3f909730\": container with ID starting with 4cba2cf73c11517188c1be2665dab18ac13ee318240e134f06d866bf3f909730 not found: ID does not exist" containerID="4cba2cf73c11517188c1be2665dab18ac13ee318240e134f06d866bf3f909730" Dec 11 08:45:29 crc kubenswrapper[4860]: I1211 08:45:29.661215 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4cba2cf73c11517188c1be2665dab18ac13ee318240e134f06d866bf3f909730"} err="failed to get container status \"4cba2cf73c11517188c1be2665dab18ac13ee318240e134f06d866bf3f909730\": rpc error: code = NotFound desc = could not find container \"4cba2cf73c11517188c1be2665dab18ac13ee318240e134f06d866bf3f909730\": container with ID starting with 4cba2cf73c11517188c1be2665dab18ac13ee318240e134f06d866bf3f909730 not found: ID does not exist" Dec 11 08:45:29 crc kubenswrapper[4860]: I1211 08:45:29.661234 4860 scope.go:117] "RemoveContainer" containerID="968584d44e01eacc893a8b322c54f726010ab703d15f4485b97327f52fa355a1" Dec 11 08:45:29 crc kubenswrapper[4860]: E1211 08:45:29.663169 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"968584d44e01eacc893a8b322c54f726010ab703d15f4485b97327f52fa355a1\": container with ID starting with 968584d44e01eacc893a8b322c54f726010ab703d15f4485b97327f52fa355a1 not found: ID does not exist" containerID="968584d44e01eacc893a8b322c54f726010ab703d15f4485b97327f52fa355a1" Dec 11 08:45:29 crc kubenswrapper[4860]: I1211 08:45:29.663212 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"968584d44e01eacc893a8b322c54f726010ab703d15f4485b97327f52fa355a1"} err="failed to get container status \"968584d44e01eacc893a8b322c54f726010ab703d15f4485b97327f52fa355a1\": rpc error: code = NotFound desc = could not find container \"968584d44e01eacc893a8b322c54f726010ab703d15f4485b97327f52fa355a1\": container with ID starting with 968584d44e01eacc893a8b322c54f726010ab703d15f4485b97327f52fa355a1 not found: ID does not exist" Dec 11 08:45:31 crc kubenswrapper[4860]: I1211 08:45:31.596839 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3847111f-a08e-4089-8ee7-74648c6f43f6" path="/var/lib/kubelet/pods/3847111f-a08e-4089-8ee7-74648c6f43f6/volumes" Dec 11 08:45:33 crc kubenswrapper[4860]: I1211 08:45:33.507098 4860 scope.go:117] "RemoveContainer" containerID="f5f79575a837d49cfff4df714dbab8b7364ce4d99d9446a09a0108cb3027e312" Dec 11 08:45:39 crc kubenswrapper[4860]: I1211 08:45:39.663437 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-g4h74"] Dec 11 08:45:39 crc kubenswrapper[4860]: E1211 08:45:39.664499 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81ffd57d-3106-4052-b985-5a181330c8ec" containerName="registry-server" Dec 11 08:45:39 crc kubenswrapper[4860]: I1211 08:45:39.664512 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="81ffd57d-3106-4052-b985-5a181330c8ec" containerName="registry-server" Dec 11 08:45:39 crc kubenswrapper[4860]: E1211 08:45:39.664530 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3847111f-a08e-4089-8ee7-74648c6f43f6" containerName="registry-server" Dec 11 08:45:39 crc kubenswrapper[4860]: I1211 08:45:39.664536 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="3847111f-a08e-4089-8ee7-74648c6f43f6" containerName="registry-server" Dec 11 08:45:39 crc kubenswrapper[4860]: E1211 08:45:39.664545 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81ffd57d-3106-4052-b985-5a181330c8ec" containerName="extract-content" Dec 11 08:45:39 crc kubenswrapper[4860]: I1211 08:45:39.664551 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="81ffd57d-3106-4052-b985-5a181330c8ec" containerName="extract-content" Dec 11 08:45:39 crc kubenswrapper[4860]: E1211 08:45:39.664574 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81ffd57d-3106-4052-b985-5a181330c8ec" containerName="extract-utilities" Dec 11 08:45:39 crc kubenswrapper[4860]: I1211 08:45:39.664580 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="81ffd57d-3106-4052-b985-5a181330c8ec" containerName="extract-utilities" Dec 11 08:45:39 crc kubenswrapper[4860]: E1211 08:45:39.664594 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3847111f-a08e-4089-8ee7-74648c6f43f6" containerName="extract-content" Dec 11 08:45:39 crc kubenswrapper[4860]: I1211 08:45:39.664601 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="3847111f-a08e-4089-8ee7-74648c6f43f6" containerName="extract-content" Dec 11 08:45:39 crc kubenswrapper[4860]: E1211 08:45:39.664613 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3847111f-a08e-4089-8ee7-74648c6f43f6" containerName="extract-utilities" Dec 11 08:45:39 crc kubenswrapper[4860]: I1211 08:45:39.664618 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="3847111f-a08e-4089-8ee7-74648c6f43f6" containerName="extract-utilities" Dec 11 08:45:39 crc kubenswrapper[4860]: I1211 08:45:39.664817 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="3847111f-a08e-4089-8ee7-74648c6f43f6" containerName="registry-server" Dec 11 08:45:39 crc kubenswrapper[4860]: I1211 08:45:39.664838 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="81ffd57d-3106-4052-b985-5a181330c8ec" containerName="registry-server" Dec 11 08:45:39 crc kubenswrapper[4860]: I1211 08:45:39.666167 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-g4h74" Dec 11 08:45:39 crc kubenswrapper[4860]: I1211 08:45:39.685382 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-g4h74"] Dec 11 08:45:39 crc kubenswrapper[4860]: I1211 08:45:39.847232 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/508571f0-279b-45a1-baf5-b8d6527b97bb-utilities\") pod \"certified-operators-g4h74\" (UID: \"508571f0-279b-45a1-baf5-b8d6527b97bb\") " pod="openshift-marketplace/certified-operators-g4h74" Dec 11 08:45:39 crc kubenswrapper[4860]: I1211 08:45:39.847425 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sbk6p\" (UniqueName: \"kubernetes.io/projected/508571f0-279b-45a1-baf5-b8d6527b97bb-kube-api-access-sbk6p\") pod \"certified-operators-g4h74\" (UID: \"508571f0-279b-45a1-baf5-b8d6527b97bb\") " pod="openshift-marketplace/certified-operators-g4h74" Dec 11 08:45:39 crc kubenswrapper[4860]: I1211 08:45:39.847513 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/508571f0-279b-45a1-baf5-b8d6527b97bb-catalog-content\") pod \"certified-operators-g4h74\" (UID: \"508571f0-279b-45a1-baf5-b8d6527b97bb\") " pod="openshift-marketplace/certified-operators-g4h74" Dec 11 08:45:39 crc kubenswrapper[4860]: I1211 08:45:39.950246 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sbk6p\" (UniqueName: \"kubernetes.io/projected/508571f0-279b-45a1-baf5-b8d6527b97bb-kube-api-access-sbk6p\") pod \"certified-operators-g4h74\" (UID: \"508571f0-279b-45a1-baf5-b8d6527b97bb\") " pod="openshift-marketplace/certified-operators-g4h74" Dec 11 08:45:39 crc kubenswrapper[4860]: I1211 08:45:39.950394 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/508571f0-279b-45a1-baf5-b8d6527b97bb-catalog-content\") pod \"certified-operators-g4h74\" (UID: \"508571f0-279b-45a1-baf5-b8d6527b97bb\") " pod="openshift-marketplace/certified-operators-g4h74" Dec 11 08:45:39 crc kubenswrapper[4860]: I1211 08:45:39.950555 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/508571f0-279b-45a1-baf5-b8d6527b97bb-utilities\") pod \"certified-operators-g4h74\" (UID: \"508571f0-279b-45a1-baf5-b8d6527b97bb\") " pod="openshift-marketplace/certified-operators-g4h74" Dec 11 08:45:39 crc kubenswrapper[4860]: I1211 08:45:39.950892 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/508571f0-279b-45a1-baf5-b8d6527b97bb-catalog-content\") pod \"certified-operators-g4h74\" (UID: \"508571f0-279b-45a1-baf5-b8d6527b97bb\") " pod="openshift-marketplace/certified-operators-g4h74" Dec 11 08:45:39 crc kubenswrapper[4860]: I1211 08:45:39.950989 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/508571f0-279b-45a1-baf5-b8d6527b97bb-utilities\") pod \"certified-operators-g4h74\" (UID: \"508571f0-279b-45a1-baf5-b8d6527b97bb\") " pod="openshift-marketplace/certified-operators-g4h74" Dec 11 08:45:39 crc kubenswrapper[4860]: I1211 08:45:39.972749 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sbk6p\" (UniqueName: \"kubernetes.io/projected/508571f0-279b-45a1-baf5-b8d6527b97bb-kube-api-access-sbk6p\") pod \"certified-operators-g4h74\" (UID: \"508571f0-279b-45a1-baf5-b8d6527b97bb\") " pod="openshift-marketplace/certified-operators-g4h74" Dec 11 08:45:39 crc kubenswrapper[4860]: I1211 08:45:39.991916 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-g4h74" Dec 11 08:45:40 crc kubenswrapper[4860]: I1211 08:45:40.480042 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-g4h74"] Dec 11 08:45:40 crc kubenswrapper[4860]: I1211 08:45:40.671530 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g4h74" event={"ID":"508571f0-279b-45a1-baf5-b8d6527b97bb","Type":"ContainerStarted","Data":"d2e4d9976d2741a831542c5b4d93696183903a0d40b0311557fc37304977954c"} Dec 11 08:45:41 crc kubenswrapper[4860]: I1211 08:45:41.681668 4860 generic.go:334] "Generic (PLEG): container finished" podID="508571f0-279b-45a1-baf5-b8d6527b97bb" containerID="d754bf7e4c506ae46455edb8c76c5ac3f62a6e9d78b4937446aed26ac8884b0f" exitCode=0 Dec 11 08:45:41 crc kubenswrapper[4860]: I1211 08:45:41.681718 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g4h74" event={"ID":"508571f0-279b-45a1-baf5-b8d6527b97bb","Type":"ContainerDied","Data":"d754bf7e4c506ae46455edb8c76c5ac3f62a6e9d78b4937446aed26ac8884b0f"} Dec 11 08:45:42 crc kubenswrapper[4860]: I1211 08:45:42.692793 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g4h74" event={"ID":"508571f0-279b-45a1-baf5-b8d6527b97bb","Type":"ContainerStarted","Data":"888c55fbb2b668cbeecb78cd274a385054e490f2ac21f8d96e319b7fc43ceffb"} Dec 11 08:45:43 crc kubenswrapper[4860]: I1211 08:45:43.703678 4860 generic.go:334] "Generic (PLEG): container finished" podID="508571f0-279b-45a1-baf5-b8d6527b97bb" containerID="888c55fbb2b668cbeecb78cd274a385054e490f2ac21f8d96e319b7fc43ceffb" exitCode=0 Dec 11 08:45:43 crc kubenswrapper[4860]: I1211 08:45:43.703767 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g4h74" event={"ID":"508571f0-279b-45a1-baf5-b8d6527b97bb","Type":"ContainerDied","Data":"888c55fbb2b668cbeecb78cd274a385054e490f2ac21f8d96e319b7fc43ceffb"} Dec 11 08:45:44 crc kubenswrapper[4860]: I1211 08:45:44.713262 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g4h74" event={"ID":"508571f0-279b-45a1-baf5-b8d6527b97bb","Type":"ContainerStarted","Data":"e7823e95deb7e4cd8a3bed8fb76c458694176d4f12649f3d849cf84064b5b93e"} Dec 11 08:45:44 crc kubenswrapper[4860]: I1211 08:45:44.735241 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-g4h74" podStartSLOduration=3.039940541 podStartE2EDuration="5.735226209s" podCreationTimestamp="2025-12-11 08:45:39 +0000 UTC" firstStartedPulling="2025-12-11 08:45:41.684976127 +0000 UTC m=+2074.413495182" lastFinishedPulling="2025-12-11 08:45:44.380261795 +0000 UTC m=+2077.108780850" observedRunningTime="2025-12-11 08:45:44.733447024 +0000 UTC m=+2077.461966079" watchObservedRunningTime="2025-12-11 08:45:44.735226209 +0000 UTC m=+2077.463745264" Dec 11 08:45:49 crc kubenswrapper[4860]: I1211 08:45:49.992138 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-g4h74" Dec 11 08:45:49 crc kubenswrapper[4860]: I1211 08:45:49.992627 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-g4h74" Dec 11 08:45:50 crc kubenswrapper[4860]: I1211 08:45:50.038223 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-g4h74" Dec 11 08:45:50 crc kubenswrapper[4860]: I1211 08:45:50.843318 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-g4h74" Dec 11 08:45:51 crc kubenswrapper[4860]: I1211 08:45:51.430223 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-g4h74"] Dec 11 08:45:52 crc kubenswrapper[4860]: I1211 08:45:52.807820 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-g4h74" podUID="508571f0-279b-45a1-baf5-b8d6527b97bb" containerName="registry-server" containerID="cri-o://e7823e95deb7e4cd8a3bed8fb76c458694176d4f12649f3d849cf84064b5b93e" gracePeriod=2 Dec 11 08:45:53 crc kubenswrapper[4860]: I1211 08:45:53.281523 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-g4h74" Dec 11 08:45:53 crc kubenswrapper[4860]: I1211 08:45:53.416009 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sbk6p\" (UniqueName: \"kubernetes.io/projected/508571f0-279b-45a1-baf5-b8d6527b97bb-kube-api-access-sbk6p\") pod \"508571f0-279b-45a1-baf5-b8d6527b97bb\" (UID: \"508571f0-279b-45a1-baf5-b8d6527b97bb\") " Dec 11 08:45:53 crc kubenswrapper[4860]: I1211 08:45:53.417133 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/508571f0-279b-45a1-baf5-b8d6527b97bb-utilities\") pod \"508571f0-279b-45a1-baf5-b8d6527b97bb\" (UID: \"508571f0-279b-45a1-baf5-b8d6527b97bb\") " Dec 11 08:45:53 crc kubenswrapper[4860]: I1211 08:45:53.417196 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/508571f0-279b-45a1-baf5-b8d6527b97bb-catalog-content\") pod \"508571f0-279b-45a1-baf5-b8d6527b97bb\" (UID: \"508571f0-279b-45a1-baf5-b8d6527b97bb\") " Dec 11 08:45:53 crc kubenswrapper[4860]: I1211 08:45:53.417387 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/508571f0-279b-45a1-baf5-b8d6527b97bb-utilities" (OuterVolumeSpecName: "utilities") pod "508571f0-279b-45a1-baf5-b8d6527b97bb" (UID: "508571f0-279b-45a1-baf5-b8d6527b97bb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:45:53 crc kubenswrapper[4860]: I1211 08:45:53.417671 4860 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/508571f0-279b-45a1-baf5-b8d6527b97bb-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 08:45:53 crc kubenswrapper[4860]: I1211 08:45:53.427961 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/508571f0-279b-45a1-baf5-b8d6527b97bb-kube-api-access-sbk6p" (OuterVolumeSpecName: "kube-api-access-sbk6p") pod "508571f0-279b-45a1-baf5-b8d6527b97bb" (UID: "508571f0-279b-45a1-baf5-b8d6527b97bb"). InnerVolumeSpecName "kube-api-access-sbk6p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:45:53 crc kubenswrapper[4860]: I1211 08:45:53.474966 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/508571f0-279b-45a1-baf5-b8d6527b97bb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "508571f0-279b-45a1-baf5-b8d6527b97bb" (UID: "508571f0-279b-45a1-baf5-b8d6527b97bb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:45:53 crc kubenswrapper[4860]: I1211 08:45:53.519378 4860 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/508571f0-279b-45a1-baf5-b8d6527b97bb-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 08:45:53 crc kubenswrapper[4860]: I1211 08:45:53.519415 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sbk6p\" (UniqueName: \"kubernetes.io/projected/508571f0-279b-45a1-baf5-b8d6527b97bb-kube-api-access-sbk6p\") on node \"crc\" DevicePath \"\"" Dec 11 08:45:53 crc kubenswrapper[4860]: I1211 08:45:53.817330 4860 generic.go:334] "Generic (PLEG): container finished" podID="508571f0-279b-45a1-baf5-b8d6527b97bb" containerID="e7823e95deb7e4cd8a3bed8fb76c458694176d4f12649f3d849cf84064b5b93e" exitCode=0 Dec 11 08:45:53 crc kubenswrapper[4860]: I1211 08:45:53.817381 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g4h74" event={"ID":"508571f0-279b-45a1-baf5-b8d6527b97bb","Type":"ContainerDied","Data":"e7823e95deb7e4cd8a3bed8fb76c458694176d4f12649f3d849cf84064b5b93e"} Dec 11 08:45:53 crc kubenswrapper[4860]: I1211 08:45:53.817416 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g4h74" event={"ID":"508571f0-279b-45a1-baf5-b8d6527b97bb","Type":"ContainerDied","Data":"d2e4d9976d2741a831542c5b4d93696183903a0d40b0311557fc37304977954c"} Dec 11 08:45:53 crc kubenswrapper[4860]: I1211 08:45:53.817439 4860 scope.go:117] "RemoveContainer" containerID="e7823e95deb7e4cd8a3bed8fb76c458694176d4f12649f3d849cf84064b5b93e" Dec 11 08:45:53 crc kubenswrapper[4860]: I1211 08:45:53.817440 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-g4h74" Dec 11 08:45:53 crc kubenswrapper[4860]: I1211 08:45:53.845126 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-g4h74"] Dec 11 08:45:53 crc kubenswrapper[4860]: I1211 08:45:53.845866 4860 scope.go:117] "RemoveContainer" containerID="888c55fbb2b668cbeecb78cd274a385054e490f2ac21f8d96e319b7fc43ceffb" Dec 11 08:45:53 crc kubenswrapper[4860]: I1211 08:45:53.864514 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-g4h74"] Dec 11 08:45:53 crc kubenswrapper[4860]: I1211 08:45:53.869887 4860 scope.go:117] "RemoveContainer" containerID="d754bf7e4c506ae46455edb8c76c5ac3f62a6e9d78b4937446aed26ac8884b0f" Dec 11 08:45:53 crc kubenswrapper[4860]: I1211 08:45:53.912296 4860 scope.go:117] "RemoveContainer" containerID="e7823e95deb7e4cd8a3bed8fb76c458694176d4f12649f3d849cf84064b5b93e" Dec 11 08:45:53 crc kubenswrapper[4860]: E1211 08:45:53.912831 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e7823e95deb7e4cd8a3bed8fb76c458694176d4f12649f3d849cf84064b5b93e\": container with ID starting with e7823e95deb7e4cd8a3bed8fb76c458694176d4f12649f3d849cf84064b5b93e not found: ID does not exist" containerID="e7823e95deb7e4cd8a3bed8fb76c458694176d4f12649f3d849cf84064b5b93e" Dec 11 08:45:53 crc kubenswrapper[4860]: I1211 08:45:53.912870 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e7823e95deb7e4cd8a3bed8fb76c458694176d4f12649f3d849cf84064b5b93e"} err="failed to get container status \"e7823e95deb7e4cd8a3bed8fb76c458694176d4f12649f3d849cf84064b5b93e\": rpc error: code = NotFound desc = could not find container \"e7823e95deb7e4cd8a3bed8fb76c458694176d4f12649f3d849cf84064b5b93e\": container with ID starting with e7823e95deb7e4cd8a3bed8fb76c458694176d4f12649f3d849cf84064b5b93e not found: ID does not exist" Dec 11 08:45:53 crc kubenswrapper[4860]: I1211 08:45:53.912899 4860 scope.go:117] "RemoveContainer" containerID="888c55fbb2b668cbeecb78cd274a385054e490f2ac21f8d96e319b7fc43ceffb" Dec 11 08:45:53 crc kubenswrapper[4860]: E1211 08:45:53.913296 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"888c55fbb2b668cbeecb78cd274a385054e490f2ac21f8d96e319b7fc43ceffb\": container with ID starting with 888c55fbb2b668cbeecb78cd274a385054e490f2ac21f8d96e319b7fc43ceffb not found: ID does not exist" containerID="888c55fbb2b668cbeecb78cd274a385054e490f2ac21f8d96e319b7fc43ceffb" Dec 11 08:45:53 crc kubenswrapper[4860]: I1211 08:45:53.913330 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"888c55fbb2b668cbeecb78cd274a385054e490f2ac21f8d96e319b7fc43ceffb"} err="failed to get container status \"888c55fbb2b668cbeecb78cd274a385054e490f2ac21f8d96e319b7fc43ceffb\": rpc error: code = NotFound desc = could not find container \"888c55fbb2b668cbeecb78cd274a385054e490f2ac21f8d96e319b7fc43ceffb\": container with ID starting with 888c55fbb2b668cbeecb78cd274a385054e490f2ac21f8d96e319b7fc43ceffb not found: ID does not exist" Dec 11 08:45:53 crc kubenswrapper[4860]: I1211 08:45:53.913347 4860 scope.go:117] "RemoveContainer" containerID="d754bf7e4c506ae46455edb8c76c5ac3f62a6e9d78b4937446aed26ac8884b0f" Dec 11 08:45:53 crc kubenswrapper[4860]: E1211 08:45:53.913722 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d754bf7e4c506ae46455edb8c76c5ac3f62a6e9d78b4937446aed26ac8884b0f\": container with ID starting with d754bf7e4c506ae46455edb8c76c5ac3f62a6e9d78b4937446aed26ac8884b0f not found: ID does not exist" containerID="d754bf7e4c506ae46455edb8c76c5ac3f62a6e9d78b4937446aed26ac8884b0f" Dec 11 08:45:53 crc kubenswrapper[4860]: I1211 08:45:53.913773 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d754bf7e4c506ae46455edb8c76c5ac3f62a6e9d78b4937446aed26ac8884b0f"} err="failed to get container status \"d754bf7e4c506ae46455edb8c76c5ac3f62a6e9d78b4937446aed26ac8884b0f\": rpc error: code = NotFound desc = could not find container \"d754bf7e4c506ae46455edb8c76c5ac3f62a6e9d78b4937446aed26ac8884b0f\": container with ID starting with d754bf7e4c506ae46455edb8c76c5ac3f62a6e9d78b4937446aed26ac8884b0f not found: ID does not exist" Dec 11 08:45:55 crc kubenswrapper[4860]: I1211 08:45:55.615763 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="508571f0-279b-45a1-baf5-b8d6527b97bb" path="/var/lib/kubelet/pods/508571f0-279b-45a1-baf5-b8d6527b97bb/volumes" Dec 11 08:46:38 crc kubenswrapper[4860]: I1211 08:46:38.794912 4860 patch_prober.go:28] interesting pod/machine-config-daemon-99qgp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 08:46:38 crc kubenswrapper[4860]: I1211 08:46:38.795551 4860 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 08:47:06 crc kubenswrapper[4860]: I1211 08:47:06.752266 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-nwvfx"] Dec 11 08:47:06 crc kubenswrapper[4860]: E1211 08:47:06.755198 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="508571f0-279b-45a1-baf5-b8d6527b97bb" containerName="extract-content" Dec 11 08:47:06 crc kubenswrapper[4860]: I1211 08:47:06.755298 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="508571f0-279b-45a1-baf5-b8d6527b97bb" containerName="extract-content" Dec 11 08:47:06 crc kubenswrapper[4860]: E1211 08:47:06.755391 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="508571f0-279b-45a1-baf5-b8d6527b97bb" containerName="registry-server" Dec 11 08:47:06 crc kubenswrapper[4860]: I1211 08:47:06.755464 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="508571f0-279b-45a1-baf5-b8d6527b97bb" containerName="registry-server" Dec 11 08:47:06 crc kubenswrapper[4860]: E1211 08:47:06.755555 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="508571f0-279b-45a1-baf5-b8d6527b97bb" containerName="extract-utilities" Dec 11 08:47:06 crc kubenswrapper[4860]: I1211 08:47:06.755627 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="508571f0-279b-45a1-baf5-b8d6527b97bb" containerName="extract-utilities" Dec 11 08:47:06 crc kubenswrapper[4860]: I1211 08:47:06.755997 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="508571f0-279b-45a1-baf5-b8d6527b97bb" containerName="registry-server" Dec 11 08:47:06 crc kubenswrapper[4860]: I1211 08:47:06.757538 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nwvfx" Dec 11 08:47:06 crc kubenswrapper[4860]: I1211 08:47:06.786779 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nwvfx"] Dec 11 08:47:06 crc kubenswrapper[4860]: I1211 08:47:06.899730 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9fc23803-fbf5-4b31-9aaa-7bd4a5d42903-utilities\") pod \"redhat-marketplace-nwvfx\" (UID: \"9fc23803-fbf5-4b31-9aaa-7bd4a5d42903\") " pod="openshift-marketplace/redhat-marketplace-nwvfx" Dec 11 08:47:06 crc kubenswrapper[4860]: I1211 08:47:06.899824 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9fc23803-fbf5-4b31-9aaa-7bd4a5d42903-catalog-content\") pod \"redhat-marketplace-nwvfx\" (UID: \"9fc23803-fbf5-4b31-9aaa-7bd4a5d42903\") " pod="openshift-marketplace/redhat-marketplace-nwvfx" Dec 11 08:47:06 crc kubenswrapper[4860]: I1211 08:47:06.899920 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-944jc\" (UniqueName: \"kubernetes.io/projected/9fc23803-fbf5-4b31-9aaa-7bd4a5d42903-kube-api-access-944jc\") pod \"redhat-marketplace-nwvfx\" (UID: \"9fc23803-fbf5-4b31-9aaa-7bd4a5d42903\") " pod="openshift-marketplace/redhat-marketplace-nwvfx" Dec 11 08:47:07 crc kubenswrapper[4860]: I1211 08:47:07.001520 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9fc23803-fbf5-4b31-9aaa-7bd4a5d42903-utilities\") pod \"redhat-marketplace-nwvfx\" (UID: \"9fc23803-fbf5-4b31-9aaa-7bd4a5d42903\") " pod="openshift-marketplace/redhat-marketplace-nwvfx" Dec 11 08:47:07 crc kubenswrapper[4860]: I1211 08:47:07.002170 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9fc23803-fbf5-4b31-9aaa-7bd4a5d42903-utilities\") pod \"redhat-marketplace-nwvfx\" (UID: \"9fc23803-fbf5-4b31-9aaa-7bd4a5d42903\") " pod="openshift-marketplace/redhat-marketplace-nwvfx" Dec 11 08:47:07 crc kubenswrapper[4860]: I1211 08:47:07.002284 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9fc23803-fbf5-4b31-9aaa-7bd4a5d42903-catalog-content\") pod \"redhat-marketplace-nwvfx\" (UID: \"9fc23803-fbf5-4b31-9aaa-7bd4a5d42903\") " pod="openshift-marketplace/redhat-marketplace-nwvfx" Dec 11 08:47:07 crc kubenswrapper[4860]: I1211 08:47:07.002404 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-944jc\" (UniqueName: \"kubernetes.io/projected/9fc23803-fbf5-4b31-9aaa-7bd4a5d42903-kube-api-access-944jc\") pod \"redhat-marketplace-nwvfx\" (UID: \"9fc23803-fbf5-4b31-9aaa-7bd4a5d42903\") " pod="openshift-marketplace/redhat-marketplace-nwvfx" Dec 11 08:47:07 crc kubenswrapper[4860]: I1211 08:47:07.002603 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9fc23803-fbf5-4b31-9aaa-7bd4a5d42903-catalog-content\") pod \"redhat-marketplace-nwvfx\" (UID: \"9fc23803-fbf5-4b31-9aaa-7bd4a5d42903\") " pod="openshift-marketplace/redhat-marketplace-nwvfx" Dec 11 08:47:07 crc kubenswrapper[4860]: I1211 08:47:07.022635 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-944jc\" (UniqueName: \"kubernetes.io/projected/9fc23803-fbf5-4b31-9aaa-7bd4a5d42903-kube-api-access-944jc\") pod \"redhat-marketplace-nwvfx\" (UID: \"9fc23803-fbf5-4b31-9aaa-7bd4a5d42903\") " pod="openshift-marketplace/redhat-marketplace-nwvfx" Dec 11 08:47:07 crc kubenswrapper[4860]: I1211 08:47:07.092465 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nwvfx" Dec 11 08:47:07 crc kubenswrapper[4860]: I1211 08:47:07.618466 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nwvfx"] Dec 11 08:47:08 crc kubenswrapper[4860]: I1211 08:47:08.533319 4860 generic.go:334] "Generic (PLEG): container finished" podID="9fc23803-fbf5-4b31-9aaa-7bd4a5d42903" containerID="66ad550e94ce18fb66387660e0710214e3ffad2b401d74c9e24d91000027591f" exitCode=0 Dec 11 08:47:08 crc kubenswrapper[4860]: I1211 08:47:08.533810 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nwvfx" event={"ID":"9fc23803-fbf5-4b31-9aaa-7bd4a5d42903","Type":"ContainerDied","Data":"66ad550e94ce18fb66387660e0710214e3ffad2b401d74c9e24d91000027591f"} Dec 11 08:47:08 crc kubenswrapper[4860]: I1211 08:47:08.534138 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nwvfx" event={"ID":"9fc23803-fbf5-4b31-9aaa-7bd4a5d42903","Type":"ContainerStarted","Data":"4b0c29e1dcd4cbc94bc7b5a24eea94ea87036b805c13d0a44e0572fe10380c13"} Dec 11 08:47:08 crc kubenswrapper[4860]: I1211 08:47:08.795586 4860 patch_prober.go:28] interesting pod/machine-config-daemon-99qgp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 08:47:08 crc kubenswrapper[4860]: I1211 08:47:08.795675 4860 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 08:47:10 crc kubenswrapper[4860]: I1211 08:47:10.555955 4860 generic.go:334] "Generic (PLEG): container finished" podID="9fc23803-fbf5-4b31-9aaa-7bd4a5d42903" containerID="e4a567144c09bb59215ac07ebb72e99d8a8786ebea447edb2fd1389463eb31b2" exitCode=0 Dec 11 08:47:10 crc kubenswrapper[4860]: I1211 08:47:10.556184 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nwvfx" event={"ID":"9fc23803-fbf5-4b31-9aaa-7bd4a5d42903","Type":"ContainerDied","Data":"e4a567144c09bb59215ac07ebb72e99d8a8786ebea447edb2fd1389463eb31b2"} Dec 11 08:47:12 crc kubenswrapper[4860]: I1211 08:47:12.578872 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nwvfx" event={"ID":"9fc23803-fbf5-4b31-9aaa-7bd4a5d42903","Type":"ContainerStarted","Data":"ec708df1fa09efbdb08749532318ad44b9ae9633b0a77737b7b2a257ef3e5d1a"} Dec 11 08:47:12 crc kubenswrapper[4860]: I1211 08:47:12.611234 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-nwvfx" podStartSLOduration=2.933370768 podStartE2EDuration="6.611210383s" podCreationTimestamp="2025-12-11 08:47:06 +0000 UTC" firstStartedPulling="2025-12-11 08:47:08.539635947 +0000 UTC m=+2161.268155022" lastFinishedPulling="2025-12-11 08:47:12.217475542 +0000 UTC m=+2164.945994637" observedRunningTime="2025-12-11 08:47:12.600678292 +0000 UTC m=+2165.329197367" watchObservedRunningTime="2025-12-11 08:47:12.611210383 +0000 UTC m=+2165.339729438" Dec 11 08:47:17 crc kubenswrapper[4860]: I1211 08:47:17.103879 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-nwvfx" Dec 11 08:47:17 crc kubenswrapper[4860]: I1211 08:47:17.104849 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-nwvfx" Dec 11 08:47:17 crc kubenswrapper[4860]: I1211 08:47:17.159487 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-nwvfx" Dec 11 08:47:17 crc kubenswrapper[4860]: I1211 08:47:17.672166 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-nwvfx" Dec 11 08:47:17 crc kubenswrapper[4860]: I1211 08:47:17.728017 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-nwvfx"] Dec 11 08:47:19 crc kubenswrapper[4860]: I1211 08:47:19.647075 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-nwvfx" podUID="9fc23803-fbf5-4b31-9aaa-7bd4a5d42903" containerName="registry-server" containerID="cri-o://ec708df1fa09efbdb08749532318ad44b9ae9633b0a77737b7b2a257ef3e5d1a" gracePeriod=2 Dec 11 08:47:20 crc kubenswrapper[4860]: I1211 08:47:20.109791 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nwvfx" Dec 11 08:47:20 crc kubenswrapper[4860]: I1211 08:47:20.296747 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-944jc\" (UniqueName: \"kubernetes.io/projected/9fc23803-fbf5-4b31-9aaa-7bd4a5d42903-kube-api-access-944jc\") pod \"9fc23803-fbf5-4b31-9aaa-7bd4a5d42903\" (UID: \"9fc23803-fbf5-4b31-9aaa-7bd4a5d42903\") " Dec 11 08:47:20 crc kubenswrapper[4860]: I1211 08:47:20.296924 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9fc23803-fbf5-4b31-9aaa-7bd4a5d42903-utilities\") pod \"9fc23803-fbf5-4b31-9aaa-7bd4a5d42903\" (UID: \"9fc23803-fbf5-4b31-9aaa-7bd4a5d42903\") " Dec 11 08:47:20 crc kubenswrapper[4860]: I1211 08:47:20.297026 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9fc23803-fbf5-4b31-9aaa-7bd4a5d42903-catalog-content\") pod \"9fc23803-fbf5-4b31-9aaa-7bd4a5d42903\" (UID: \"9fc23803-fbf5-4b31-9aaa-7bd4a5d42903\") " Dec 11 08:47:20 crc kubenswrapper[4860]: I1211 08:47:20.297702 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9fc23803-fbf5-4b31-9aaa-7bd4a5d42903-utilities" (OuterVolumeSpecName: "utilities") pod "9fc23803-fbf5-4b31-9aaa-7bd4a5d42903" (UID: "9fc23803-fbf5-4b31-9aaa-7bd4a5d42903"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:47:20 crc kubenswrapper[4860]: I1211 08:47:20.299974 4860 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9fc23803-fbf5-4b31-9aaa-7bd4a5d42903-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 08:47:20 crc kubenswrapper[4860]: I1211 08:47:20.302854 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9fc23803-fbf5-4b31-9aaa-7bd4a5d42903-kube-api-access-944jc" (OuterVolumeSpecName: "kube-api-access-944jc") pod "9fc23803-fbf5-4b31-9aaa-7bd4a5d42903" (UID: "9fc23803-fbf5-4b31-9aaa-7bd4a5d42903"). InnerVolumeSpecName "kube-api-access-944jc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:47:20 crc kubenswrapper[4860]: I1211 08:47:20.331392 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9fc23803-fbf5-4b31-9aaa-7bd4a5d42903-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9fc23803-fbf5-4b31-9aaa-7bd4a5d42903" (UID: "9fc23803-fbf5-4b31-9aaa-7bd4a5d42903"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:47:20 crc kubenswrapper[4860]: I1211 08:47:20.402014 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-944jc\" (UniqueName: \"kubernetes.io/projected/9fc23803-fbf5-4b31-9aaa-7bd4a5d42903-kube-api-access-944jc\") on node \"crc\" DevicePath \"\"" Dec 11 08:47:20 crc kubenswrapper[4860]: I1211 08:47:20.402053 4860 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9fc23803-fbf5-4b31-9aaa-7bd4a5d42903-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 08:47:20 crc kubenswrapper[4860]: I1211 08:47:20.657599 4860 generic.go:334] "Generic (PLEG): container finished" podID="9fc23803-fbf5-4b31-9aaa-7bd4a5d42903" containerID="ec708df1fa09efbdb08749532318ad44b9ae9633b0a77737b7b2a257ef3e5d1a" exitCode=0 Dec 11 08:47:20 crc kubenswrapper[4860]: I1211 08:47:20.657664 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nwvfx" event={"ID":"9fc23803-fbf5-4b31-9aaa-7bd4a5d42903","Type":"ContainerDied","Data":"ec708df1fa09efbdb08749532318ad44b9ae9633b0a77737b7b2a257ef3e5d1a"} Dec 11 08:47:20 crc kubenswrapper[4860]: I1211 08:47:20.657703 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nwvfx" event={"ID":"9fc23803-fbf5-4b31-9aaa-7bd4a5d42903","Type":"ContainerDied","Data":"4b0c29e1dcd4cbc94bc7b5a24eea94ea87036b805c13d0a44e0572fe10380c13"} Dec 11 08:47:20 crc kubenswrapper[4860]: I1211 08:47:20.657753 4860 scope.go:117] "RemoveContainer" containerID="ec708df1fa09efbdb08749532318ad44b9ae9633b0a77737b7b2a257ef3e5d1a" Dec 11 08:47:20 crc kubenswrapper[4860]: I1211 08:47:20.657757 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nwvfx" Dec 11 08:47:20 crc kubenswrapper[4860]: I1211 08:47:20.692346 4860 scope.go:117] "RemoveContainer" containerID="e4a567144c09bb59215ac07ebb72e99d8a8786ebea447edb2fd1389463eb31b2" Dec 11 08:47:20 crc kubenswrapper[4860]: I1211 08:47:20.710131 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-nwvfx"] Dec 11 08:47:20 crc kubenswrapper[4860]: I1211 08:47:20.722103 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-nwvfx"] Dec 11 08:47:20 crc kubenswrapper[4860]: I1211 08:47:20.735790 4860 scope.go:117] "RemoveContainer" containerID="66ad550e94ce18fb66387660e0710214e3ffad2b401d74c9e24d91000027591f" Dec 11 08:47:20 crc kubenswrapper[4860]: I1211 08:47:20.769229 4860 scope.go:117] "RemoveContainer" containerID="ec708df1fa09efbdb08749532318ad44b9ae9633b0a77737b7b2a257ef3e5d1a" Dec 11 08:47:20 crc kubenswrapper[4860]: E1211 08:47:20.770205 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec708df1fa09efbdb08749532318ad44b9ae9633b0a77737b7b2a257ef3e5d1a\": container with ID starting with ec708df1fa09efbdb08749532318ad44b9ae9633b0a77737b7b2a257ef3e5d1a not found: ID does not exist" containerID="ec708df1fa09efbdb08749532318ad44b9ae9633b0a77737b7b2a257ef3e5d1a" Dec 11 08:47:20 crc kubenswrapper[4860]: I1211 08:47:20.770280 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec708df1fa09efbdb08749532318ad44b9ae9633b0a77737b7b2a257ef3e5d1a"} err="failed to get container status \"ec708df1fa09efbdb08749532318ad44b9ae9633b0a77737b7b2a257ef3e5d1a\": rpc error: code = NotFound desc = could not find container \"ec708df1fa09efbdb08749532318ad44b9ae9633b0a77737b7b2a257ef3e5d1a\": container with ID starting with ec708df1fa09efbdb08749532318ad44b9ae9633b0a77737b7b2a257ef3e5d1a not found: ID does not exist" Dec 11 08:47:20 crc kubenswrapper[4860]: I1211 08:47:20.770338 4860 scope.go:117] "RemoveContainer" containerID="e4a567144c09bb59215ac07ebb72e99d8a8786ebea447edb2fd1389463eb31b2" Dec 11 08:47:20 crc kubenswrapper[4860]: E1211 08:47:20.770734 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e4a567144c09bb59215ac07ebb72e99d8a8786ebea447edb2fd1389463eb31b2\": container with ID starting with e4a567144c09bb59215ac07ebb72e99d8a8786ebea447edb2fd1389463eb31b2 not found: ID does not exist" containerID="e4a567144c09bb59215ac07ebb72e99d8a8786ebea447edb2fd1389463eb31b2" Dec 11 08:47:20 crc kubenswrapper[4860]: I1211 08:47:20.770764 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e4a567144c09bb59215ac07ebb72e99d8a8786ebea447edb2fd1389463eb31b2"} err="failed to get container status \"e4a567144c09bb59215ac07ebb72e99d8a8786ebea447edb2fd1389463eb31b2\": rpc error: code = NotFound desc = could not find container \"e4a567144c09bb59215ac07ebb72e99d8a8786ebea447edb2fd1389463eb31b2\": container with ID starting with e4a567144c09bb59215ac07ebb72e99d8a8786ebea447edb2fd1389463eb31b2 not found: ID does not exist" Dec 11 08:47:20 crc kubenswrapper[4860]: I1211 08:47:20.770781 4860 scope.go:117] "RemoveContainer" containerID="66ad550e94ce18fb66387660e0710214e3ffad2b401d74c9e24d91000027591f" Dec 11 08:47:20 crc kubenswrapper[4860]: E1211 08:47:20.771053 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"66ad550e94ce18fb66387660e0710214e3ffad2b401d74c9e24d91000027591f\": container with ID starting with 66ad550e94ce18fb66387660e0710214e3ffad2b401d74c9e24d91000027591f not found: ID does not exist" containerID="66ad550e94ce18fb66387660e0710214e3ffad2b401d74c9e24d91000027591f" Dec 11 08:47:20 crc kubenswrapper[4860]: I1211 08:47:20.771104 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"66ad550e94ce18fb66387660e0710214e3ffad2b401d74c9e24d91000027591f"} err="failed to get container status \"66ad550e94ce18fb66387660e0710214e3ffad2b401d74c9e24d91000027591f\": rpc error: code = NotFound desc = could not find container \"66ad550e94ce18fb66387660e0710214e3ffad2b401d74c9e24d91000027591f\": container with ID starting with 66ad550e94ce18fb66387660e0710214e3ffad2b401d74c9e24d91000027591f not found: ID does not exist" Dec 11 08:47:21 crc kubenswrapper[4860]: I1211 08:47:21.597956 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9fc23803-fbf5-4b31-9aaa-7bd4a5d42903" path="/var/lib/kubelet/pods/9fc23803-fbf5-4b31-9aaa-7bd4a5d42903/volumes" Dec 11 08:47:38 crc kubenswrapper[4860]: I1211 08:47:38.794931 4860 patch_prober.go:28] interesting pod/machine-config-daemon-99qgp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 08:47:38 crc kubenswrapper[4860]: I1211 08:47:38.795625 4860 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 08:47:38 crc kubenswrapper[4860]: I1211 08:47:38.795719 4860 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" Dec 11 08:47:38 crc kubenswrapper[4860]: I1211 08:47:38.796847 4860 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"360fac818f30e5d5bec8360d7df78a5f152a69c33b8db5723f4b538448a80e00"} pod="openshift-machine-config-operator/machine-config-daemon-99qgp" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 11 08:47:38 crc kubenswrapper[4860]: I1211 08:47:38.796968 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" containerName="machine-config-daemon" containerID="cri-o://360fac818f30e5d5bec8360d7df78a5f152a69c33b8db5723f4b538448a80e00" gracePeriod=600 Dec 11 08:47:39 crc kubenswrapper[4860]: I1211 08:47:39.837863 4860 generic.go:334] "Generic (PLEG): container finished" podID="31c30642-6e60-41b4-a477-0d802424e0aa" containerID="360fac818f30e5d5bec8360d7df78a5f152a69c33b8db5723f4b538448a80e00" exitCode=0 Dec 11 08:47:39 crc kubenswrapper[4860]: I1211 08:47:39.837942 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" event={"ID":"31c30642-6e60-41b4-a477-0d802424e0aa","Type":"ContainerDied","Data":"360fac818f30e5d5bec8360d7df78a5f152a69c33b8db5723f4b538448a80e00"} Dec 11 08:47:39 crc kubenswrapper[4860]: I1211 08:47:39.838432 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" event={"ID":"31c30642-6e60-41b4-a477-0d802424e0aa","Type":"ContainerStarted","Data":"b747bbead8584985306f457adcbaf89164c99c29fe92a3c021617afef9e04489"} Dec 11 08:47:39 crc kubenswrapper[4860]: I1211 08:47:39.838455 4860 scope.go:117] "RemoveContainer" containerID="af76095360a11870b804fb7f16c40dad604dbf93152022249293687b1519ca1b" Dec 11 08:49:09 crc kubenswrapper[4860]: I1211 08:49:09.668109 4860 generic.go:334] "Generic (PLEG): container finished" podID="b0eed689-a827-47b6-81ba-91e76db6f35b" containerID="3d85f48b0ba72df4ee804d19f66b272e661e9956e61cf14b2692c6cda570fbe3" exitCode=0 Dec 11 08:49:09 crc kubenswrapper[4860]: I1211 08:49:09.668174 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5z2qw" event={"ID":"b0eed689-a827-47b6-81ba-91e76db6f35b","Type":"ContainerDied","Data":"3d85f48b0ba72df4ee804d19f66b272e661e9956e61cf14b2692c6cda570fbe3"} Dec 11 08:49:11 crc kubenswrapper[4860]: I1211 08:49:11.173792 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5z2qw" Dec 11 08:49:11 crc kubenswrapper[4860]: I1211 08:49:11.331625 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0eed689-a827-47b6-81ba-91e76db6f35b-libvirt-combined-ca-bundle\") pod \"b0eed689-a827-47b6-81ba-91e76db6f35b\" (UID: \"b0eed689-a827-47b6-81ba-91e76db6f35b\") " Dec 11 08:49:11 crc kubenswrapper[4860]: I1211 08:49:11.331744 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b0eed689-a827-47b6-81ba-91e76db6f35b-inventory\") pod \"b0eed689-a827-47b6-81ba-91e76db6f35b\" (UID: \"b0eed689-a827-47b6-81ba-91e76db6f35b\") " Dec 11 08:49:11 crc kubenswrapper[4860]: I1211 08:49:11.331797 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfh9v\" (UniqueName: \"kubernetes.io/projected/b0eed689-a827-47b6-81ba-91e76db6f35b-kube-api-access-kfh9v\") pod \"b0eed689-a827-47b6-81ba-91e76db6f35b\" (UID: \"b0eed689-a827-47b6-81ba-91e76db6f35b\") " Dec 11 08:49:11 crc kubenswrapper[4860]: I1211 08:49:11.331887 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/b0eed689-a827-47b6-81ba-91e76db6f35b-libvirt-secret-0\") pod \"b0eed689-a827-47b6-81ba-91e76db6f35b\" (UID: \"b0eed689-a827-47b6-81ba-91e76db6f35b\") " Dec 11 08:49:11 crc kubenswrapper[4860]: I1211 08:49:11.331949 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b0eed689-a827-47b6-81ba-91e76db6f35b-ssh-key\") pod \"b0eed689-a827-47b6-81ba-91e76db6f35b\" (UID: \"b0eed689-a827-47b6-81ba-91e76db6f35b\") " Dec 11 08:49:11 crc kubenswrapper[4860]: I1211 08:49:11.340784 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b0eed689-a827-47b6-81ba-91e76db6f35b-kube-api-access-kfh9v" (OuterVolumeSpecName: "kube-api-access-kfh9v") pod "b0eed689-a827-47b6-81ba-91e76db6f35b" (UID: "b0eed689-a827-47b6-81ba-91e76db6f35b"). InnerVolumeSpecName "kube-api-access-kfh9v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:49:11 crc kubenswrapper[4860]: I1211 08:49:11.341580 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b0eed689-a827-47b6-81ba-91e76db6f35b-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "b0eed689-a827-47b6-81ba-91e76db6f35b" (UID: "b0eed689-a827-47b6-81ba-91e76db6f35b"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:49:11 crc kubenswrapper[4860]: I1211 08:49:11.364081 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b0eed689-a827-47b6-81ba-91e76db6f35b-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "b0eed689-a827-47b6-81ba-91e76db6f35b" (UID: "b0eed689-a827-47b6-81ba-91e76db6f35b"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:49:11 crc kubenswrapper[4860]: I1211 08:49:11.366209 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b0eed689-a827-47b6-81ba-91e76db6f35b-inventory" (OuterVolumeSpecName: "inventory") pod "b0eed689-a827-47b6-81ba-91e76db6f35b" (UID: "b0eed689-a827-47b6-81ba-91e76db6f35b"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:49:11 crc kubenswrapper[4860]: I1211 08:49:11.368064 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b0eed689-a827-47b6-81ba-91e76db6f35b-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "b0eed689-a827-47b6-81ba-91e76db6f35b" (UID: "b0eed689-a827-47b6-81ba-91e76db6f35b"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:49:11 crc kubenswrapper[4860]: I1211 08:49:11.434573 4860 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0eed689-a827-47b6-81ba-91e76db6f35b-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 08:49:11 crc kubenswrapper[4860]: I1211 08:49:11.434604 4860 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b0eed689-a827-47b6-81ba-91e76db6f35b-inventory\") on node \"crc\" DevicePath \"\"" Dec 11 08:49:11 crc kubenswrapper[4860]: I1211 08:49:11.434615 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfh9v\" (UniqueName: \"kubernetes.io/projected/b0eed689-a827-47b6-81ba-91e76db6f35b-kube-api-access-kfh9v\") on node \"crc\" DevicePath \"\"" Dec 11 08:49:11 crc kubenswrapper[4860]: I1211 08:49:11.434624 4860 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/b0eed689-a827-47b6-81ba-91e76db6f35b-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Dec 11 08:49:11 crc kubenswrapper[4860]: I1211 08:49:11.434632 4860 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b0eed689-a827-47b6-81ba-91e76db6f35b-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 11 08:49:11 crc kubenswrapper[4860]: I1211 08:49:11.691305 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5z2qw" event={"ID":"b0eed689-a827-47b6-81ba-91e76db6f35b","Type":"ContainerDied","Data":"aef7abc9548de978941adae34fe118c396e036943bb572847ba4ceabf84d0761"} Dec 11 08:49:11 crc kubenswrapper[4860]: I1211 08:49:11.691595 4860 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aef7abc9548de978941adae34fe118c396e036943bb572847ba4ceabf84d0761" Dec 11 08:49:11 crc kubenswrapper[4860]: I1211 08:49:11.691377 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5z2qw" Dec 11 08:49:11 crc kubenswrapper[4860]: I1211 08:49:11.787944 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-x42ls"] Dec 11 08:49:11 crc kubenswrapper[4860]: E1211 08:49:11.788437 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9fc23803-fbf5-4b31-9aaa-7bd4a5d42903" containerName="registry-server" Dec 11 08:49:11 crc kubenswrapper[4860]: I1211 08:49:11.788464 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="9fc23803-fbf5-4b31-9aaa-7bd4a5d42903" containerName="registry-server" Dec 11 08:49:11 crc kubenswrapper[4860]: E1211 08:49:11.788500 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0eed689-a827-47b6-81ba-91e76db6f35b" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 11 08:49:11 crc kubenswrapper[4860]: I1211 08:49:11.788510 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0eed689-a827-47b6-81ba-91e76db6f35b" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 11 08:49:11 crc kubenswrapper[4860]: E1211 08:49:11.788523 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9fc23803-fbf5-4b31-9aaa-7bd4a5d42903" containerName="extract-utilities" Dec 11 08:49:11 crc kubenswrapper[4860]: I1211 08:49:11.788531 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="9fc23803-fbf5-4b31-9aaa-7bd4a5d42903" containerName="extract-utilities" Dec 11 08:49:11 crc kubenswrapper[4860]: E1211 08:49:11.788549 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9fc23803-fbf5-4b31-9aaa-7bd4a5d42903" containerName="extract-content" Dec 11 08:49:11 crc kubenswrapper[4860]: I1211 08:49:11.788556 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="9fc23803-fbf5-4b31-9aaa-7bd4a5d42903" containerName="extract-content" Dec 11 08:49:11 crc kubenswrapper[4860]: I1211 08:49:11.788830 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="b0eed689-a827-47b6-81ba-91e76db6f35b" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 11 08:49:11 crc kubenswrapper[4860]: I1211 08:49:11.788859 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="9fc23803-fbf5-4b31-9aaa-7bd4a5d42903" containerName="registry-server" Dec 11 08:49:11 crc kubenswrapper[4860]: I1211 08:49:11.789634 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-x42ls" Dec 11 08:49:11 crc kubenswrapper[4860]: I1211 08:49:11.794275 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Dec 11 08:49:11 crc kubenswrapper[4860]: I1211 08:49:11.794486 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 11 08:49:11 crc kubenswrapper[4860]: I1211 08:49:11.794540 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 11 08:49:11 crc kubenswrapper[4860]: I1211 08:49:11.794542 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Dec 11 08:49:11 crc kubenswrapper[4860]: I1211 08:49:11.794806 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-7w7kh" Dec 11 08:49:11 crc kubenswrapper[4860]: I1211 08:49:11.794846 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Dec 11 08:49:11 crc kubenswrapper[4860]: I1211 08:49:11.794924 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 11 08:49:11 crc kubenswrapper[4860]: I1211 08:49:11.798323 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-x42ls"] Dec 11 08:49:11 crc kubenswrapper[4860]: I1211 08:49:11.942708 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/fac07a15-8332-4374-aaca-6577650e3183-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-x42ls\" (UID: \"fac07a15-8332-4374-aaca-6577650e3183\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-x42ls" Dec 11 08:49:11 crc kubenswrapper[4860]: I1211 08:49:11.942807 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/fac07a15-8332-4374-aaca-6577650e3183-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-x42ls\" (UID: \"fac07a15-8332-4374-aaca-6577650e3183\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-x42ls" Dec 11 08:49:11 crc kubenswrapper[4860]: I1211 08:49:11.942866 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/fac07a15-8332-4374-aaca-6577650e3183-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-x42ls\" (UID: \"fac07a15-8332-4374-aaca-6577650e3183\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-x42ls" Dec 11 08:49:11 crc kubenswrapper[4860]: I1211 08:49:11.942903 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/fac07a15-8332-4374-aaca-6577650e3183-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-x42ls\" (UID: \"fac07a15-8332-4374-aaca-6577650e3183\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-x42ls" Dec 11 08:49:11 crc kubenswrapper[4860]: I1211 08:49:11.942953 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-prbwd\" (UniqueName: \"kubernetes.io/projected/fac07a15-8332-4374-aaca-6577650e3183-kube-api-access-prbwd\") pod \"nova-edpm-deployment-openstack-edpm-ipam-x42ls\" (UID: \"fac07a15-8332-4374-aaca-6577650e3183\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-x42ls" Dec 11 08:49:11 crc kubenswrapper[4860]: I1211 08:49:11.942995 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fac07a15-8332-4374-aaca-6577650e3183-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-x42ls\" (UID: \"fac07a15-8332-4374-aaca-6577650e3183\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-x42ls" Dec 11 08:49:11 crc kubenswrapper[4860]: I1211 08:49:11.943014 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/fac07a15-8332-4374-aaca-6577650e3183-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-x42ls\" (UID: \"fac07a15-8332-4374-aaca-6577650e3183\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-x42ls" Dec 11 08:49:11 crc kubenswrapper[4860]: I1211 08:49:11.943062 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fac07a15-8332-4374-aaca-6577650e3183-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-x42ls\" (UID: \"fac07a15-8332-4374-aaca-6577650e3183\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-x42ls" Dec 11 08:49:11 crc kubenswrapper[4860]: I1211 08:49:11.943129 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fac07a15-8332-4374-aaca-6577650e3183-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-x42ls\" (UID: \"fac07a15-8332-4374-aaca-6577650e3183\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-x42ls" Dec 11 08:49:12 crc kubenswrapper[4860]: I1211 08:49:12.045130 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/fac07a15-8332-4374-aaca-6577650e3183-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-x42ls\" (UID: \"fac07a15-8332-4374-aaca-6577650e3183\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-x42ls" Dec 11 08:49:12 crc kubenswrapper[4860]: I1211 08:49:12.045212 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/fac07a15-8332-4374-aaca-6577650e3183-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-x42ls\" (UID: \"fac07a15-8332-4374-aaca-6577650e3183\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-x42ls" Dec 11 08:49:12 crc kubenswrapper[4860]: I1211 08:49:12.045238 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-prbwd\" (UniqueName: \"kubernetes.io/projected/fac07a15-8332-4374-aaca-6577650e3183-kube-api-access-prbwd\") pod \"nova-edpm-deployment-openstack-edpm-ipam-x42ls\" (UID: \"fac07a15-8332-4374-aaca-6577650e3183\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-x42ls" Dec 11 08:49:12 crc kubenswrapper[4860]: I1211 08:49:12.045257 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fac07a15-8332-4374-aaca-6577650e3183-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-x42ls\" (UID: \"fac07a15-8332-4374-aaca-6577650e3183\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-x42ls" Dec 11 08:49:12 crc kubenswrapper[4860]: I1211 08:49:12.045279 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/fac07a15-8332-4374-aaca-6577650e3183-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-x42ls\" (UID: \"fac07a15-8332-4374-aaca-6577650e3183\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-x42ls" Dec 11 08:49:12 crc kubenswrapper[4860]: I1211 08:49:12.045307 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fac07a15-8332-4374-aaca-6577650e3183-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-x42ls\" (UID: \"fac07a15-8332-4374-aaca-6577650e3183\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-x42ls" Dec 11 08:49:12 crc kubenswrapper[4860]: I1211 08:49:12.045354 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fac07a15-8332-4374-aaca-6577650e3183-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-x42ls\" (UID: \"fac07a15-8332-4374-aaca-6577650e3183\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-x42ls" Dec 11 08:49:12 crc kubenswrapper[4860]: I1211 08:49:12.045392 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/fac07a15-8332-4374-aaca-6577650e3183-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-x42ls\" (UID: \"fac07a15-8332-4374-aaca-6577650e3183\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-x42ls" Dec 11 08:49:12 crc kubenswrapper[4860]: I1211 08:49:12.045453 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/fac07a15-8332-4374-aaca-6577650e3183-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-x42ls\" (UID: \"fac07a15-8332-4374-aaca-6577650e3183\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-x42ls" Dec 11 08:49:12 crc kubenswrapper[4860]: I1211 08:49:12.046216 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/fac07a15-8332-4374-aaca-6577650e3183-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-x42ls\" (UID: \"fac07a15-8332-4374-aaca-6577650e3183\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-x42ls" Dec 11 08:49:12 crc kubenswrapper[4860]: I1211 08:49:12.051715 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/fac07a15-8332-4374-aaca-6577650e3183-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-x42ls\" (UID: \"fac07a15-8332-4374-aaca-6577650e3183\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-x42ls" Dec 11 08:49:12 crc kubenswrapper[4860]: I1211 08:49:12.052140 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fac07a15-8332-4374-aaca-6577650e3183-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-x42ls\" (UID: \"fac07a15-8332-4374-aaca-6577650e3183\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-x42ls" Dec 11 08:49:12 crc kubenswrapper[4860]: I1211 08:49:12.052218 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/fac07a15-8332-4374-aaca-6577650e3183-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-x42ls\" (UID: \"fac07a15-8332-4374-aaca-6577650e3183\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-x42ls" Dec 11 08:49:12 crc kubenswrapper[4860]: I1211 08:49:12.054446 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fac07a15-8332-4374-aaca-6577650e3183-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-x42ls\" (UID: \"fac07a15-8332-4374-aaca-6577650e3183\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-x42ls" Dec 11 08:49:12 crc kubenswrapper[4860]: I1211 08:49:12.057691 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/fac07a15-8332-4374-aaca-6577650e3183-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-x42ls\" (UID: \"fac07a15-8332-4374-aaca-6577650e3183\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-x42ls" Dec 11 08:49:12 crc kubenswrapper[4860]: I1211 08:49:12.058332 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/fac07a15-8332-4374-aaca-6577650e3183-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-x42ls\" (UID: \"fac07a15-8332-4374-aaca-6577650e3183\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-x42ls" Dec 11 08:49:12 crc kubenswrapper[4860]: I1211 08:49:12.063295 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fac07a15-8332-4374-aaca-6577650e3183-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-x42ls\" (UID: \"fac07a15-8332-4374-aaca-6577650e3183\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-x42ls" Dec 11 08:49:12 crc kubenswrapper[4860]: I1211 08:49:12.069872 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-prbwd\" (UniqueName: \"kubernetes.io/projected/fac07a15-8332-4374-aaca-6577650e3183-kube-api-access-prbwd\") pod \"nova-edpm-deployment-openstack-edpm-ipam-x42ls\" (UID: \"fac07a15-8332-4374-aaca-6577650e3183\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-x42ls" Dec 11 08:49:12 crc kubenswrapper[4860]: I1211 08:49:12.108486 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-x42ls" Dec 11 08:49:12 crc kubenswrapper[4860]: I1211 08:49:12.659424 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-x42ls"] Dec 11 08:49:12 crc kubenswrapper[4860]: I1211 08:49:12.702836 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-x42ls" event={"ID":"fac07a15-8332-4374-aaca-6577650e3183","Type":"ContainerStarted","Data":"4eb252880694d1d4cd93eb24e7ceaaa3f84c99cfac98223654caff4a1791273f"} Dec 11 08:49:13 crc kubenswrapper[4860]: I1211 08:49:13.712052 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-x42ls" event={"ID":"fac07a15-8332-4374-aaca-6577650e3183","Type":"ContainerStarted","Data":"4c28d14ad310e911d545d787e0bf15c099187767a791a8b97768464101c42ef5"} Dec 11 08:49:13 crc kubenswrapper[4860]: I1211 08:49:13.733562 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-x42ls" podStartSLOduration=2.271482882 podStartE2EDuration="2.733544708s" podCreationTimestamp="2025-12-11 08:49:11 +0000 UTC" firstStartedPulling="2025-12-11 08:49:12.663127091 +0000 UTC m=+2285.391646146" lastFinishedPulling="2025-12-11 08:49:13.125188907 +0000 UTC m=+2285.853707972" observedRunningTime="2025-12-11 08:49:13.727435439 +0000 UTC m=+2286.455954494" watchObservedRunningTime="2025-12-11 08:49:13.733544708 +0000 UTC m=+2286.462063763" Dec 11 08:50:08 crc kubenswrapper[4860]: I1211 08:50:08.795481 4860 patch_prober.go:28] interesting pod/machine-config-daemon-99qgp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 08:50:08 crc kubenswrapper[4860]: I1211 08:50:08.796052 4860 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 08:50:38 crc kubenswrapper[4860]: I1211 08:50:38.795715 4860 patch_prober.go:28] interesting pod/machine-config-daemon-99qgp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 08:50:38 crc kubenswrapper[4860]: I1211 08:50:38.796231 4860 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 08:51:08 crc kubenswrapper[4860]: I1211 08:51:08.795515 4860 patch_prober.go:28] interesting pod/machine-config-daemon-99qgp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 08:51:09 crc kubenswrapper[4860]: I1211 08:51:08.796463 4860 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 08:51:09 crc kubenswrapper[4860]: I1211 08:51:08.796554 4860 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" Dec 11 08:51:09 crc kubenswrapper[4860]: I1211 08:51:08.797936 4860 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b747bbead8584985306f457adcbaf89164c99c29fe92a3c021617afef9e04489"} pod="openshift-machine-config-operator/machine-config-daemon-99qgp" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 11 08:51:09 crc kubenswrapper[4860]: I1211 08:51:08.798076 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" containerName="machine-config-daemon" containerID="cri-o://b747bbead8584985306f457adcbaf89164c99c29fe92a3c021617afef9e04489" gracePeriod=600 Dec 11 08:51:09 crc kubenswrapper[4860]: E1211 08:51:09.440077 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-99qgp_openshift-machine-config-operator(31c30642-6e60-41b4-a477-0d802424e0aa)\"" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" Dec 11 08:51:09 crc kubenswrapper[4860]: I1211 08:51:09.726631 4860 generic.go:334] "Generic (PLEG): container finished" podID="31c30642-6e60-41b4-a477-0d802424e0aa" containerID="b747bbead8584985306f457adcbaf89164c99c29fe92a3c021617afef9e04489" exitCode=0 Dec 11 08:51:09 crc kubenswrapper[4860]: I1211 08:51:09.726677 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" event={"ID":"31c30642-6e60-41b4-a477-0d802424e0aa","Type":"ContainerDied","Data":"b747bbead8584985306f457adcbaf89164c99c29fe92a3c021617afef9e04489"} Dec 11 08:51:09 crc kubenswrapper[4860]: I1211 08:51:09.726741 4860 scope.go:117] "RemoveContainer" containerID="360fac818f30e5d5bec8360d7df78a5f152a69c33b8db5723f4b538448a80e00" Dec 11 08:51:09 crc kubenswrapper[4860]: I1211 08:51:09.727475 4860 scope.go:117] "RemoveContainer" containerID="b747bbead8584985306f457adcbaf89164c99c29fe92a3c021617afef9e04489" Dec 11 08:51:09 crc kubenswrapper[4860]: E1211 08:51:09.727919 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-99qgp_openshift-machine-config-operator(31c30642-6e60-41b4-a477-0d802424e0aa)\"" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" Dec 11 08:51:22 crc kubenswrapper[4860]: I1211 08:51:22.578614 4860 scope.go:117] "RemoveContainer" containerID="b747bbead8584985306f457adcbaf89164c99c29fe92a3c021617afef9e04489" Dec 11 08:51:22 crc kubenswrapper[4860]: E1211 08:51:22.579547 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-99qgp_openshift-machine-config-operator(31c30642-6e60-41b4-a477-0d802424e0aa)\"" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" Dec 11 08:51:33 crc kubenswrapper[4860]: I1211 08:51:33.579251 4860 scope.go:117] "RemoveContainer" containerID="b747bbead8584985306f457adcbaf89164c99c29fe92a3c021617afef9e04489" Dec 11 08:51:33 crc kubenswrapper[4860]: E1211 08:51:33.580256 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-99qgp_openshift-machine-config-operator(31c30642-6e60-41b4-a477-0d802424e0aa)\"" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" Dec 11 08:51:46 crc kubenswrapper[4860]: I1211 08:51:46.044116 4860 generic.go:334] "Generic (PLEG): container finished" podID="fac07a15-8332-4374-aaca-6577650e3183" containerID="4c28d14ad310e911d545d787e0bf15c099187767a791a8b97768464101c42ef5" exitCode=0 Dec 11 08:51:46 crc kubenswrapper[4860]: I1211 08:51:46.044214 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-x42ls" event={"ID":"fac07a15-8332-4374-aaca-6577650e3183","Type":"ContainerDied","Data":"4c28d14ad310e911d545d787e0bf15c099187767a791a8b97768464101c42ef5"} Dec 11 08:51:47 crc kubenswrapper[4860]: I1211 08:51:47.459635 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-x42ls" Dec 11 08:51:47 crc kubenswrapper[4860]: I1211 08:51:47.552723 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fac07a15-8332-4374-aaca-6577650e3183-nova-combined-ca-bundle\") pod \"fac07a15-8332-4374-aaca-6577650e3183\" (UID: \"fac07a15-8332-4374-aaca-6577650e3183\") " Dec 11 08:51:47 crc kubenswrapper[4860]: I1211 08:51:47.553040 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-prbwd\" (UniqueName: \"kubernetes.io/projected/fac07a15-8332-4374-aaca-6577650e3183-kube-api-access-prbwd\") pod \"fac07a15-8332-4374-aaca-6577650e3183\" (UID: \"fac07a15-8332-4374-aaca-6577650e3183\") " Dec 11 08:51:47 crc kubenswrapper[4860]: I1211 08:51:47.553081 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/fac07a15-8332-4374-aaca-6577650e3183-nova-migration-ssh-key-1\") pod \"fac07a15-8332-4374-aaca-6577650e3183\" (UID: \"fac07a15-8332-4374-aaca-6577650e3183\") " Dec 11 08:51:47 crc kubenswrapper[4860]: I1211 08:51:47.553119 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/fac07a15-8332-4374-aaca-6577650e3183-nova-cell1-compute-config-1\") pod \"fac07a15-8332-4374-aaca-6577650e3183\" (UID: \"fac07a15-8332-4374-aaca-6577650e3183\") " Dec 11 08:51:47 crc kubenswrapper[4860]: I1211 08:51:47.553152 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fac07a15-8332-4374-aaca-6577650e3183-inventory\") pod \"fac07a15-8332-4374-aaca-6577650e3183\" (UID: \"fac07a15-8332-4374-aaca-6577650e3183\") " Dec 11 08:51:47 crc kubenswrapper[4860]: I1211 08:51:47.553182 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/fac07a15-8332-4374-aaca-6577650e3183-nova-extra-config-0\") pod \"fac07a15-8332-4374-aaca-6577650e3183\" (UID: \"fac07a15-8332-4374-aaca-6577650e3183\") " Dec 11 08:51:47 crc kubenswrapper[4860]: I1211 08:51:47.553294 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fac07a15-8332-4374-aaca-6577650e3183-ssh-key\") pod \"fac07a15-8332-4374-aaca-6577650e3183\" (UID: \"fac07a15-8332-4374-aaca-6577650e3183\") " Dec 11 08:51:47 crc kubenswrapper[4860]: I1211 08:51:47.553323 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/fac07a15-8332-4374-aaca-6577650e3183-nova-migration-ssh-key-0\") pod \"fac07a15-8332-4374-aaca-6577650e3183\" (UID: \"fac07a15-8332-4374-aaca-6577650e3183\") " Dec 11 08:51:47 crc kubenswrapper[4860]: I1211 08:51:47.553347 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/fac07a15-8332-4374-aaca-6577650e3183-nova-cell1-compute-config-0\") pod \"fac07a15-8332-4374-aaca-6577650e3183\" (UID: \"fac07a15-8332-4374-aaca-6577650e3183\") " Dec 11 08:51:47 crc kubenswrapper[4860]: I1211 08:51:47.559807 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fac07a15-8332-4374-aaca-6577650e3183-kube-api-access-prbwd" (OuterVolumeSpecName: "kube-api-access-prbwd") pod "fac07a15-8332-4374-aaca-6577650e3183" (UID: "fac07a15-8332-4374-aaca-6577650e3183"). InnerVolumeSpecName "kube-api-access-prbwd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:51:47 crc kubenswrapper[4860]: I1211 08:51:47.565927 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fac07a15-8332-4374-aaca-6577650e3183-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "fac07a15-8332-4374-aaca-6577650e3183" (UID: "fac07a15-8332-4374-aaca-6577650e3183"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:51:47 crc kubenswrapper[4860]: I1211 08:51:47.587762 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fac07a15-8332-4374-aaca-6577650e3183-inventory" (OuterVolumeSpecName: "inventory") pod "fac07a15-8332-4374-aaca-6577650e3183" (UID: "fac07a15-8332-4374-aaca-6577650e3183"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:51:47 crc kubenswrapper[4860]: I1211 08:51:47.588416 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fac07a15-8332-4374-aaca-6577650e3183-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "fac07a15-8332-4374-aaca-6577650e3183" (UID: "fac07a15-8332-4374-aaca-6577650e3183"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:51:47 crc kubenswrapper[4860]: I1211 08:51:47.589362 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fac07a15-8332-4374-aaca-6577650e3183-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "fac07a15-8332-4374-aaca-6577650e3183" (UID: "fac07a15-8332-4374-aaca-6577650e3183"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 08:51:47 crc kubenswrapper[4860]: I1211 08:51:47.595881 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fac07a15-8332-4374-aaca-6577650e3183-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "fac07a15-8332-4374-aaca-6577650e3183" (UID: "fac07a15-8332-4374-aaca-6577650e3183"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:51:47 crc kubenswrapper[4860]: I1211 08:51:47.597025 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fac07a15-8332-4374-aaca-6577650e3183-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "fac07a15-8332-4374-aaca-6577650e3183" (UID: "fac07a15-8332-4374-aaca-6577650e3183"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:51:47 crc kubenswrapper[4860]: I1211 08:51:47.603823 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fac07a15-8332-4374-aaca-6577650e3183-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "fac07a15-8332-4374-aaca-6577650e3183" (UID: "fac07a15-8332-4374-aaca-6577650e3183"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:51:47 crc kubenswrapper[4860]: I1211 08:51:47.609687 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fac07a15-8332-4374-aaca-6577650e3183-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "fac07a15-8332-4374-aaca-6577650e3183" (UID: "fac07a15-8332-4374-aaca-6577650e3183"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:51:47 crc kubenswrapper[4860]: I1211 08:51:47.655471 4860 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fac07a15-8332-4374-aaca-6577650e3183-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 08:51:47 crc kubenswrapper[4860]: I1211 08:51:47.655522 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-prbwd\" (UniqueName: \"kubernetes.io/projected/fac07a15-8332-4374-aaca-6577650e3183-kube-api-access-prbwd\") on node \"crc\" DevicePath \"\"" Dec 11 08:51:47 crc kubenswrapper[4860]: I1211 08:51:47.655535 4860 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/fac07a15-8332-4374-aaca-6577650e3183-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Dec 11 08:51:47 crc kubenswrapper[4860]: I1211 08:51:47.655546 4860 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/fac07a15-8332-4374-aaca-6577650e3183-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Dec 11 08:51:47 crc kubenswrapper[4860]: I1211 08:51:47.655558 4860 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fac07a15-8332-4374-aaca-6577650e3183-inventory\") on node \"crc\" DevicePath \"\"" Dec 11 08:51:47 crc kubenswrapper[4860]: I1211 08:51:47.655570 4860 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/fac07a15-8332-4374-aaca-6577650e3183-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Dec 11 08:51:47 crc kubenswrapper[4860]: I1211 08:51:47.655584 4860 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fac07a15-8332-4374-aaca-6577650e3183-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 11 08:51:47 crc kubenswrapper[4860]: I1211 08:51:47.655595 4860 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/fac07a15-8332-4374-aaca-6577650e3183-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Dec 11 08:51:47 crc kubenswrapper[4860]: I1211 08:51:47.655606 4860 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/fac07a15-8332-4374-aaca-6577650e3183-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Dec 11 08:51:48 crc kubenswrapper[4860]: I1211 08:51:48.062101 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-x42ls" event={"ID":"fac07a15-8332-4374-aaca-6577650e3183","Type":"ContainerDied","Data":"4eb252880694d1d4cd93eb24e7ceaaa3f84c99cfac98223654caff4a1791273f"} Dec 11 08:51:48 crc kubenswrapper[4860]: I1211 08:51:48.062133 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-x42ls" Dec 11 08:51:48 crc kubenswrapper[4860]: I1211 08:51:48.062143 4860 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4eb252880694d1d4cd93eb24e7ceaaa3f84c99cfac98223654caff4a1791273f" Dec 11 08:51:48 crc kubenswrapper[4860]: I1211 08:51:48.172034 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wmwhw"] Dec 11 08:51:48 crc kubenswrapper[4860]: E1211 08:51:48.172562 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fac07a15-8332-4374-aaca-6577650e3183" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 11 08:51:48 crc kubenswrapper[4860]: I1211 08:51:48.172585 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="fac07a15-8332-4374-aaca-6577650e3183" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 11 08:51:48 crc kubenswrapper[4860]: I1211 08:51:48.172852 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="fac07a15-8332-4374-aaca-6577650e3183" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 11 08:51:48 crc kubenswrapper[4860]: I1211 08:51:48.173699 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wmwhw" Dec 11 08:51:48 crc kubenswrapper[4860]: I1211 08:51:48.178020 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Dec 11 08:51:48 crc kubenswrapper[4860]: I1211 08:51:48.178259 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 11 08:51:48 crc kubenswrapper[4860]: I1211 08:51:48.178547 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 11 08:51:48 crc kubenswrapper[4860]: I1211 08:51:48.178789 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 11 08:51:48 crc kubenswrapper[4860]: I1211 08:51:48.178965 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-7w7kh" Dec 11 08:51:48 crc kubenswrapper[4860]: I1211 08:51:48.187717 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wmwhw"] Dec 11 08:51:48 crc kubenswrapper[4860]: I1211 08:51:48.265698 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/beb319f7-9b48-413d-9561-e6c0dc62238f-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wmwhw\" (UID: \"beb319f7-9b48-413d-9561-e6c0dc62238f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wmwhw" Dec 11 08:51:48 crc kubenswrapper[4860]: I1211 08:51:48.265767 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/beb319f7-9b48-413d-9561-e6c0dc62238f-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wmwhw\" (UID: \"beb319f7-9b48-413d-9561-e6c0dc62238f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wmwhw" Dec 11 08:51:48 crc kubenswrapper[4860]: I1211 08:51:48.265838 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/beb319f7-9b48-413d-9561-e6c0dc62238f-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wmwhw\" (UID: \"beb319f7-9b48-413d-9561-e6c0dc62238f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wmwhw" Dec 11 08:51:48 crc kubenswrapper[4860]: I1211 08:51:48.265867 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-chvbx\" (UniqueName: \"kubernetes.io/projected/beb319f7-9b48-413d-9561-e6c0dc62238f-kube-api-access-chvbx\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wmwhw\" (UID: \"beb319f7-9b48-413d-9561-e6c0dc62238f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wmwhw" Dec 11 08:51:48 crc kubenswrapper[4860]: I1211 08:51:48.265935 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/beb319f7-9b48-413d-9561-e6c0dc62238f-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wmwhw\" (UID: \"beb319f7-9b48-413d-9561-e6c0dc62238f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wmwhw" Dec 11 08:51:48 crc kubenswrapper[4860]: I1211 08:51:48.265975 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/beb319f7-9b48-413d-9561-e6c0dc62238f-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wmwhw\" (UID: \"beb319f7-9b48-413d-9561-e6c0dc62238f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wmwhw" Dec 11 08:51:48 crc kubenswrapper[4860]: I1211 08:51:48.265997 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/beb319f7-9b48-413d-9561-e6c0dc62238f-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wmwhw\" (UID: \"beb319f7-9b48-413d-9561-e6c0dc62238f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wmwhw" Dec 11 08:51:48 crc kubenswrapper[4860]: I1211 08:51:48.368368 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/beb319f7-9b48-413d-9561-e6c0dc62238f-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wmwhw\" (UID: \"beb319f7-9b48-413d-9561-e6c0dc62238f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wmwhw" Dec 11 08:51:48 crc kubenswrapper[4860]: I1211 08:51:48.368436 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-chvbx\" (UniqueName: \"kubernetes.io/projected/beb319f7-9b48-413d-9561-e6c0dc62238f-kube-api-access-chvbx\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wmwhw\" (UID: \"beb319f7-9b48-413d-9561-e6c0dc62238f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wmwhw" Dec 11 08:51:48 crc kubenswrapper[4860]: I1211 08:51:48.368520 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/beb319f7-9b48-413d-9561-e6c0dc62238f-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wmwhw\" (UID: \"beb319f7-9b48-413d-9561-e6c0dc62238f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wmwhw" Dec 11 08:51:48 crc kubenswrapper[4860]: I1211 08:51:48.368559 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/beb319f7-9b48-413d-9561-e6c0dc62238f-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wmwhw\" (UID: \"beb319f7-9b48-413d-9561-e6c0dc62238f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wmwhw" Dec 11 08:51:48 crc kubenswrapper[4860]: I1211 08:51:48.368586 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/beb319f7-9b48-413d-9561-e6c0dc62238f-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wmwhw\" (UID: \"beb319f7-9b48-413d-9561-e6c0dc62238f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wmwhw" Dec 11 08:51:48 crc kubenswrapper[4860]: I1211 08:51:48.368690 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/beb319f7-9b48-413d-9561-e6c0dc62238f-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wmwhw\" (UID: \"beb319f7-9b48-413d-9561-e6c0dc62238f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wmwhw" Dec 11 08:51:48 crc kubenswrapper[4860]: I1211 08:51:48.368735 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/beb319f7-9b48-413d-9561-e6c0dc62238f-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wmwhw\" (UID: \"beb319f7-9b48-413d-9561-e6c0dc62238f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wmwhw" Dec 11 08:51:48 crc kubenswrapper[4860]: I1211 08:51:48.372528 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/beb319f7-9b48-413d-9561-e6c0dc62238f-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wmwhw\" (UID: \"beb319f7-9b48-413d-9561-e6c0dc62238f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wmwhw" Dec 11 08:51:48 crc kubenswrapper[4860]: I1211 08:51:48.372677 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/beb319f7-9b48-413d-9561-e6c0dc62238f-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wmwhw\" (UID: \"beb319f7-9b48-413d-9561-e6c0dc62238f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wmwhw" Dec 11 08:51:48 crc kubenswrapper[4860]: I1211 08:51:48.373011 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/beb319f7-9b48-413d-9561-e6c0dc62238f-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wmwhw\" (UID: \"beb319f7-9b48-413d-9561-e6c0dc62238f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wmwhw" Dec 11 08:51:48 crc kubenswrapper[4860]: I1211 08:51:48.374905 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/beb319f7-9b48-413d-9561-e6c0dc62238f-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wmwhw\" (UID: \"beb319f7-9b48-413d-9561-e6c0dc62238f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wmwhw" Dec 11 08:51:48 crc kubenswrapper[4860]: I1211 08:51:48.377028 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/beb319f7-9b48-413d-9561-e6c0dc62238f-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wmwhw\" (UID: \"beb319f7-9b48-413d-9561-e6c0dc62238f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wmwhw" Dec 11 08:51:48 crc kubenswrapper[4860]: I1211 08:51:48.383149 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/beb319f7-9b48-413d-9561-e6c0dc62238f-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wmwhw\" (UID: \"beb319f7-9b48-413d-9561-e6c0dc62238f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wmwhw" Dec 11 08:51:48 crc kubenswrapper[4860]: I1211 08:51:48.391202 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-chvbx\" (UniqueName: \"kubernetes.io/projected/beb319f7-9b48-413d-9561-e6c0dc62238f-kube-api-access-chvbx\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wmwhw\" (UID: \"beb319f7-9b48-413d-9561-e6c0dc62238f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wmwhw" Dec 11 08:51:48 crc kubenswrapper[4860]: I1211 08:51:48.491340 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wmwhw" Dec 11 08:51:48 crc kubenswrapper[4860]: I1211 08:51:48.579190 4860 scope.go:117] "RemoveContainer" containerID="b747bbead8584985306f457adcbaf89164c99c29fe92a3c021617afef9e04489" Dec 11 08:51:48 crc kubenswrapper[4860]: E1211 08:51:48.579721 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-99qgp_openshift-machine-config-operator(31c30642-6e60-41b4-a477-0d802424e0aa)\"" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" Dec 11 08:51:49 crc kubenswrapper[4860]: I1211 08:51:49.003545 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wmwhw"] Dec 11 08:51:49 crc kubenswrapper[4860]: I1211 08:51:49.024570 4860 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 11 08:51:49 crc kubenswrapper[4860]: I1211 08:51:49.070546 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wmwhw" event={"ID":"beb319f7-9b48-413d-9561-e6c0dc62238f","Type":"ContainerStarted","Data":"9122eb55022eb140a74b6e783cfa164a2bf9ce1c585685cce65a0149893c77e7"} Dec 11 08:51:50 crc kubenswrapper[4860]: I1211 08:51:50.080335 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wmwhw" event={"ID":"beb319f7-9b48-413d-9561-e6c0dc62238f","Type":"ContainerStarted","Data":"6196c26e3f071ed113fd2033776c67cdefc31691cb70334a2e2c54bb233fd779"} Dec 11 08:51:50 crc kubenswrapper[4860]: I1211 08:51:50.099695 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wmwhw" podStartSLOduration=1.596655267 podStartE2EDuration="2.099675053s" podCreationTimestamp="2025-12-11 08:51:48 +0000 UTC" firstStartedPulling="2025-12-11 08:51:49.024321722 +0000 UTC m=+2441.752840777" lastFinishedPulling="2025-12-11 08:51:49.527341508 +0000 UTC m=+2442.255860563" observedRunningTime="2025-12-11 08:51:50.097691481 +0000 UTC m=+2442.826210546" watchObservedRunningTime="2025-12-11 08:51:50.099675053 +0000 UTC m=+2442.828194118" Dec 11 08:51:59 crc kubenswrapper[4860]: I1211 08:51:59.579130 4860 scope.go:117] "RemoveContainer" containerID="b747bbead8584985306f457adcbaf89164c99c29fe92a3c021617afef9e04489" Dec 11 08:51:59 crc kubenswrapper[4860]: E1211 08:51:59.579949 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-99qgp_openshift-machine-config-operator(31c30642-6e60-41b4-a477-0d802424e0aa)\"" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" Dec 11 08:52:14 crc kubenswrapper[4860]: I1211 08:52:14.579908 4860 scope.go:117] "RemoveContainer" containerID="b747bbead8584985306f457adcbaf89164c99c29fe92a3c021617afef9e04489" Dec 11 08:52:14 crc kubenswrapper[4860]: E1211 08:52:14.580937 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-99qgp_openshift-machine-config-operator(31c30642-6e60-41b4-a477-0d802424e0aa)\"" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" Dec 11 08:52:26 crc kubenswrapper[4860]: I1211 08:52:26.579459 4860 scope.go:117] "RemoveContainer" containerID="b747bbead8584985306f457adcbaf89164c99c29fe92a3c021617afef9e04489" Dec 11 08:52:26 crc kubenswrapper[4860]: E1211 08:52:26.581473 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-99qgp_openshift-machine-config-operator(31c30642-6e60-41b4-a477-0d802424e0aa)\"" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" Dec 11 08:52:38 crc kubenswrapper[4860]: I1211 08:52:38.579231 4860 scope.go:117] "RemoveContainer" containerID="b747bbead8584985306f457adcbaf89164c99c29fe92a3c021617afef9e04489" Dec 11 08:52:38 crc kubenswrapper[4860]: E1211 08:52:38.580095 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-99qgp_openshift-machine-config-operator(31c30642-6e60-41b4-a477-0d802424e0aa)\"" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" Dec 11 08:52:51 crc kubenswrapper[4860]: I1211 08:52:51.578381 4860 scope.go:117] "RemoveContainer" containerID="b747bbead8584985306f457adcbaf89164c99c29fe92a3c021617afef9e04489" Dec 11 08:52:51 crc kubenswrapper[4860]: E1211 08:52:51.579113 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-99qgp_openshift-machine-config-operator(31c30642-6e60-41b4-a477-0d802424e0aa)\"" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" Dec 11 08:53:04 crc kubenswrapper[4860]: I1211 08:53:04.579382 4860 scope.go:117] "RemoveContainer" containerID="b747bbead8584985306f457adcbaf89164c99c29fe92a3c021617afef9e04489" Dec 11 08:53:04 crc kubenswrapper[4860]: E1211 08:53:04.580159 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-99qgp_openshift-machine-config-operator(31c30642-6e60-41b4-a477-0d802424e0aa)\"" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" Dec 11 08:53:16 crc kubenswrapper[4860]: I1211 08:53:16.580036 4860 scope.go:117] "RemoveContainer" containerID="b747bbead8584985306f457adcbaf89164c99c29fe92a3c021617afef9e04489" Dec 11 08:53:16 crc kubenswrapper[4860]: E1211 08:53:16.581173 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-99qgp_openshift-machine-config-operator(31c30642-6e60-41b4-a477-0d802424e0aa)\"" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" Dec 11 08:53:31 crc kubenswrapper[4860]: I1211 08:53:31.578891 4860 scope.go:117] "RemoveContainer" containerID="b747bbead8584985306f457adcbaf89164c99c29fe92a3c021617afef9e04489" Dec 11 08:53:31 crc kubenswrapper[4860]: E1211 08:53:31.581106 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-99qgp_openshift-machine-config-operator(31c30642-6e60-41b4-a477-0d802424e0aa)\"" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" Dec 11 08:53:45 crc kubenswrapper[4860]: I1211 08:53:45.579710 4860 scope.go:117] "RemoveContainer" containerID="b747bbead8584985306f457adcbaf89164c99c29fe92a3c021617afef9e04489" Dec 11 08:53:45 crc kubenswrapper[4860]: E1211 08:53:45.580613 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-99qgp_openshift-machine-config-operator(31c30642-6e60-41b4-a477-0d802424e0aa)\"" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" Dec 11 08:53:59 crc kubenswrapper[4860]: I1211 08:53:59.579849 4860 scope.go:117] "RemoveContainer" containerID="b747bbead8584985306f457adcbaf89164c99c29fe92a3c021617afef9e04489" Dec 11 08:53:59 crc kubenswrapper[4860]: E1211 08:53:59.580572 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-99qgp_openshift-machine-config-operator(31c30642-6e60-41b4-a477-0d802424e0aa)\"" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" Dec 11 08:54:03 crc kubenswrapper[4860]: I1211 08:54:03.313317 4860 generic.go:334] "Generic (PLEG): container finished" podID="beb319f7-9b48-413d-9561-e6c0dc62238f" containerID="6196c26e3f071ed113fd2033776c67cdefc31691cb70334a2e2c54bb233fd779" exitCode=0 Dec 11 08:54:03 crc kubenswrapper[4860]: I1211 08:54:03.313420 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wmwhw" event={"ID":"beb319f7-9b48-413d-9561-e6c0dc62238f","Type":"ContainerDied","Data":"6196c26e3f071ed113fd2033776c67cdefc31691cb70334a2e2c54bb233fd779"} Dec 11 08:54:04 crc kubenswrapper[4860]: I1211 08:54:04.726506 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wmwhw" Dec 11 08:54:04 crc kubenswrapper[4860]: I1211 08:54:04.817522 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/beb319f7-9b48-413d-9561-e6c0dc62238f-ceilometer-compute-config-data-1\") pod \"beb319f7-9b48-413d-9561-e6c0dc62238f\" (UID: \"beb319f7-9b48-413d-9561-e6c0dc62238f\") " Dec 11 08:54:04 crc kubenswrapper[4860]: I1211 08:54:04.817584 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/beb319f7-9b48-413d-9561-e6c0dc62238f-telemetry-combined-ca-bundle\") pod \"beb319f7-9b48-413d-9561-e6c0dc62238f\" (UID: \"beb319f7-9b48-413d-9561-e6c0dc62238f\") " Dec 11 08:54:04 crc kubenswrapper[4860]: I1211 08:54:04.817644 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-chvbx\" (UniqueName: \"kubernetes.io/projected/beb319f7-9b48-413d-9561-e6c0dc62238f-kube-api-access-chvbx\") pod \"beb319f7-9b48-413d-9561-e6c0dc62238f\" (UID: \"beb319f7-9b48-413d-9561-e6c0dc62238f\") " Dec 11 08:54:04 crc kubenswrapper[4860]: I1211 08:54:04.817685 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/beb319f7-9b48-413d-9561-e6c0dc62238f-ceilometer-compute-config-data-2\") pod \"beb319f7-9b48-413d-9561-e6c0dc62238f\" (UID: \"beb319f7-9b48-413d-9561-e6c0dc62238f\") " Dec 11 08:54:04 crc kubenswrapper[4860]: I1211 08:54:04.817797 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/beb319f7-9b48-413d-9561-e6c0dc62238f-ssh-key\") pod \"beb319f7-9b48-413d-9561-e6c0dc62238f\" (UID: \"beb319f7-9b48-413d-9561-e6c0dc62238f\") " Dec 11 08:54:04 crc kubenswrapper[4860]: I1211 08:54:04.817898 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/beb319f7-9b48-413d-9561-e6c0dc62238f-ceilometer-compute-config-data-0\") pod \"beb319f7-9b48-413d-9561-e6c0dc62238f\" (UID: \"beb319f7-9b48-413d-9561-e6c0dc62238f\") " Dec 11 08:54:04 crc kubenswrapper[4860]: I1211 08:54:04.817945 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/beb319f7-9b48-413d-9561-e6c0dc62238f-inventory\") pod \"beb319f7-9b48-413d-9561-e6c0dc62238f\" (UID: \"beb319f7-9b48-413d-9561-e6c0dc62238f\") " Dec 11 08:54:04 crc kubenswrapper[4860]: I1211 08:54:04.823550 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/beb319f7-9b48-413d-9561-e6c0dc62238f-kube-api-access-chvbx" (OuterVolumeSpecName: "kube-api-access-chvbx") pod "beb319f7-9b48-413d-9561-e6c0dc62238f" (UID: "beb319f7-9b48-413d-9561-e6c0dc62238f"). InnerVolumeSpecName "kube-api-access-chvbx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:54:04 crc kubenswrapper[4860]: I1211 08:54:04.824892 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/beb319f7-9b48-413d-9561-e6c0dc62238f-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "beb319f7-9b48-413d-9561-e6c0dc62238f" (UID: "beb319f7-9b48-413d-9561-e6c0dc62238f"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:54:04 crc kubenswrapper[4860]: I1211 08:54:04.845793 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/beb319f7-9b48-413d-9561-e6c0dc62238f-inventory" (OuterVolumeSpecName: "inventory") pod "beb319f7-9b48-413d-9561-e6c0dc62238f" (UID: "beb319f7-9b48-413d-9561-e6c0dc62238f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:54:04 crc kubenswrapper[4860]: I1211 08:54:04.849472 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/beb319f7-9b48-413d-9561-e6c0dc62238f-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "beb319f7-9b48-413d-9561-e6c0dc62238f" (UID: "beb319f7-9b48-413d-9561-e6c0dc62238f"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:54:04 crc kubenswrapper[4860]: I1211 08:54:04.850004 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/beb319f7-9b48-413d-9561-e6c0dc62238f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "beb319f7-9b48-413d-9561-e6c0dc62238f" (UID: "beb319f7-9b48-413d-9561-e6c0dc62238f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:54:04 crc kubenswrapper[4860]: I1211 08:54:04.863809 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/beb319f7-9b48-413d-9561-e6c0dc62238f-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "beb319f7-9b48-413d-9561-e6c0dc62238f" (UID: "beb319f7-9b48-413d-9561-e6c0dc62238f"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:54:04 crc kubenswrapper[4860]: I1211 08:54:04.866343 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/beb319f7-9b48-413d-9561-e6c0dc62238f-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "beb319f7-9b48-413d-9561-e6c0dc62238f" (UID: "beb319f7-9b48-413d-9561-e6c0dc62238f"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 08:54:04 crc kubenswrapper[4860]: I1211 08:54:04.920379 4860 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/beb319f7-9b48-413d-9561-e6c0dc62238f-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Dec 11 08:54:04 crc kubenswrapper[4860]: I1211 08:54:04.920648 4860 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/beb319f7-9b48-413d-9561-e6c0dc62238f-inventory\") on node \"crc\" DevicePath \"\"" Dec 11 08:54:04 crc kubenswrapper[4860]: I1211 08:54:04.920735 4860 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/beb319f7-9b48-413d-9561-e6c0dc62238f-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Dec 11 08:54:04 crc kubenswrapper[4860]: I1211 08:54:04.920793 4860 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/beb319f7-9b48-413d-9561-e6c0dc62238f-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 08:54:04 crc kubenswrapper[4860]: I1211 08:54:04.920851 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-chvbx\" (UniqueName: \"kubernetes.io/projected/beb319f7-9b48-413d-9561-e6c0dc62238f-kube-api-access-chvbx\") on node \"crc\" DevicePath \"\"" Dec 11 08:54:04 crc kubenswrapper[4860]: I1211 08:54:04.920904 4860 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/beb319f7-9b48-413d-9561-e6c0dc62238f-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Dec 11 08:54:04 crc kubenswrapper[4860]: I1211 08:54:04.920960 4860 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/beb319f7-9b48-413d-9561-e6c0dc62238f-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 11 08:54:05 crc kubenswrapper[4860]: I1211 08:54:05.331291 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wmwhw" event={"ID":"beb319f7-9b48-413d-9561-e6c0dc62238f","Type":"ContainerDied","Data":"9122eb55022eb140a74b6e783cfa164a2bf9ce1c585685cce65a0149893c77e7"} Dec 11 08:54:05 crc kubenswrapper[4860]: I1211 08:54:05.331331 4860 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9122eb55022eb140a74b6e783cfa164a2bf9ce1c585685cce65a0149893c77e7" Dec 11 08:54:05 crc kubenswrapper[4860]: I1211 08:54:05.331359 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wmwhw" Dec 11 08:54:14 crc kubenswrapper[4860]: I1211 08:54:14.156449 4860 scope.go:117] "RemoveContainer" containerID="b747bbead8584985306f457adcbaf89164c99c29fe92a3c021617afef9e04489" Dec 11 08:54:14 crc kubenswrapper[4860]: E1211 08:54:14.157428 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-99qgp_openshift-machine-config-operator(31c30642-6e60-41b4-a477-0d802424e0aa)\"" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" Dec 11 08:54:25 crc kubenswrapper[4860]: I1211 08:54:25.579494 4860 scope.go:117] "RemoveContainer" containerID="b747bbead8584985306f457adcbaf89164c99c29fe92a3c021617afef9e04489" Dec 11 08:54:25 crc kubenswrapper[4860]: E1211 08:54:25.580414 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-99qgp_openshift-machine-config-operator(31c30642-6e60-41b4-a477-0d802424e0aa)\"" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" Dec 11 08:54:39 crc kubenswrapper[4860]: I1211 08:54:39.579526 4860 scope.go:117] "RemoveContainer" containerID="b747bbead8584985306f457adcbaf89164c99c29fe92a3c021617afef9e04489" Dec 11 08:54:39 crc kubenswrapper[4860]: E1211 08:54:39.580502 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-99qgp_openshift-machine-config-operator(31c30642-6e60-41b4-a477-0d802424e0aa)\"" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" Dec 11 08:54:50 crc kubenswrapper[4860]: I1211 08:54:50.093180 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Dec 11 08:54:50 crc kubenswrapper[4860]: E1211 08:54:50.095035 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="beb319f7-9b48-413d-9561-e6c0dc62238f" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 11 08:54:50 crc kubenswrapper[4860]: I1211 08:54:50.095056 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="beb319f7-9b48-413d-9561-e6c0dc62238f" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 11 08:54:50 crc kubenswrapper[4860]: I1211 08:54:50.095214 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="beb319f7-9b48-413d-9561-e6c0dc62238f" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 11 08:54:50 crc kubenswrapper[4860]: I1211 08:54:50.095950 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 11 08:54:50 crc kubenswrapper[4860]: I1211 08:54:50.098109 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-hlscw" Dec 11 08:54:50 crc kubenswrapper[4860]: I1211 08:54:50.098496 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Dec 11 08:54:50 crc kubenswrapper[4860]: I1211 08:54:50.098993 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Dec 11 08:54:50 crc kubenswrapper[4860]: I1211 08:54:50.099008 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Dec 11 08:54:50 crc kubenswrapper[4860]: I1211 08:54:50.104496 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Dec 11 08:54:50 crc kubenswrapper[4860]: I1211 08:54:50.233247 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/3b659d05-76c9-4b04-a05e-a7dd488e9852-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"3b659d05-76c9-4b04-a05e-a7dd488e9852\") " pod="openstack/tempest-tests-tempest" Dec 11 08:54:50 crc kubenswrapper[4860]: I1211 08:54:50.233298 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"tempest-tests-tempest\" (UID: \"3b659d05-76c9-4b04-a05e-a7dd488e9852\") " pod="openstack/tempest-tests-tempest" Dec 11 08:54:50 crc kubenswrapper[4860]: I1211 08:54:50.233328 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/3b659d05-76c9-4b04-a05e-a7dd488e9852-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"3b659d05-76c9-4b04-a05e-a7dd488e9852\") " pod="openstack/tempest-tests-tempest" Dec 11 08:54:50 crc kubenswrapper[4860]: I1211 08:54:50.233551 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q4rtg\" (UniqueName: \"kubernetes.io/projected/3b659d05-76c9-4b04-a05e-a7dd488e9852-kube-api-access-q4rtg\") pod \"tempest-tests-tempest\" (UID: \"3b659d05-76c9-4b04-a05e-a7dd488e9852\") " pod="openstack/tempest-tests-tempest" Dec 11 08:54:50 crc kubenswrapper[4860]: I1211 08:54:50.233749 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3b659d05-76c9-4b04-a05e-a7dd488e9852-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"3b659d05-76c9-4b04-a05e-a7dd488e9852\") " pod="openstack/tempest-tests-tempest" Dec 11 08:54:50 crc kubenswrapper[4860]: I1211 08:54:50.233817 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/3b659d05-76c9-4b04-a05e-a7dd488e9852-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"3b659d05-76c9-4b04-a05e-a7dd488e9852\") " pod="openstack/tempest-tests-tempest" Dec 11 08:54:50 crc kubenswrapper[4860]: I1211 08:54:50.233880 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/3b659d05-76c9-4b04-a05e-a7dd488e9852-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"3b659d05-76c9-4b04-a05e-a7dd488e9852\") " pod="openstack/tempest-tests-tempest" Dec 11 08:54:50 crc kubenswrapper[4860]: I1211 08:54:50.233910 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/3b659d05-76c9-4b04-a05e-a7dd488e9852-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"3b659d05-76c9-4b04-a05e-a7dd488e9852\") " pod="openstack/tempest-tests-tempest" Dec 11 08:54:50 crc kubenswrapper[4860]: I1211 08:54:50.234131 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3b659d05-76c9-4b04-a05e-a7dd488e9852-config-data\") pod \"tempest-tests-tempest\" (UID: \"3b659d05-76c9-4b04-a05e-a7dd488e9852\") " pod="openstack/tempest-tests-tempest" Dec 11 08:54:50 crc kubenswrapper[4860]: I1211 08:54:50.335693 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q4rtg\" (UniqueName: \"kubernetes.io/projected/3b659d05-76c9-4b04-a05e-a7dd488e9852-kube-api-access-q4rtg\") pod \"tempest-tests-tempest\" (UID: \"3b659d05-76c9-4b04-a05e-a7dd488e9852\") " pod="openstack/tempest-tests-tempest" Dec 11 08:54:50 crc kubenswrapper[4860]: I1211 08:54:50.335805 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3b659d05-76c9-4b04-a05e-a7dd488e9852-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"3b659d05-76c9-4b04-a05e-a7dd488e9852\") " pod="openstack/tempest-tests-tempest" Dec 11 08:54:50 crc kubenswrapper[4860]: I1211 08:54:50.335846 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/3b659d05-76c9-4b04-a05e-a7dd488e9852-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"3b659d05-76c9-4b04-a05e-a7dd488e9852\") " pod="openstack/tempest-tests-tempest" Dec 11 08:54:50 crc kubenswrapper[4860]: I1211 08:54:50.335881 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/3b659d05-76c9-4b04-a05e-a7dd488e9852-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"3b659d05-76c9-4b04-a05e-a7dd488e9852\") " pod="openstack/tempest-tests-tempest" Dec 11 08:54:50 crc kubenswrapper[4860]: I1211 08:54:50.335913 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/3b659d05-76c9-4b04-a05e-a7dd488e9852-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"3b659d05-76c9-4b04-a05e-a7dd488e9852\") " pod="openstack/tempest-tests-tempest" Dec 11 08:54:50 crc kubenswrapper[4860]: I1211 08:54:50.336004 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3b659d05-76c9-4b04-a05e-a7dd488e9852-config-data\") pod \"tempest-tests-tempest\" (UID: \"3b659d05-76c9-4b04-a05e-a7dd488e9852\") " pod="openstack/tempest-tests-tempest" Dec 11 08:54:50 crc kubenswrapper[4860]: I1211 08:54:50.336072 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/3b659d05-76c9-4b04-a05e-a7dd488e9852-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"3b659d05-76c9-4b04-a05e-a7dd488e9852\") " pod="openstack/tempest-tests-tempest" Dec 11 08:54:50 crc kubenswrapper[4860]: I1211 08:54:50.336101 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"tempest-tests-tempest\" (UID: \"3b659d05-76c9-4b04-a05e-a7dd488e9852\") " pod="openstack/tempest-tests-tempest" Dec 11 08:54:50 crc kubenswrapper[4860]: I1211 08:54:50.336137 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/3b659d05-76c9-4b04-a05e-a7dd488e9852-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"3b659d05-76c9-4b04-a05e-a7dd488e9852\") " pod="openstack/tempest-tests-tempest" Dec 11 08:54:50 crc kubenswrapper[4860]: I1211 08:54:50.336623 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/3b659d05-76c9-4b04-a05e-a7dd488e9852-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"3b659d05-76c9-4b04-a05e-a7dd488e9852\") " pod="openstack/tempest-tests-tempest" Dec 11 08:54:50 crc kubenswrapper[4860]: I1211 08:54:50.337138 4860 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"tempest-tests-tempest\" (UID: \"3b659d05-76c9-4b04-a05e-a7dd488e9852\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/tempest-tests-tempest" Dec 11 08:54:50 crc kubenswrapper[4860]: I1211 08:54:50.337145 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/3b659d05-76c9-4b04-a05e-a7dd488e9852-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"3b659d05-76c9-4b04-a05e-a7dd488e9852\") " pod="openstack/tempest-tests-tempest" Dec 11 08:54:50 crc kubenswrapper[4860]: I1211 08:54:50.337240 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/3b659d05-76c9-4b04-a05e-a7dd488e9852-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"3b659d05-76c9-4b04-a05e-a7dd488e9852\") " pod="openstack/tempest-tests-tempest" Dec 11 08:54:50 crc kubenswrapper[4860]: I1211 08:54:50.337871 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3b659d05-76c9-4b04-a05e-a7dd488e9852-config-data\") pod \"tempest-tests-tempest\" (UID: \"3b659d05-76c9-4b04-a05e-a7dd488e9852\") " pod="openstack/tempest-tests-tempest" Dec 11 08:54:50 crc kubenswrapper[4860]: I1211 08:54:50.341770 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3b659d05-76c9-4b04-a05e-a7dd488e9852-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"3b659d05-76c9-4b04-a05e-a7dd488e9852\") " pod="openstack/tempest-tests-tempest" Dec 11 08:54:50 crc kubenswrapper[4860]: I1211 08:54:50.342247 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/3b659d05-76c9-4b04-a05e-a7dd488e9852-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"3b659d05-76c9-4b04-a05e-a7dd488e9852\") " pod="openstack/tempest-tests-tempest" Dec 11 08:54:50 crc kubenswrapper[4860]: I1211 08:54:50.355526 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/3b659d05-76c9-4b04-a05e-a7dd488e9852-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"3b659d05-76c9-4b04-a05e-a7dd488e9852\") " pod="openstack/tempest-tests-tempest" Dec 11 08:54:50 crc kubenswrapper[4860]: I1211 08:54:50.363191 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q4rtg\" (UniqueName: \"kubernetes.io/projected/3b659d05-76c9-4b04-a05e-a7dd488e9852-kube-api-access-q4rtg\") pod \"tempest-tests-tempest\" (UID: \"3b659d05-76c9-4b04-a05e-a7dd488e9852\") " pod="openstack/tempest-tests-tempest" Dec 11 08:54:50 crc kubenswrapper[4860]: I1211 08:54:50.395330 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"tempest-tests-tempest\" (UID: \"3b659d05-76c9-4b04-a05e-a7dd488e9852\") " pod="openstack/tempest-tests-tempest" Dec 11 08:54:50 crc kubenswrapper[4860]: I1211 08:54:50.420067 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 11 08:54:50 crc kubenswrapper[4860]: I1211 08:54:50.899807 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Dec 11 08:54:51 crc kubenswrapper[4860]: I1211 08:54:51.524628 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"3b659d05-76c9-4b04-a05e-a7dd488e9852","Type":"ContainerStarted","Data":"bd540183ed09af10fb87bc2ab950c566a34bdf06b893dac04f7da1ec42044907"} Dec 11 08:54:52 crc kubenswrapper[4860]: I1211 08:54:52.580189 4860 scope.go:117] "RemoveContainer" containerID="b747bbead8584985306f457adcbaf89164c99c29fe92a3c021617afef9e04489" Dec 11 08:54:52 crc kubenswrapper[4860]: E1211 08:54:52.580439 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-99qgp_openshift-machine-config-operator(31c30642-6e60-41b4-a477-0d802424e0aa)\"" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" Dec 11 08:55:06 crc kubenswrapper[4860]: I1211 08:55:06.579044 4860 scope.go:117] "RemoveContainer" containerID="b747bbead8584985306f457adcbaf89164c99c29fe92a3c021617afef9e04489" Dec 11 08:55:06 crc kubenswrapper[4860]: E1211 08:55:06.579852 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-99qgp_openshift-machine-config-operator(31c30642-6e60-41b4-a477-0d802424e0aa)\"" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" Dec 11 08:55:19 crc kubenswrapper[4860]: I1211 08:55:19.579803 4860 scope.go:117] "RemoveContainer" containerID="b747bbead8584985306f457adcbaf89164c99c29fe92a3c021617afef9e04489" Dec 11 08:55:19 crc kubenswrapper[4860]: E1211 08:55:19.580495 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-99qgp_openshift-machine-config-operator(31c30642-6e60-41b4-a477-0d802424e0aa)\"" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" Dec 11 08:55:22 crc kubenswrapper[4860]: E1211 08:55:22.133000 4860 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Dec 11 08:55:22 crc kubenswrapper[4860]: E1211 08:55:22.133440 4860 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-q4rtg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(3b659d05-76c9-4b04-a05e-a7dd488e9852): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 11 08:55:22 crc kubenswrapper[4860]: E1211 08:55:22.134669 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="3b659d05-76c9-4b04-a05e-a7dd488e9852" Dec 11 08:55:22 crc kubenswrapper[4860]: E1211 08:55:22.823453 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest" podUID="3b659d05-76c9-4b04-a05e-a7dd488e9852" Dec 11 08:55:27 crc kubenswrapper[4860]: I1211 08:55:27.906194 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-slv27"] Dec 11 08:55:27 crc kubenswrapper[4860]: I1211 08:55:27.910524 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-slv27" Dec 11 08:55:27 crc kubenswrapper[4860]: I1211 08:55:27.918196 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-slv27"] Dec 11 08:55:28 crc kubenswrapper[4860]: I1211 08:55:28.072653 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb7f71cc-a761-48f6-8749-25a33a2d5c5d-utilities\") pod \"community-operators-slv27\" (UID: \"bb7f71cc-a761-48f6-8749-25a33a2d5c5d\") " pod="openshift-marketplace/community-operators-slv27" Dec 11 08:55:28 crc kubenswrapper[4860]: I1211 08:55:28.072707 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb7f71cc-a761-48f6-8749-25a33a2d5c5d-catalog-content\") pod \"community-operators-slv27\" (UID: \"bb7f71cc-a761-48f6-8749-25a33a2d5c5d\") " pod="openshift-marketplace/community-operators-slv27" Dec 11 08:55:28 crc kubenswrapper[4860]: I1211 08:55:28.072901 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9hdlt\" (UniqueName: \"kubernetes.io/projected/bb7f71cc-a761-48f6-8749-25a33a2d5c5d-kube-api-access-9hdlt\") pod \"community-operators-slv27\" (UID: \"bb7f71cc-a761-48f6-8749-25a33a2d5c5d\") " pod="openshift-marketplace/community-operators-slv27" Dec 11 08:55:28 crc kubenswrapper[4860]: I1211 08:55:28.174669 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9hdlt\" (UniqueName: \"kubernetes.io/projected/bb7f71cc-a761-48f6-8749-25a33a2d5c5d-kube-api-access-9hdlt\") pod \"community-operators-slv27\" (UID: \"bb7f71cc-a761-48f6-8749-25a33a2d5c5d\") " pod="openshift-marketplace/community-operators-slv27" Dec 11 08:55:28 crc kubenswrapper[4860]: I1211 08:55:28.174769 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb7f71cc-a761-48f6-8749-25a33a2d5c5d-utilities\") pod \"community-operators-slv27\" (UID: \"bb7f71cc-a761-48f6-8749-25a33a2d5c5d\") " pod="openshift-marketplace/community-operators-slv27" Dec 11 08:55:28 crc kubenswrapper[4860]: I1211 08:55:28.174802 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb7f71cc-a761-48f6-8749-25a33a2d5c5d-catalog-content\") pod \"community-operators-slv27\" (UID: \"bb7f71cc-a761-48f6-8749-25a33a2d5c5d\") " pod="openshift-marketplace/community-operators-slv27" Dec 11 08:55:28 crc kubenswrapper[4860]: I1211 08:55:28.175474 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb7f71cc-a761-48f6-8749-25a33a2d5c5d-catalog-content\") pod \"community-operators-slv27\" (UID: \"bb7f71cc-a761-48f6-8749-25a33a2d5c5d\") " pod="openshift-marketplace/community-operators-slv27" Dec 11 08:55:28 crc kubenswrapper[4860]: I1211 08:55:28.175496 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb7f71cc-a761-48f6-8749-25a33a2d5c5d-utilities\") pod \"community-operators-slv27\" (UID: \"bb7f71cc-a761-48f6-8749-25a33a2d5c5d\") " pod="openshift-marketplace/community-operators-slv27" Dec 11 08:55:28 crc kubenswrapper[4860]: I1211 08:55:28.195770 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9hdlt\" (UniqueName: \"kubernetes.io/projected/bb7f71cc-a761-48f6-8749-25a33a2d5c5d-kube-api-access-9hdlt\") pod \"community-operators-slv27\" (UID: \"bb7f71cc-a761-48f6-8749-25a33a2d5c5d\") " pod="openshift-marketplace/community-operators-slv27" Dec 11 08:55:28 crc kubenswrapper[4860]: I1211 08:55:28.242056 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-slv27" Dec 11 08:55:28 crc kubenswrapper[4860]: W1211 08:55:28.795553 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbb7f71cc_a761_48f6_8749_25a33a2d5c5d.slice/crio-276da5110b78d113d924568f9428137f140cf053814d7c571671a39d304569be WatchSource:0}: Error finding container 276da5110b78d113d924568f9428137f140cf053814d7c571671a39d304569be: Status 404 returned error can't find the container with id 276da5110b78d113d924568f9428137f140cf053814d7c571671a39d304569be Dec 11 08:55:28 crc kubenswrapper[4860]: I1211 08:55:28.798036 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-slv27"] Dec 11 08:55:28 crc kubenswrapper[4860]: I1211 08:55:28.876378 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-slv27" event={"ID":"bb7f71cc-a761-48f6-8749-25a33a2d5c5d","Type":"ContainerStarted","Data":"276da5110b78d113d924568f9428137f140cf053814d7c571671a39d304569be"} Dec 11 08:55:29 crc kubenswrapper[4860]: I1211 08:55:29.887708 4860 generic.go:334] "Generic (PLEG): container finished" podID="bb7f71cc-a761-48f6-8749-25a33a2d5c5d" containerID="51bac1372f9d26b98b99f3a545a65774c8fb124c6d5ea2621e53c5ac9570767a" exitCode=0 Dec 11 08:55:29 crc kubenswrapper[4860]: I1211 08:55:29.887794 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-slv27" event={"ID":"bb7f71cc-a761-48f6-8749-25a33a2d5c5d","Type":"ContainerDied","Data":"51bac1372f9d26b98b99f3a545a65774c8fb124c6d5ea2621e53c5ac9570767a"} Dec 11 08:55:30 crc kubenswrapper[4860]: I1211 08:55:30.579296 4860 scope.go:117] "RemoveContainer" containerID="b747bbead8584985306f457adcbaf89164c99c29fe92a3c021617afef9e04489" Dec 11 08:55:30 crc kubenswrapper[4860]: E1211 08:55:30.579810 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-99qgp_openshift-machine-config-operator(31c30642-6e60-41b4-a477-0d802424e0aa)\"" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" Dec 11 08:55:30 crc kubenswrapper[4860]: I1211 08:55:30.899255 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-slv27" event={"ID":"bb7f71cc-a761-48f6-8749-25a33a2d5c5d","Type":"ContainerStarted","Data":"10b1058083dc1341da92b2116ff70a327f303e2595dd670f7741ae20dabecace"} Dec 11 08:55:31 crc kubenswrapper[4860]: I1211 08:55:31.924182 4860 generic.go:334] "Generic (PLEG): container finished" podID="bb7f71cc-a761-48f6-8749-25a33a2d5c5d" containerID="10b1058083dc1341da92b2116ff70a327f303e2595dd670f7741ae20dabecace" exitCode=0 Dec 11 08:55:31 crc kubenswrapper[4860]: I1211 08:55:31.924506 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-slv27" event={"ID":"bb7f71cc-a761-48f6-8749-25a33a2d5c5d","Type":"ContainerDied","Data":"10b1058083dc1341da92b2116ff70a327f303e2595dd670f7741ae20dabecace"} Dec 11 08:55:32 crc kubenswrapper[4860]: I1211 08:55:32.936898 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-slv27" event={"ID":"bb7f71cc-a761-48f6-8749-25a33a2d5c5d","Type":"ContainerStarted","Data":"ecd79f18a97bd96fb9100822e84a8ad6d15595be73a89217eab927ef489ffc40"} Dec 11 08:55:32 crc kubenswrapper[4860]: I1211 08:55:32.958252 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-slv27" podStartSLOduration=3.378531026 podStartE2EDuration="5.958229265s" podCreationTimestamp="2025-12-11 08:55:27 +0000 UTC" firstStartedPulling="2025-12-11 08:55:29.890608334 +0000 UTC m=+2662.619127379" lastFinishedPulling="2025-12-11 08:55:32.470306563 +0000 UTC m=+2665.198825618" observedRunningTime="2025-12-11 08:55:32.952483445 +0000 UTC m=+2665.681002500" watchObservedRunningTime="2025-12-11 08:55:32.958229265 +0000 UTC m=+2665.686748320" Dec 11 08:55:36 crc kubenswrapper[4860]: I1211 08:55:36.475191 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Dec 11 08:55:37 crc kubenswrapper[4860]: I1211 08:55:37.978242 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"3b659d05-76c9-4b04-a05e-a7dd488e9852","Type":"ContainerStarted","Data":"e8b2b959a1b90a4db8fbd7a58476c9c82fb1f896125d0763a74f9579fa5726c2"} Dec 11 08:55:37 crc kubenswrapper[4860]: I1211 08:55:37.999476 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=3.428886343 podStartE2EDuration="48.999459145s" podCreationTimestamp="2025-12-11 08:54:49 +0000 UTC" firstStartedPulling="2025-12-11 08:54:50.902224886 +0000 UTC m=+2623.630743941" lastFinishedPulling="2025-12-11 08:55:36.472797688 +0000 UTC m=+2669.201316743" observedRunningTime="2025-12-11 08:55:37.992681124 +0000 UTC m=+2670.721200199" watchObservedRunningTime="2025-12-11 08:55:37.999459145 +0000 UTC m=+2670.727978200" Dec 11 08:55:38 crc kubenswrapper[4860]: I1211 08:55:38.242732 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-slv27" Dec 11 08:55:38 crc kubenswrapper[4860]: I1211 08:55:38.243084 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-slv27" Dec 11 08:55:38 crc kubenswrapper[4860]: I1211 08:55:38.294619 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-slv27" Dec 11 08:55:39 crc kubenswrapper[4860]: I1211 08:55:39.038594 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-slv27" Dec 11 08:55:39 crc kubenswrapper[4860]: I1211 08:55:39.090863 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-slv27"] Dec 11 08:55:41 crc kubenswrapper[4860]: I1211 08:55:41.003838 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-slv27" podUID="bb7f71cc-a761-48f6-8749-25a33a2d5c5d" containerName="registry-server" containerID="cri-o://ecd79f18a97bd96fb9100822e84a8ad6d15595be73a89217eab927ef489ffc40" gracePeriod=2 Dec 11 08:55:41 crc kubenswrapper[4860]: I1211 08:55:41.430766 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-llgls"] Dec 11 08:55:41 crc kubenswrapper[4860]: I1211 08:55:41.434390 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-llgls" Dec 11 08:55:41 crc kubenswrapper[4860]: I1211 08:55:41.446855 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-llgls"] Dec 11 08:55:41 crc kubenswrapper[4860]: I1211 08:55:41.567057 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8tn8c\" (UniqueName: \"kubernetes.io/projected/b83bd05e-eca8-484b-a317-7d9c064e46c4-kube-api-access-8tn8c\") pod \"certified-operators-llgls\" (UID: \"b83bd05e-eca8-484b-a317-7d9c064e46c4\") " pod="openshift-marketplace/certified-operators-llgls" Dec 11 08:55:41 crc kubenswrapper[4860]: I1211 08:55:41.567381 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b83bd05e-eca8-484b-a317-7d9c064e46c4-catalog-content\") pod \"certified-operators-llgls\" (UID: \"b83bd05e-eca8-484b-a317-7d9c064e46c4\") " pod="openshift-marketplace/certified-operators-llgls" Dec 11 08:55:41 crc kubenswrapper[4860]: I1211 08:55:41.567509 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b83bd05e-eca8-484b-a317-7d9c064e46c4-utilities\") pod \"certified-operators-llgls\" (UID: \"b83bd05e-eca8-484b-a317-7d9c064e46c4\") " pod="openshift-marketplace/certified-operators-llgls" Dec 11 08:55:41 crc kubenswrapper[4860]: I1211 08:55:41.669417 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b83bd05e-eca8-484b-a317-7d9c064e46c4-utilities\") pod \"certified-operators-llgls\" (UID: \"b83bd05e-eca8-484b-a317-7d9c064e46c4\") " pod="openshift-marketplace/certified-operators-llgls" Dec 11 08:55:41 crc kubenswrapper[4860]: I1211 08:55:41.669513 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8tn8c\" (UniqueName: \"kubernetes.io/projected/b83bd05e-eca8-484b-a317-7d9c064e46c4-kube-api-access-8tn8c\") pod \"certified-operators-llgls\" (UID: \"b83bd05e-eca8-484b-a317-7d9c064e46c4\") " pod="openshift-marketplace/certified-operators-llgls" Dec 11 08:55:41 crc kubenswrapper[4860]: I1211 08:55:41.669548 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b83bd05e-eca8-484b-a317-7d9c064e46c4-catalog-content\") pod \"certified-operators-llgls\" (UID: \"b83bd05e-eca8-484b-a317-7d9c064e46c4\") " pod="openshift-marketplace/certified-operators-llgls" Dec 11 08:55:41 crc kubenswrapper[4860]: I1211 08:55:41.670009 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b83bd05e-eca8-484b-a317-7d9c064e46c4-utilities\") pod \"certified-operators-llgls\" (UID: \"b83bd05e-eca8-484b-a317-7d9c064e46c4\") " pod="openshift-marketplace/certified-operators-llgls" Dec 11 08:55:41 crc kubenswrapper[4860]: I1211 08:55:41.670165 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b83bd05e-eca8-484b-a317-7d9c064e46c4-catalog-content\") pod \"certified-operators-llgls\" (UID: \"b83bd05e-eca8-484b-a317-7d9c064e46c4\") " pod="openshift-marketplace/certified-operators-llgls" Dec 11 08:55:41 crc kubenswrapper[4860]: I1211 08:55:41.708119 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8tn8c\" (UniqueName: \"kubernetes.io/projected/b83bd05e-eca8-484b-a317-7d9c064e46c4-kube-api-access-8tn8c\") pod \"certified-operators-llgls\" (UID: \"b83bd05e-eca8-484b-a317-7d9c064e46c4\") " pod="openshift-marketplace/certified-operators-llgls" Dec 11 08:55:41 crc kubenswrapper[4860]: I1211 08:55:41.819891 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-llgls" Dec 11 08:55:42 crc kubenswrapper[4860]: I1211 08:55:42.062665 4860 generic.go:334] "Generic (PLEG): container finished" podID="bb7f71cc-a761-48f6-8749-25a33a2d5c5d" containerID="ecd79f18a97bd96fb9100822e84a8ad6d15595be73a89217eab927ef489ffc40" exitCode=0 Dec 11 08:55:42 crc kubenswrapper[4860]: I1211 08:55:42.063008 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-slv27" event={"ID":"bb7f71cc-a761-48f6-8749-25a33a2d5c5d","Type":"ContainerDied","Data":"ecd79f18a97bd96fb9100822e84a8ad6d15595be73a89217eab927ef489ffc40"} Dec 11 08:55:42 crc kubenswrapper[4860]: I1211 08:55:42.063041 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-slv27" event={"ID":"bb7f71cc-a761-48f6-8749-25a33a2d5c5d","Type":"ContainerDied","Data":"276da5110b78d113d924568f9428137f140cf053814d7c571671a39d304569be"} Dec 11 08:55:42 crc kubenswrapper[4860]: I1211 08:55:42.063055 4860 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="276da5110b78d113d924568f9428137f140cf053814d7c571671a39d304569be" Dec 11 08:55:42 crc kubenswrapper[4860]: I1211 08:55:42.067713 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-slv27" Dec 11 08:55:42 crc kubenswrapper[4860]: I1211 08:55:42.186677 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb7f71cc-a761-48f6-8749-25a33a2d5c5d-catalog-content\") pod \"bb7f71cc-a761-48f6-8749-25a33a2d5c5d\" (UID: \"bb7f71cc-a761-48f6-8749-25a33a2d5c5d\") " Dec 11 08:55:42 crc kubenswrapper[4860]: I1211 08:55:42.186839 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb7f71cc-a761-48f6-8749-25a33a2d5c5d-utilities\") pod \"bb7f71cc-a761-48f6-8749-25a33a2d5c5d\" (UID: \"bb7f71cc-a761-48f6-8749-25a33a2d5c5d\") " Dec 11 08:55:42 crc kubenswrapper[4860]: I1211 08:55:42.186915 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9hdlt\" (UniqueName: \"kubernetes.io/projected/bb7f71cc-a761-48f6-8749-25a33a2d5c5d-kube-api-access-9hdlt\") pod \"bb7f71cc-a761-48f6-8749-25a33a2d5c5d\" (UID: \"bb7f71cc-a761-48f6-8749-25a33a2d5c5d\") " Dec 11 08:55:42 crc kubenswrapper[4860]: I1211 08:55:42.188586 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bb7f71cc-a761-48f6-8749-25a33a2d5c5d-utilities" (OuterVolumeSpecName: "utilities") pod "bb7f71cc-a761-48f6-8749-25a33a2d5c5d" (UID: "bb7f71cc-a761-48f6-8749-25a33a2d5c5d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:55:42 crc kubenswrapper[4860]: I1211 08:55:42.194863 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb7f71cc-a761-48f6-8749-25a33a2d5c5d-kube-api-access-9hdlt" (OuterVolumeSpecName: "kube-api-access-9hdlt") pod "bb7f71cc-a761-48f6-8749-25a33a2d5c5d" (UID: "bb7f71cc-a761-48f6-8749-25a33a2d5c5d"). InnerVolumeSpecName "kube-api-access-9hdlt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:55:42 crc kubenswrapper[4860]: I1211 08:55:42.249525 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bb7f71cc-a761-48f6-8749-25a33a2d5c5d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bb7f71cc-a761-48f6-8749-25a33a2d5c5d" (UID: "bb7f71cc-a761-48f6-8749-25a33a2d5c5d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:55:42 crc kubenswrapper[4860]: I1211 08:55:42.289354 4860 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb7f71cc-a761-48f6-8749-25a33a2d5c5d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 08:55:42 crc kubenswrapper[4860]: I1211 08:55:42.289387 4860 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb7f71cc-a761-48f6-8749-25a33a2d5c5d-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 08:55:42 crc kubenswrapper[4860]: I1211 08:55:42.289397 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9hdlt\" (UniqueName: \"kubernetes.io/projected/bb7f71cc-a761-48f6-8749-25a33a2d5c5d-kube-api-access-9hdlt\") on node \"crc\" DevicePath \"\"" Dec 11 08:55:42 crc kubenswrapper[4860]: W1211 08:55:42.365618 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb83bd05e_eca8_484b_a317_7d9c064e46c4.slice/crio-decee745d5e5ce82c59c370ff36edc9e72830ef8dd9ab2613880a8d7821753a6 WatchSource:0}: Error finding container decee745d5e5ce82c59c370ff36edc9e72830ef8dd9ab2613880a8d7821753a6: Status 404 returned error can't find the container with id decee745d5e5ce82c59c370ff36edc9e72830ef8dd9ab2613880a8d7821753a6 Dec 11 08:55:42 crc kubenswrapper[4860]: I1211 08:55:42.367083 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-llgls"] Dec 11 08:55:43 crc kubenswrapper[4860]: I1211 08:55:43.073166 4860 generic.go:334] "Generic (PLEG): container finished" podID="b83bd05e-eca8-484b-a317-7d9c064e46c4" containerID="752a7f9d746534bb1290590edb12593b2e2c726ccfbbd761f89a1d14a9573913" exitCode=0 Dec 11 08:55:43 crc kubenswrapper[4860]: I1211 08:55:43.073208 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-llgls" event={"ID":"b83bd05e-eca8-484b-a317-7d9c064e46c4","Type":"ContainerDied","Data":"752a7f9d746534bb1290590edb12593b2e2c726ccfbbd761f89a1d14a9573913"} Dec 11 08:55:43 crc kubenswrapper[4860]: I1211 08:55:43.073625 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-llgls" event={"ID":"b83bd05e-eca8-484b-a317-7d9c064e46c4","Type":"ContainerStarted","Data":"decee745d5e5ce82c59c370ff36edc9e72830ef8dd9ab2613880a8d7821753a6"} Dec 11 08:55:43 crc kubenswrapper[4860]: I1211 08:55:43.073961 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-slv27" Dec 11 08:55:43 crc kubenswrapper[4860]: I1211 08:55:43.150257 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-slv27"] Dec 11 08:55:43 crc kubenswrapper[4860]: I1211 08:55:43.160523 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-slv27"] Dec 11 08:55:43 crc kubenswrapper[4860]: I1211 08:55:43.591280 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bb7f71cc-a761-48f6-8749-25a33a2d5c5d" path="/var/lib/kubelet/pods/bb7f71cc-a761-48f6-8749-25a33a2d5c5d/volumes" Dec 11 08:55:44 crc kubenswrapper[4860]: I1211 08:55:44.084932 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-llgls" event={"ID":"b83bd05e-eca8-484b-a317-7d9c064e46c4","Type":"ContainerStarted","Data":"79e640166977a97f4009dc09c574f72b0689faec8a134e2e86fa92b08a3a228d"} Dec 11 08:55:45 crc kubenswrapper[4860]: I1211 08:55:45.096539 4860 generic.go:334] "Generic (PLEG): container finished" podID="b83bd05e-eca8-484b-a317-7d9c064e46c4" containerID="79e640166977a97f4009dc09c574f72b0689faec8a134e2e86fa92b08a3a228d" exitCode=0 Dec 11 08:55:45 crc kubenswrapper[4860]: I1211 08:55:45.096981 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-llgls" event={"ID":"b83bd05e-eca8-484b-a317-7d9c064e46c4","Type":"ContainerDied","Data":"79e640166977a97f4009dc09c574f72b0689faec8a134e2e86fa92b08a3a228d"} Dec 11 08:55:45 crc kubenswrapper[4860]: I1211 08:55:45.579687 4860 scope.go:117] "RemoveContainer" containerID="b747bbead8584985306f457adcbaf89164c99c29fe92a3c021617afef9e04489" Dec 11 08:55:45 crc kubenswrapper[4860]: E1211 08:55:45.580048 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-99qgp_openshift-machine-config-operator(31c30642-6e60-41b4-a477-0d802424e0aa)\"" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" Dec 11 08:55:46 crc kubenswrapper[4860]: I1211 08:55:46.108341 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-llgls" event={"ID":"b83bd05e-eca8-484b-a317-7d9c064e46c4","Type":"ContainerStarted","Data":"6dd260dda271921aae21fbe73358133a2bca903c412e41ac2ca51993f32f444b"} Dec 11 08:55:46 crc kubenswrapper[4860]: I1211 08:55:46.127801 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-llgls" podStartSLOduration=2.374308509 podStartE2EDuration="5.127779224s" podCreationTimestamp="2025-12-11 08:55:41 +0000 UTC" firstStartedPulling="2025-12-11 08:55:43.074950174 +0000 UTC m=+2675.803469229" lastFinishedPulling="2025-12-11 08:55:45.828420889 +0000 UTC m=+2678.556939944" observedRunningTime="2025-12-11 08:55:46.123371167 +0000 UTC m=+2678.851890222" watchObservedRunningTime="2025-12-11 08:55:46.127779224 +0000 UTC m=+2678.856298279" Dec 11 08:55:51 crc kubenswrapper[4860]: I1211 08:55:51.821098 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-llgls" Dec 11 08:55:51 crc kubenswrapper[4860]: I1211 08:55:51.821698 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-llgls" Dec 11 08:55:51 crc kubenswrapper[4860]: I1211 08:55:51.866286 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-llgls" Dec 11 08:55:52 crc kubenswrapper[4860]: I1211 08:55:52.215692 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-llgls" Dec 11 08:55:52 crc kubenswrapper[4860]: I1211 08:55:52.262338 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-llgls"] Dec 11 08:55:54 crc kubenswrapper[4860]: I1211 08:55:54.177142 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-llgls" podUID="b83bd05e-eca8-484b-a317-7d9c064e46c4" containerName="registry-server" containerID="cri-o://6dd260dda271921aae21fbe73358133a2bca903c412e41ac2ca51993f32f444b" gracePeriod=2 Dec 11 08:55:55 crc kubenswrapper[4860]: I1211 08:55:55.195414 4860 generic.go:334] "Generic (PLEG): container finished" podID="b83bd05e-eca8-484b-a317-7d9c064e46c4" containerID="6dd260dda271921aae21fbe73358133a2bca903c412e41ac2ca51993f32f444b" exitCode=0 Dec 11 08:55:55 crc kubenswrapper[4860]: I1211 08:55:55.195829 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-llgls" event={"ID":"b83bd05e-eca8-484b-a317-7d9c064e46c4","Type":"ContainerDied","Data":"6dd260dda271921aae21fbe73358133a2bca903c412e41ac2ca51993f32f444b"} Dec 11 08:55:55 crc kubenswrapper[4860]: I1211 08:55:55.780575 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-llgls" Dec 11 08:55:55 crc kubenswrapper[4860]: I1211 08:55:55.871940 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b83bd05e-eca8-484b-a317-7d9c064e46c4-utilities\") pod \"b83bd05e-eca8-484b-a317-7d9c064e46c4\" (UID: \"b83bd05e-eca8-484b-a317-7d9c064e46c4\") " Dec 11 08:55:55 crc kubenswrapper[4860]: I1211 08:55:55.872016 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tn8c\" (UniqueName: \"kubernetes.io/projected/b83bd05e-eca8-484b-a317-7d9c064e46c4-kube-api-access-8tn8c\") pod \"b83bd05e-eca8-484b-a317-7d9c064e46c4\" (UID: \"b83bd05e-eca8-484b-a317-7d9c064e46c4\") " Dec 11 08:55:55 crc kubenswrapper[4860]: I1211 08:55:55.872133 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b83bd05e-eca8-484b-a317-7d9c064e46c4-catalog-content\") pod \"b83bd05e-eca8-484b-a317-7d9c064e46c4\" (UID: \"b83bd05e-eca8-484b-a317-7d9c064e46c4\") " Dec 11 08:55:55 crc kubenswrapper[4860]: I1211 08:55:55.872980 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b83bd05e-eca8-484b-a317-7d9c064e46c4-utilities" (OuterVolumeSpecName: "utilities") pod "b83bd05e-eca8-484b-a317-7d9c064e46c4" (UID: "b83bd05e-eca8-484b-a317-7d9c064e46c4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:55:55 crc kubenswrapper[4860]: I1211 08:55:55.877843 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b83bd05e-eca8-484b-a317-7d9c064e46c4-kube-api-access-8tn8c" (OuterVolumeSpecName: "kube-api-access-8tn8c") pod "b83bd05e-eca8-484b-a317-7d9c064e46c4" (UID: "b83bd05e-eca8-484b-a317-7d9c064e46c4"). InnerVolumeSpecName "kube-api-access-8tn8c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:55:55 crc kubenswrapper[4860]: I1211 08:55:55.931760 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b83bd05e-eca8-484b-a317-7d9c064e46c4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b83bd05e-eca8-484b-a317-7d9c064e46c4" (UID: "b83bd05e-eca8-484b-a317-7d9c064e46c4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:55:55 crc kubenswrapper[4860]: I1211 08:55:55.974097 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tn8c\" (UniqueName: \"kubernetes.io/projected/b83bd05e-eca8-484b-a317-7d9c064e46c4-kube-api-access-8tn8c\") on node \"crc\" DevicePath \"\"" Dec 11 08:55:55 crc kubenswrapper[4860]: I1211 08:55:55.974134 4860 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b83bd05e-eca8-484b-a317-7d9c064e46c4-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 08:55:55 crc kubenswrapper[4860]: I1211 08:55:55.974148 4860 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b83bd05e-eca8-484b-a317-7d9c064e46c4-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 08:55:56 crc kubenswrapper[4860]: I1211 08:55:56.209428 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-llgls" event={"ID":"b83bd05e-eca8-484b-a317-7d9c064e46c4","Type":"ContainerDied","Data":"decee745d5e5ce82c59c370ff36edc9e72830ef8dd9ab2613880a8d7821753a6"} Dec 11 08:55:56 crc kubenswrapper[4860]: I1211 08:55:56.209501 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-llgls" Dec 11 08:55:56 crc kubenswrapper[4860]: I1211 08:55:56.209843 4860 scope.go:117] "RemoveContainer" containerID="6dd260dda271921aae21fbe73358133a2bca903c412e41ac2ca51993f32f444b" Dec 11 08:55:56 crc kubenswrapper[4860]: I1211 08:55:56.245601 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-llgls"] Dec 11 08:55:56 crc kubenswrapper[4860]: I1211 08:55:56.247834 4860 scope.go:117] "RemoveContainer" containerID="79e640166977a97f4009dc09c574f72b0689faec8a134e2e86fa92b08a3a228d" Dec 11 08:55:56 crc kubenswrapper[4860]: I1211 08:55:56.254896 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-llgls"] Dec 11 08:55:56 crc kubenswrapper[4860]: I1211 08:55:56.299170 4860 scope.go:117] "RemoveContainer" containerID="752a7f9d746534bb1290590edb12593b2e2c726ccfbbd761f89a1d14a9573913" Dec 11 08:55:57 crc kubenswrapper[4860]: I1211 08:55:57.591698 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b83bd05e-eca8-484b-a317-7d9c064e46c4" path="/var/lib/kubelet/pods/b83bd05e-eca8-484b-a317-7d9c064e46c4/volumes" Dec 11 08:55:58 crc kubenswrapper[4860]: I1211 08:55:58.579713 4860 scope.go:117] "RemoveContainer" containerID="b747bbead8584985306f457adcbaf89164c99c29fe92a3c021617afef9e04489" Dec 11 08:55:58 crc kubenswrapper[4860]: E1211 08:55:58.580445 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-99qgp_openshift-machine-config-operator(31c30642-6e60-41b4-a477-0d802424e0aa)\"" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" Dec 11 08:56:02 crc kubenswrapper[4860]: I1211 08:56:02.782485 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-ndvvj"] Dec 11 08:56:02 crc kubenswrapper[4860]: E1211 08:56:02.785267 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb7f71cc-a761-48f6-8749-25a33a2d5c5d" containerName="extract-content" Dec 11 08:56:02 crc kubenswrapper[4860]: I1211 08:56:02.786353 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb7f71cc-a761-48f6-8749-25a33a2d5c5d" containerName="extract-content" Dec 11 08:56:02 crc kubenswrapper[4860]: E1211 08:56:02.786486 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb7f71cc-a761-48f6-8749-25a33a2d5c5d" containerName="extract-utilities" Dec 11 08:56:02 crc kubenswrapper[4860]: I1211 08:56:02.786565 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb7f71cc-a761-48f6-8749-25a33a2d5c5d" containerName="extract-utilities" Dec 11 08:56:02 crc kubenswrapper[4860]: E1211 08:56:02.786682 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b83bd05e-eca8-484b-a317-7d9c064e46c4" containerName="extract-utilities" Dec 11 08:56:02 crc kubenswrapper[4860]: I1211 08:56:02.786772 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="b83bd05e-eca8-484b-a317-7d9c064e46c4" containerName="extract-utilities" Dec 11 08:56:02 crc kubenswrapper[4860]: E1211 08:56:02.786872 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b83bd05e-eca8-484b-a317-7d9c064e46c4" containerName="extract-content" Dec 11 08:56:02 crc kubenswrapper[4860]: I1211 08:56:02.786950 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="b83bd05e-eca8-484b-a317-7d9c064e46c4" containerName="extract-content" Dec 11 08:56:02 crc kubenswrapper[4860]: E1211 08:56:02.787041 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb7f71cc-a761-48f6-8749-25a33a2d5c5d" containerName="registry-server" Dec 11 08:56:02 crc kubenswrapper[4860]: I1211 08:56:02.787123 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb7f71cc-a761-48f6-8749-25a33a2d5c5d" containerName="registry-server" Dec 11 08:56:02 crc kubenswrapper[4860]: E1211 08:56:02.787206 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b83bd05e-eca8-484b-a317-7d9c064e46c4" containerName="registry-server" Dec 11 08:56:02 crc kubenswrapper[4860]: I1211 08:56:02.787289 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="b83bd05e-eca8-484b-a317-7d9c064e46c4" containerName="registry-server" Dec 11 08:56:02 crc kubenswrapper[4860]: I1211 08:56:02.787684 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="b83bd05e-eca8-484b-a317-7d9c064e46c4" containerName="registry-server" Dec 11 08:56:02 crc kubenswrapper[4860]: I1211 08:56:02.787821 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb7f71cc-a761-48f6-8749-25a33a2d5c5d" containerName="registry-server" Dec 11 08:56:02 crc kubenswrapper[4860]: I1211 08:56:02.789792 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ndvvj" Dec 11 08:56:02 crc kubenswrapper[4860]: I1211 08:56:02.791142 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ndvvj"] Dec 11 08:56:02 crc kubenswrapper[4860]: I1211 08:56:02.848081 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7ac2ba0-677d-45e5-8842-c64b96c8fed6-catalog-content\") pod \"redhat-operators-ndvvj\" (UID: \"e7ac2ba0-677d-45e5-8842-c64b96c8fed6\") " pod="openshift-marketplace/redhat-operators-ndvvj" Dec 11 08:56:02 crc kubenswrapper[4860]: I1211 08:56:02.848431 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7ac2ba0-677d-45e5-8842-c64b96c8fed6-utilities\") pod \"redhat-operators-ndvvj\" (UID: \"e7ac2ba0-677d-45e5-8842-c64b96c8fed6\") " pod="openshift-marketplace/redhat-operators-ndvvj" Dec 11 08:56:02 crc kubenswrapper[4860]: I1211 08:56:02.848520 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hls6l\" (UniqueName: \"kubernetes.io/projected/e7ac2ba0-677d-45e5-8842-c64b96c8fed6-kube-api-access-hls6l\") pod \"redhat-operators-ndvvj\" (UID: \"e7ac2ba0-677d-45e5-8842-c64b96c8fed6\") " pod="openshift-marketplace/redhat-operators-ndvvj" Dec 11 08:56:02 crc kubenswrapper[4860]: I1211 08:56:02.950994 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7ac2ba0-677d-45e5-8842-c64b96c8fed6-catalog-content\") pod \"redhat-operators-ndvvj\" (UID: \"e7ac2ba0-677d-45e5-8842-c64b96c8fed6\") " pod="openshift-marketplace/redhat-operators-ndvvj" Dec 11 08:56:02 crc kubenswrapper[4860]: I1211 08:56:02.951062 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7ac2ba0-677d-45e5-8842-c64b96c8fed6-utilities\") pod \"redhat-operators-ndvvj\" (UID: \"e7ac2ba0-677d-45e5-8842-c64b96c8fed6\") " pod="openshift-marketplace/redhat-operators-ndvvj" Dec 11 08:56:02 crc kubenswrapper[4860]: I1211 08:56:02.951084 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hls6l\" (UniqueName: \"kubernetes.io/projected/e7ac2ba0-677d-45e5-8842-c64b96c8fed6-kube-api-access-hls6l\") pod \"redhat-operators-ndvvj\" (UID: \"e7ac2ba0-677d-45e5-8842-c64b96c8fed6\") " pod="openshift-marketplace/redhat-operators-ndvvj" Dec 11 08:56:02 crc kubenswrapper[4860]: I1211 08:56:02.951737 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7ac2ba0-677d-45e5-8842-c64b96c8fed6-catalog-content\") pod \"redhat-operators-ndvvj\" (UID: \"e7ac2ba0-677d-45e5-8842-c64b96c8fed6\") " pod="openshift-marketplace/redhat-operators-ndvvj" Dec 11 08:56:02 crc kubenswrapper[4860]: I1211 08:56:02.951911 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7ac2ba0-677d-45e5-8842-c64b96c8fed6-utilities\") pod \"redhat-operators-ndvvj\" (UID: \"e7ac2ba0-677d-45e5-8842-c64b96c8fed6\") " pod="openshift-marketplace/redhat-operators-ndvvj" Dec 11 08:56:02 crc kubenswrapper[4860]: I1211 08:56:02.971335 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hls6l\" (UniqueName: \"kubernetes.io/projected/e7ac2ba0-677d-45e5-8842-c64b96c8fed6-kube-api-access-hls6l\") pod \"redhat-operators-ndvvj\" (UID: \"e7ac2ba0-677d-45e5-8842-c64b96c8fed6\") " pod="openshift-marketplace/redhat-operators-ndvvj" Dec 11 08:56:03 crc kubenswrapper[4860]: I1211 08:56:03.114762 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ndvvj" Dec 11 08:56:03 crc kubenswrapper[4860]: I1211 08:56:03.613479 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ndvvj"] Dec 11 08:56:04 crc kubenswrapper[4860]: I1211 08:56:04.301753 4860 generic.go:334] "Generic (PLEG): container finished" podID="e7ac2ba0-677d-45e5-8842-c64b96c8fed6" containerID="8f72ec4bd21e4d739d48d19d24c83d9e62ae3937d71afb494367aa4158e1e2d1" exitCode=0 Dec 11 08:56:04 crc kubenswrapper[4860]: I1211 08:56:04.302017 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ndvvj" event={"ID":"e7ac2ba0-677d-45e5-8842-c64b96c8fed6","Type":"ContainerDied","Data":"8f72ec4bd21e4d739d48d19d24c83d9e62ae3937d71afb494367aa4158e1e2d1"} Dec 11 08:56:04 crc kubenswrapper[4860]: I1211 08:56:04.302043 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ndvvj" event={"ID":"e7ac2ba0-677d-45e5-8842-c64b96c8fed6","Type":"ContainerStarted","Data":"0498de5835002fa9c7a9b53fd54e14c2836ebe7c5841b740d6c9b0a1e1562a8f"} Dec 11 08:56:07 crc kubenswrapper[4860]: I1211 08:56:07.337688 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ndvvj" event={"ID":"e7ac2ba0-677d-45e5-8842-c64b96c8fed6","Type":"ContainerStarted","Data":"4a12e21ba3a980ba052f7f68d15536850049694a6bf0aee4b4f8e7ba8f83af91"} Dec 11 08:56:09 crc kubenswrapper[4860]: I1211 08:56:09.358565 4860 generic.go:334] "Generic (PLEG): container finished" podID="e7ac2ba0-677d-45e5-8842-c64b96c8fed6" containerID="4a12e21ba3a980ba052f7f68d15536850049694a6bf0aee4b4f8e7ba8f83af91" exitCode=0 Dec 11 08:56:09 crc kubenswrapper[4860]: I1211 08:56:09.358653 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ndvvj" event={"ID":"e7ac2ba0-677d-45e5-8842-c64b96c8fed6","Type":"ContainerDied","Data":"4a12e21ba3a980ba052f7f68d15536850049694a6bf0aee4b4f8e7ba8f83af91"} Dec 11 08:56:10 crc kubenswrapper[4860]: I1211 08:56:10.374719 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ndvvj" event={"ID":"e7ac2ba0-677d-45e5-8842-c64b96c8fed6","Type":"ContainerStarted","Data":"138aee03b6c0ca7708a2ad0a9bd763d1586f59878c7543251e484377824bd308"} Dec 11 08:56:10 crc kubenswrapper[4860]: I1211 08:56:10.397887 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-ndvvj" podStartSLOduration=2.746013284 podStartE2EDuration="8.397871036s" podCreationTimestamp="2025-12-11 08:56:02 +0000 UTC" firstStartedPulling="2025-12-11 08:56:04.303676516 +0000 UTC m=+2697.032195571" lastFinishedPulling="2025-12-11 08:56:09.955534268 +0000 UTC m=+2702.684053323" observedRunningTime="2025-12-11 08:56:10.39319945 +0000 UTC m=+2703.121718505" watchObservedRunningTime="2025-12-11 08:56:10.397871036 +0000 UTC m=+2703.126390091" Dec 11 08:56:13 crc kubenswrapper[4860]: I1211 08:56:13.115155 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-ndvvj" Dec 11 08:56:13 crc kubenswrapper[4860]: I1211 08:56:13.115512 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-ndvvj" Dec 11 08:56:13 crc kubenswrapper[4860]: I1211 08:56:13.579298 4860 scope.go:117] "RemoveContainer" containerID="b747bbead8584985306f457adcbaf89164c99c29fe92a3c021617afef9e04489" Dec 11 08:56:14 crc kubenswrapper[4860]: I1211 08:56:14.167971 4860 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-ndvvj" podUID="e7ac2ba0-677d-45e5-8842-c64b96c8fed6" containerName="registry-server" probeResult="failure" output=< Dec 11 08:56:14 crc kubenswrapper[4860]: timeout: failed to connect service ":50051" within 1s Dec 11 08:56:14 crc kubenswrapper[4860]: > Dec 11 08:56:14 crc kubenswrapper[4860]: I1211 08:56:14.414895 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" event={"ID":"31c30642-6e60-41b4-a477-0d802424e0aa","Type":"ContainerStarted","Data":"a784a9513182b229e3bc2bf76a6f95c5e4bcc9f5b196234f21cfe4fd910a02e2"} Dec 11 08:56:23 crc kubenswrapper[4860]: I1211 08:56:23.179343 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-ndvvj" Dec 11 08:56:23 crc kubenswrapper[4860]: I1211 08:56:23.237131 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-ndvvj" Dec 11 08:56:23 crc kubenswrapper[4860]: I1211 08:56:23.417520 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-ndvvj"] Dec 11 08:56:24 crc kubenswrapper[4860]: I1211 08:56:24.519496 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-ndvvj" podUID="e7ac2ba0-677d-45e5-8842-c64b96c8fed6" containerName="registry-server" containerID="cri-o://138aee03b6c0ca7708a2ad0a9bd763d1586f59878c7543251e484377824bd308" gracePeriod=2 Dec 11 08:56:25 crc kubenswrapper[4860]: I1211 08:56:25.534701 4860 generic.go:334] "Generic (PLEG): container finished" podID="e7ac2ba0-677d-45e5-8842-c64b96c8fed6" containerID="138aee03b6c0ca7708a2ad0a9bd763d1586f59878c7543251e484377824bd308" exitCode=0 Dec 11 08:56:25 crc kubenswrapper[4860]: I1211 08:56:25.535010 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ndvvj" event={"ID":"e7ac2ba0-677d-45e5-8842-c64b96c8fed6","Type":"ContainerDied","Data":"138aee03b6c0ca7708a2ad0a9bd763d1586f59878c7543251e484377824bd308"} Dec 11 08:56:25 crc kubenswrapper[4860]: I1211 08:56:25.708004 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ndvvj" Dec 11 08:56:25 crc kubenswrapper[4860]: I1211 08:56:25.849802 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7ac2ba0-677d-45e5-8842-c64b96c8fed6-utilities\") pod \"e7ac2ba0-677d-45e5-8842-c64b96c8fed6\" (UID: \"e7ac2ba0-677d-45e5-8842-c64b96c8fed6\") " Dec 11 08:56:25 crc kubenswrapper[4860]: I1211 08:56:25.850192 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7ac2ba0-677d-45e5-8842-c64b96c8fed6-catalog-content\") pod \"e7ac2ba0-677d-45e5-8842-c64b96c8fed6\" (UID: \"e7ac2ba0-677d-45e5-8842-c64b96c8fed6\") " Dec 11 08:56:25 crc kubenswrapper[4860]: I1211 08:56:25.850243 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hls6l\" (UniqueName: \"kubernetes.io/projected/e7ac2ba0-677d-45e5-8842-c64b96c8fed6-kube-api-access-hls6l\") pod \"e7ac2ba0-677d-45e5-8842-c64b96c8fed6\" (UID: \"e7ac2ba0-677d-45e5-8842-c64b96c8fed6\") " Dec 11 08:56:25 crc kubenswrapper[4860]: I1211 08:56:25.851719 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e7ac2ba0-677d-45e5-8842-c64b96c8fed6-utilities" (OuterVolumeSpecName: "utilities") pod "e7ac2ba0-677d-45e5-8842-c64b96c8fed6" (UID: "e7ac2ba0-677d-45e5-8842-c64b96c8fed6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:56:25 crc kubenswrapper[4860]: I1211 08:56:25.856030 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7ac2ba0-677d-45e5-8842-c64b96c8fed6-kube-api-access-hls6l" (OuterVolumeSpecName: "kube-api-access-hls6l") pod "e7ac2ba0-677d-45e5-8842-c64b96c8fed6" (UID: "e7ac2ba0-677d-45e5-8842-c64b96c8fed6"). InnerVolumeSpecName "kube-api-access-hls6l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:56:25 crc kubenswrapper[4860]: I1211 08:56:25.955295 4860 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7ac2ba0-677d-45e5-8842-c64b96c8fed6-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 08:56:25 crc kubenswrapper[4860]: I1211 08:56:25.955341 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hls6l\" (UniqueName: \"kubernetes.io/projected/e7ac2ba0-677d-45e5-8842-c64b96c8fed6-kube-api-access-hls6l\") on node \"crc\" DevicePath \"\"" Dec 11 08:56:26 crc kubenswrapper[4860]: I1211 08:56:26.000831 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e7ac2ba0-677d-45e5-8842-c64b96c8fed6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e7ac2ba0-677d-45e5-8842-c64b96c8fed6" (UID: "e7ac2ba0-677d-45e5-8842-c64b96c8fed6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:56:26 crc kubenswrapper[4860]: I1211 08:56:26.056915 4860 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7ac2ba0-677d-45e5-8842-c64b96c8fed6-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 08:56:26 crc kubenswrapper[4860]: I1211 08:56:26.547255 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ndvvj" event={"ID":"e7ac2ba0-677d-45e5-8842-c64b96c8fed6","Type":"ContainerDied","Data":"0498de5835002fa9c7a9b53fd54e14c2836ebe7c5841b740d6c9b0a1e1562a8f"} Dec 11 08:56:26 crc kubenswrapper[4860]: I1211 08:56:26.547351 4860 scope.go:117] "RemoveContainer" containerID="138aee03b6c0ca7708a2ad0a9bd763d1586f59878c7543251e484377824bd308" Dec 11 08:56:26 crc kubenswrapper[4860]: I1211 08:56:26.547394 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ndvvj" Dec 11 08:56:26 crc kubenswrapper[4860]: I1211 08:56:26.577545 4860 scope.go:117] "RemoveContainer" containerID="4a12e21ba3a980ba052f7f68d15536850049694a6bf0aee4b4f8e7ba8f83af91" Dec 11 08:56:26 crc kubenswrapper[4860]: I1211 08:56:26.585366 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-ndvvj"] Dec 11 08:56:26 crc kubenswrapper[4860]: I1211 08:56:26.597045 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-ndvvj"] Dec 11 08:56:26 crc kubenswrapper[4860]: I1211 08:56:26.609445 4860 scope.go:117] "RemoveContainer" containerID="8f72ec4bd21e4d739d48d19d24c83d9e62ae3937d71afb494367aa4158e1e2d1" Dec 11 08:56:27 crc kubenswrapper[4860]: I1211 08:56:27.615982 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7ac2ba0-677d-45e5-8842-c64b96c8fed6" path="/var/lib/kubelet/pods/e7ac2ba0-677d-45e5-8842-c64b96c8fed6/volumes" Dec 11 08:57:23 crc kubenswrapper[4860]: I1211 08:57:23.305536 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-m49qj"] Dec 11 08:57:23 crc kubenswrapper[4860]: E1211 08:57:23.306554 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7ac2ba0-677d-45e5-8842-c64b96c8fed6" containerName="extract-utilities" Dec 11 08:57:23 crc kubenswrapper[4860]: I1211 08:57:23.306572 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7ac2ba0-677d-45e5-8842-c64b96c8fed6" containerName="extract-utilities" Dec 11 08:57:23 crc kubenswrapper[4860]: E1211 08:57:23.306593 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7ac2ba0-677d-45e5-8842-c64b96c8fed6" containerName="extract-content" Dec 11 08:57:23 crc kubenswrapper[4860]: I1211 08:57:23.306600 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7ac2ba0-677d-45e5-8842-c64b96c8fed6" containerName="extract-content" Dec 11 08:57:23 crc kubenswrapper[4860]: E1211 08:57:23.306623 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7ac2ba0-677d-45e5-8842-c64b96c8fed6" containerName="registry-server" Dec 11 08:57:23 crc kubenswrapper[4860]: I1211 08:57:23.306631 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7ac2ba0-677d-45e5-8842-c64b96c8fed6" containerName="registry-server" Dec 11 08:57:23 crc kubenswrapper[4860]: I1211 08:57:23.307266 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="e7ac2ba0-677d-45e5-8842-c64b96c8fed6" containerName="registry-server" Dec 11 08:57:23 crc kubenswrapper[4860]: I1211 08:57:23.308712 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-m49qj" Dec 11 08:57:23 crc kubenswrapper[4860]: I1211 08:57:23.322477 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-m49qj"] Dec 11 08:57:23 crc kubenswrapper[4860]: I1211 08:57:23.473531 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8be3ab82-fe51-4b58-98fd-4197112aa6bf-catalog-content\") pod \"redhat-marketplace-m49qj\" (UID: \"8be3ab82-fe51-4b58-98fd-4197112aa6bf\") " pod="openshift-marketplace/redhat-marketplace-m49qj" Dec 11 08:57:23 crc kubenswrapper[4860]: I1211 08:57:23.473574 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v8mtp\" (UniqueName: \"kubernetes.io/projected/8be3ab82-fe51-4b58-98fd-4197112aa6bf-kube-api-access-v8mtp\") pod \"redhat-marketplace-m49qj\" (UID: \"8be3ab82-fe51-4b58-98fd-4197112aa6bf\") " pod="openshift-marketplace/redhat-marketplace-m49qj" Dec 11 08:57:23 crc kubenswrapper[4860]: I1211 08:57:23.473603 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8be3ab82-fe51-4b58-98fd-4197112aa6bf-utilities\") pod \"redhat-marketplace-m49qj\" (UID: \"8be3ab82-fe51-4b58-98fd-4197112aa6bf\") " pod="openshift-marketplace/redhat-marketplace-m49qj" Dec 11 08:57:23 crc kubenswrapper[4860]: I1211 08:57:23.575701 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8be3ab82-fe51-4b58-98fd-4197112aa6bf-catalog-content\") pod \"redhat-marketplace-m49qj\" (UID: \"8be3ab82-fe51-4b58-98fd-4197112aa6bf\") " pod="openshift-marketplace/redhat-marketplace-m49qj" Dec 11 08:57:23 crc kubenswrapper[4860]: I1211 08:57:23.575753 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v8mtp\" (UniqueName: \"kubernetes.io/projected/8be3ab82-fe51-4b58-98fd-4197112aa6bf-kube-api-access-v8mtp\") pod \"redhat-marketplace-m49qj\" (UID: \"8be3ab82-fe51-4b58-98fd-4197112aa6bf\") " pod="openshift-marketplace/redhat-marketplace-m49qj" Dec 11 08:57:23 crc kubenswrapper[4860]: I1211 08:57:23.575781 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8be3ab82-fe51-4b58-98fd-4197112aa6bf-utilities\") pod \"redhat-marketplace-m49qj\" (UID: \"8be3ab82-fe51-4b58-98fd-4197112aa6bf\") " pod="openshift-marketplace/redhat-marketplace-m49qj" Dec 11 08:57:23 crc kubenswrapper[4860]: I1211 08:57:23.576322 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8be3ab82-fe51-4b58-98fd-4197112aa6bf-catalog-content\") pod \"redhat-marketplace-m49qj\" (UID: \"8be3ab82-fe51-4b58-98fd-4197112aa6bf\") " pod="openshift-marketplace/redhat-marketplace-m49qj" Dec 11 08:57:23 crc kubenswrapper[4860]: I1211 08:57:23.576417 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8be3ab82-fe51-4b58-98fd-4197112aa6bf-utilities\") pod \"redhat-marketplace-m49qj\" (UID: \"8be3ab82-fe51-4b58-98fd-4197112aa6bf\") " pod="openshift-marketplace/redhat-marketplace-m49qj" Dec 11 08:57:23 crc kubenswrapper[4860]: I1211 08:57:23.610435 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v8mtp\" (UniqueName: \"kubernetes.io/projected/8be3ab82-fe51-4b58-98fd-4197112aa6bf-kube-api-access-v8mtp\") pod \"redhat-marketplace-m49qj\" (UID: \"8be3ab82-fe51-4b58-98fd-4197112aa6bf\") " pod="openshift-marketplace/redhat-marketplace-m49qj" Dec 11 08:57:23 crc kubenswrapper[4860]: I1211 08:57:23.632145 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-m49qj" Dec 11 08:57:24 crc kubenswrapper[4860]: I1211 08:57:24.089826 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-m49qj"] Dec 11 08:57:25 crc kubenswrapper[4860]: I1211 08:57:25.065132 4860 generic.go:334] "Generic (PLEG): container finished" podID="8be3ab82-fe51-4b58-98fd-4197112aa6bf" containerID="08692b38da4f3c76672c1444d99b565bc18b5a6d7be529b9cbca9e2b50924ebe" exitCode=0 Dec 11 08:57:25 crc kubenswrapper[4860]: I1211 08:57:25.065196 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m49qj" event={"ID":"8be3ab82-fe51-4b58-98fd-4197112aa6bf","Type":"ContainerDied","Data":"08692b38da4f3c76672c1444d99b565bc18b5a6d7be529b9cbca9e2b50924ebe"} Dec 11 08:57:25 crc kubenswrapper[4860]: I1211 08:57:25.065438 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m49qj" event={"ID":"8be3ab82-fe51-4b58-98fd-4197112aa6bf","Type":"ContainerStarted","Data":"27becae4c3a4ae267d42b2c89b6a105235cc0411e04ec4cfa9f7dae033078678"} Dec 11 08:57:25 crc kubenswrapper[4860]: I1211 08:57:25.068337 4860 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 11 08:57:27 crc kubenswrapper[4860]: I1211 08:57:27.086312 4860 generic.go:334] "Generic (PLEG): container finished" podID="8be3ab82-fe51-4b58-98fd-4197112aa6bf" containerID="99531bd2d79c756f3e18aa752f0b652936fe764848f745a48ef73f40f473ff3f" exitCode=0 Dec 11 08:57:27 crc kubenswrapper[4860]: I1211 08:57:27.086387 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m49qj" event={"ID":"8be3ab82-fe51-4b58-98fd-4197112aa6bf","Type":"ContainerDied","Data":"99531bd2d79c756f3e18aa752f0b652936fe764848f745a48ef73f40f473ff3f"} Dec 11 08:57:30 crc kubenswrapper[4860]: I1211 08:57:30.114127 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m49qj" event={"ID":"8be3ab82-fe51-4b58-98fd-4197112aa6bf","Type":"ContainerStarted","Data":"09f59b27f4a3a61a4ec46ac8cd751cda9178c103d169c1dc20b9dbc3e5adeba2"} Dec 11 08:57:30 crc kubenswrapper[4860]: I1211 08:57:30.137376 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-m49qj" podStartSLOduration=3.265096148 podStartE2EDuration="7.137357278s" podCreationTimestamp="2025-12-11 08:57:23 +0000 UTC" firstStartedPulling="2025-12-11 08:57:25.068100163 +0000 UTC m=+2777.796619218" lastFinishedPulling="2025-12-11 08:57:28.940361293 +0000 UTC m=+2781.668880348" observedRunningTime="2025-12-11 08:57:30.135589074 +0000 UTC m=+2782.864108149" watchObservedRunningTime="2025-12-11 08:57:30.137357278 +0000 UTC m=+2782.865876333" Dec 11 08:57:33 crc kubenswrapper[4860]: I1211 08:57:33.633039 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-m49qj" Dec 11 08:57:33 crc kubenswrapper[4860]: I1211 08:57:33.633285 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-m49qj" Dec 11 08:57:33 crc kubenswrapper[4860]: I1211 08:57:33.682342 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-m49qj" Dec 11 08:57:34 crc kubenswrapper[4860]: I1211 08:57:34.197360 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-m49qj" Dec 11 08:57:34 crc kubenswrapper[4860]: I1211 08:57:34.247978 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-m49qj"] Dec 11 08:57:36 crc kubenswrapper[4860]: I1211 08:57:36.162856 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-m49qj" podUID="8be3ab82-fe51-4b58-98fd-4197112aa6bf" containerName="registry-server" containerID="cri-o://09f59b27f4a3a61a4ec46ac8cd751cda9178c103d169c1dc20b9dbc3e5adeba2" gracePeriod=2 Dec 11 08:57:36 crc kubenswrapper[4860]: I1211 08:57:36.772687 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-m49qj" Dec 11 08:57:36 crc kubenswrapper[4860]: I1211 08:57:36.847552 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8be3ab82-fe51-4b58-98fd-4197112aa6bf-utilities\") pod \"8be3ab82-fe51-4b58-98fd-4197112aa6bf\" (UID: \"8be3ab82-fe51-4b58-98fd-4197112aa6bf\") " Dec 11 08:57:36 crc kubenswrapper[4860]: I1211 08:57:36.848085 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8be3ab82-fe51-4b58-98fd-4197112aa6bf-catalog-content\") pod \"8be3ab82-fe51-4b58-98fd-4197112aa6bf\" (UID: \"8be3ab82-fe51-4b58-98fd-4197112aa6bf\") " Dec 11 08:57:36 crc kubenswrapper[4860]: I1211 08:57:36.848137 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v8mtp\" (UniqueName: \"kubernetes.io/projected/8be3ab82-fe51-4b58-98fd-4197112aa6bf-kube-api-access-v8mtp\") pod \"8be3ab82-fe51-4b58-98fd-4197112aa6bf\" (UID: \"8be3ab82-fe51-4b58-98fd-4197112aa6bf\") " Dec 11 08:57:36 crc kubenswrapper[4860]: I1211 08:57:36.848925 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8be3ab82-fe51-4b58-98fd-4197112aa6bf-utilities" (OuterVolumeSpecName: "utilities") pod "8be3ab82-fe51-4b58-98fd-4197112aa6bf" (UID: "8be3ab82-fe51-4b58-98fd-4197112aa6bf"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:57:36 crc kubenswrapper[4860]: I1211 08:57:36.864905 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8be3ab82-fe51-4b58-98fd-4197112aa6bf-kube-api-access-v8mtp" (OuterVolumeSpecName: "kube-api-access-v8mtp") pod "8be3ab82-fe51-4b58-98fd-4197112aa6bf" (UID: "8be3ab82-fe51-4b58-98fd-4197112aa6bf"). InnerVolumeSpecName "kube-api-access-v8mtp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 08:57:36 crc kubenswrapper[4860]: I1211 08:57:36.884354 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8be3ab82-fe51-4b58-98fd-4197112aa6bf-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8be3ab82-fe51-4b58-98fd-4197112aa6bf" (UID: "8be3ab82-fe51-4b58-98fd-4197112aa6bf"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 08:57:36 crc kubenswrapper[4860]: I1211 08:57:36.951571 4860 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8be3ab82-fe51-4b58-98fd-4197112aa6bf-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 08:57:36 crc kubenswrapper[4860]: I1211 08:57:36.951637 4860 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8be3ab82-fe51-4b58-98fd-4197112aa6bf-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 08:57:36 crc kubenswrapper[4860]: I1211 08:57:36.951681 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v8mtp\" (UniqueName: \"kubernetes.io/projected/8be3ab82-fe51-4b58-98fd-4197112aa6bf-kube-api-access-v8mtp\") on node \"crc\" DevicePath \"\"" Dec 11 08:57:37 crc kubenswrapper[4860]: I1211 08:57:37.173312 4860 generic.go:334] "Generic (PLEG): container finished" podID="8be3ab82-fe51-4b58-98fd-4197112aa6bf" containerID="09f59b27f4a3a61a4ec46ac8cd751cda9178c103d169c1dc20b9dbc3e5adeba2" exitCode=0 Dec 11 08:57:37 crc kubenswrapper[4860]: I1211 08:57:37.173366 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m49qj" event={"ID":"8be3ab82-fe51-4b58-98fd-4197112aa6bf","Type":"ContainerDied","Data":"09f59b27f4a3a61a4ec46ac8cd751cda9178c103d169c1dc20b9dbc3e5adeba2"} Dec 11 08:57:37 crc kubenswrapper[4860]: I1211 08:57:37.173399 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-m49qj" Dec 11 08:57:37 crc kubenswrapper[4860]: I1211 08:57:37.173419 4860 scope.go:117] "RemoveContainer" containerID="09f59b27f4a3a61a4ec46ac8cd751cda9178c103d169c1dc20b9dbc3e5adeba2" Dec 11 08:57:37 crc kubenswrapper[4860]: I1211 08:57:37.173406 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m49qj" event={"ID":"8be3ab82-fe51-4b58-98fd-4197112aa6bf","Type":"ContainerDied","Data":"27becae4c3a4ae267d42b2c89b6a105235cc0411e04ec4cfa9f7dae033078678"} Dec 11 08:57:37 crc kubenswrapper[4860]: I1211 08:57:37.197494 4860 scope.go:117] "RemoveContainer" containerID="99531bd2d79c756f3e18aa752f0b652936fe764848f745a48ef73f40f473ff3f" Dec 11 08:57:37 crc kubenswrapper[4860]: I1211 08:57:37.210925 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-m49qj"] Dec 11 08:57:37 crc kubenswrapper[4860]: I1211 08:57:37.220314 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-m49qj"] Dec 11 08:57:37 crc kubenswrapper[4860]: I1211 08:57:37.231345 4860 scope.go:117] "RemoveContainer" containerID="08692b38da4f3c76672c1444d99b565bc18b5a6d7be529b9cbca9e2b50924ebe" Dec 11 08:57:37 crc kubenswrapper[4860]: I1211 08:57:37.265263 4860 scope.go:117] "RemoveContainer" containerID="09f59b27f4a3a61a4ec46ac8cd751cda9178c103d169c1dc20b9dbc3e5adeba2" Dec 11 08:57:37 crc kubenswrapper[4860]: E1211 08:57:37.265717 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"09f59b27f4a3a61a4ec46ac8cd751cda9178c103d169c1dc20b9dbc3e5adeba2\": container with ID starting with 09f59b27f4a3a61a4ec46ac8cd751cda9178c103d169c1dc20b9dbc3e5adeba2 not found: ID does not exist" containerID="09f59b27f4a3a61a4ec46ac8cd751cda9178c103d169c1dc20b9dbc3e5adeba2" Dec 11 08:57:37 crc kubenswrapper[4860]: I1211 08:57:37.265751 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"09f59b27f4a3a61a4ec46ac8cd751cda9178c103d169c1dc20b9dbc3e5adeba2"} err="failed to get container status \"09f59b27f4a3a61a4ec46ac8cd751cda9178c103d169c1dc20b9dbc3e5adeba2\": rpc error: code = NotFound desc = could not find container \"09f59b27f4a3a61a4ec46ac8cd751cda9178c103d169c1dc20b9dbc3e5adeba2\": container with ID starting with 09f59b27f4a3a61a4ec46ac8cd751cda9178c103d169c1dc20b9dbc3e5adeba2 not found: ID does not exist" Dec 11 08:57:37 crc kubenswrapper[4860]: I1211 08:57:37.265788 4860 scope.go:117] "RemoveContainer" containerID="99531bd2d79c756f3e18aa752f0b652936fe764848f745a48ef73f40f473ff3f" Dec 11 08:57:37 crc kubenswrapper[4860]: E1211 08:57:37.266102 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"99531bd2d79c756f3e18aa752f0b652936fe764848f745a48ef73f40f473ff3f\": container with ID starting with 99531bd2d79c756f3e18aa752f0b652936fe764848f745a48ef73f40f473ff3f not found: ID does not exist" containerID="99531bd2d79c756f3e18aa752f0b652936fe764848f745a48ef73f40f473ff3f" Dec 11 08:57:37 crc kubenswrapper[4860]: I1211 08:57:37.266154 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"99531bd2d79c756f3e18aa752f0b652936fe764848f745a48ef73f40f473ff3f"} err="failed to get container status \"99531bd2d79c756f3e18aa752f0b652936fe764848f745a48ef73f40f473ff3f\": rpc error: code = NotFound desc = could not find container \"99531bd2d79c756f3e18aa752f0b652936fe764848f745a48ef73f40f473ff3f\": container with ID starting with 99531bd2d79c756f3e18aa752f0b652936fe764848f745a48ef73f40f473ff3f not found: ID does not exist" Dec 11 08:57:37 crc kubenswrapper[4860]: I1211 08:57:37.266173 4860 scope.go:117] "RemoveContainer" containerID="08692b38da4f3c76672c1444d99b565bc18b5a6d7be529b9cbca9e2b50924ebe" Dec 11 08:57:37 crc kubenswrapper[4860]: E1211 08:57:37.266420 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"08692b38da4f3c76672c1444d99b565bc18b5a6d7be529b9cbca9e2b50924ebe\": container with ID starting with 08692b38da4f3c76672c1444d99b565bc18b5a6d7be529b9cbca9e2b50924ebe not found: ID does not exist" containerID="08692b38da4f3c76672c1444d99b565bc18b5a6d7be529b9cbca9e2b50924ebe" Dec 11 08:57:37 crc kubenswrapper[4860]: I1211 08:57:37.266446 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"08692b38da4f3c76672c1444d99b565bc18b5a6d7be529b9cbca9e2b50924ebe"} err="failed to get container status \"08692b38da4f3c76672c1444d99b565bc18b5a6d7be529b9cbca9e2b50924ebe\": rpc error: code = NotFound desc = could not find container \"08692b38da4f3c76672c1444d99b565bc18b5a6d7be529b9cbca9e2b50924ebe\": container with ID starting with 08692b38da4f3c76672c1444d99b565bc18b5a6d7be529b9cbca9e2b50924ebe not found: ID does not exist" Dec 11 08:57:37 crc kubenswrapper[4860]: I1211 08:57:37.611626 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8be3ab82-fe51-4b58-98fd-4197112aa6bf" path="/var/lib/kubelet/pods/8be3ab82-fe51-4b58-98fd-4197112aa6bf/volumes" Dec 11 08:58:38 crc kubenswrapper[4860]: I1211 08:58:38.795032 4860 patch_prober.go:28] interesting pod/machine-config-daemon-99qgp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 08:58:38 crc kubenswrapper[4860]: I1211 08:58:38.795699 4860 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 08:59:08 crc kubenswrapper[4860]: I1211 08:59:08.795630 4860 patch_prober.go:28] interesting pod/machine-config-daemon-99qgp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 08:59:08 crc kubenswrapper[4860]: I1211 08:59:08.796224 4860 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 08:59:38 crc kubenswrapper[4860]: I1211 08:59:38.795017 4860 patch_prober.go:28] interesting pod/machine-config-daemon-99qgp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 08:59:38 crc kubenswrapper[4860]: I1211 08:59:38.795618 4860 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 08:59:38 crc kubenswrapper[4860]: I1211 08:59:38.795690 4860 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" Dec 11 08:59:38 crc kubenswrapper[4860]: I1211 08:59:38.796480 4860 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a784a9513182b229e3bc2bf76a6f95c5e4bcc9f5b196234f21cfe4fd910a02e2"} pod="openshift-machine-config-operator/machine-config-daemon-99qgp" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 11 08:59:38 crc kubenswrapper[4860]: I1211 08:59:38.796538 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" containerName="machine-config-daemon" containerID="cri-o://a784a9513182b229e3bc2bf76a6f95c5e4bcc9f5b196234f21cfe4fd910a02e2" gracePeriod=600 Dec 11 08:59:40 crc kubenswrapper[4860]: I1211 08:59:40.295593 4860 generic.go:334] "Generic (PLEG): container finished" podID="31c30642-6e60-41b4-a477-0d802424e0aa" containerID="a784a9513182b229e3bc2bf76a6f95c5e4bcc9f5b196234f21cfe4fd910a02e2" exitCode=0 Dec 11 08:59:40 crc kubenswrapper[4860]: I1211 08:59:40.295628 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" event={"ID":"31c30642-6e60-41b4-a477-0d802424e0aa","Type":"ContainerDied","Data":"a784a9513182b229e3bc2bf76a6f95c5e4bcc9f5b196234f21cfe4fd910a02e2"} Dec 11 08:59:40 crc kubenswrapper[4860]: I1211 08:59:40.296169 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" event={"ID":"31c30642-6e60-41b4-a477-0d802424e0aa","Type":"ContainerStarted","Data":"76665d219a809fd146709b7837eb798329ce7fb1c53d5b15c247e7e654985e07"} Dec 11 08:59:40 crc kubenswrapper[4860]: I1211 08:59:40.296192 4860 scope.go:117] "RemoveContainer" containerID="b747bbead8584985306f457adcbaf89164c99c29fe92a3c021617afef9e04489" Dec 11 09:00:00 crc kubenswrapper[4860]: I1211 09:00:00.186734 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29424060-4dmn8"] Dec 11 09:00:00 crc kubenswrapper[4860]: E1211 09:00:00.187911 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8be3ab82-fe51-4b58-98fd-4197112aa6bf" containerName="extract-content" Dec 11 09:00:00 crc kubenswrapper[4860]: I1211 09:00:00.187932 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="8be3ab82-fe51-4b58-98fd-4197112aa6bf" containerName="extract-content" Dec 11 09:00:00 crc kubenswrapper[4860]: E1211 09:00:00.187957 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8be3ab82-fe51-4b58-98fd-4197112aa6bf" containerName="extract-utilities" Dec 11 09:00:00 crc kubenswrapper[4860]: I1211 09:00:00.187966 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="8be3ab82-fe51-4b58-98fd-4197112aa6bf" containerName="extract-utilities" Dec 11 09:00:00 crc kubenswrapper[4860]: E1211 09:00:00.187974 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8be3ab82-fe51-4b58-98fd-4197112aa6bf" containerName="registry-server" Dec 11 09:00:00 crc kubenswrapper[4860]: I1211 09:00:00.187980 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="8be3ab82-fe51-4b58-98fd-4197112aa6bf" containerName="registry-server" Dec 11 09:00:00 crc kubenswrapper[4860]: I1211 09:00:00.188294 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="8be3ab82-fe51-4b58-98fd-4197112aa6bf" containerName="registry-server" Dec 11 09:00:00 crc kubenswrapper[4860]: I1211 09:00:00.189469 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29424060-4dmn8" Dec 11 09:00:00 crc kubenswrapper[4860]: I1211 09:00:00.197686 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 11 09:00:00 crc kubenswrapper[4860]: I1211 09:00:00.197992 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 11 09:00:00 crc kubenswrapper[4860]: I1211 09:00:00.199517 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29424060-4dmn8"] Dec 11 09:00:00 crc kubenswrapper[4860]: I1211 09:00:00.325451 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vfh4z\" (UniqueName: \"kubernetes.io/projected/548c8c0d-b078-4304-b561-e04d80a5806a-kube-api-access-vfh4z\") pod \"collect-profiles-29424060-4dmn8\" (UID: \"548c8c0d-b078-4304-b561-e04d80a5806a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424060-4dmn8" Dec 11 09:00:00 crc kubenswrapper[4860]: I1211 09:00:00.325548 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/548c8c0d-b078-4304-b561-e04d80a5806a-config-volume\") pod \"collect-profiles-29424060-4dmn8\" (UID: \"548c8c0d-b078-4304-b561-e04d80a5806a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424060-4dmn8" Dec 11 09:00:00 crc kubenswrapper[4860]: I1211 09:00:00.325922 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/548c8c0d-b078-4304-b561-e04d80a5806a-secret-volume\") pod \"collect-profiles-29424060-4dmn8\" (UID: \"548c8c0d-b078-4304-b561-e04d80a5806a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424060-4dmn8" Dec 11 09:00:00 crc kubenswrapper[4860]: I1211 09:00:00.427693 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/548c8c0d-b078-4304-b561-e04d80a5806a-secret-volume\") pod \"collect-profiles-29424060-4dmn8\" (UID: \"548c8c0d-b078-4304-b561-e04d80a5806a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424060-4dmn8" Dec 11 09:00:00 crc kubenswrapper[4860]: I1211 09:00:00.427765 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vfh4z\" (UniqueName: \"kubernetes.io/projected/548c8c0d-b078-4304-b561-e04d80a5806a-kube-api-access-vfh4z\") pod \"collect-profiles-29424060-4dmn8\" (UID: \"548c8c0d-b078-4304-b561-e04d80a5806a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424060-4dmn8" Dec 11 09:00:00 crc kubenswrapper[4860]: I1211 09:00:00.427803 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/548c8c0d-b078-4304-b561-e04d80a5806a-config-volume\") pod \"collect-profiles-29424060-4dmn8\" (UID: \"548c8c0d-b078-4304-b561-e04d80a5806a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424060-4dmn8" Dec 11 09:00:00 crc kubenswrapper[4860]: I1211 09:00:00.428667 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/548c8c0d-b078-4304-b561-e04d80a5806a-config-volume\") pod \"collect-profiles-29424060-4dmn8\" (UID: \"548c8c0d-b078-4304-b561-e04d80a5806a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424060-4dmn8" Dec 11 09:00:00 crc kubenswrapper[4860]: I1211 09:00:00.434868 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/548c8c0d-b078-4304-b561-e04d80a5806a-secret-volume\") pod \"collect-profiles-29424060-4dmn8\" (UID: \"548c8c0d-b078-4304-b561-e04d80a5806a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424060-4dmn8" Dec 11 09:00:00 crc kubenswrapper[4860]: I1211 09:00:00.445915 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vfh4z\" (UniqueName: \"kubernetes.io/projected/548c8c0d-b078-4304-b561-e04d80a5806a-kube-api-access-vfh4z\") pod \"collect-profiles-29424060-4dmn8\" (UID: \"548c8c0d-b078-4304-b561-e04d80a5806a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424060-4dmn8" Dec 11 09:00:00 crc kubenswrapper[4860]: I1211 09:00:00.513989 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29424060-4dmn8" Dec 11 09:00:01 crc kubenswrapper[4860]: I1211 09:00:01.002148 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29424060-4dmn8"] Dec 11 09:00:01 crc kubenswrapper[4860]: I1211 09:00:01.493794 4860 generic.go:334] "Generic (PLEG): container finished" podID="548c8c0d-b078-4304-b561-e04d80a5806a" containerID="0cb7f3f2bb92fc746087152c145db2551c33d0fd31140629ca8cc9868079093c" exitCode=0 Dec 11 09:00:01 crc kubenswrapper[4860]: I1211 09:00:01.493852 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29424060-4dmn8" event={"ID":"548c8c0d-b078-4304-b561-e04d80a5806a","Type":"ContainerDied","Data":"0cb7f3f2bb92fc746087152c145db2551c33d0fd31140629ca8cc9868079093c"} Dec 11 09:00:01 crc kubenswrapper[4860]: I1211 09:00:01.494148 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29424060-4dmn8" event={"ID":"548c8c0d-b078-4304-b561-e04d80a5806a","Type":"ContainerStarted","Data":"583bdbbe4ff0357f329f13e31203e6f544cbc08a5cf23bd9c19376c63eb2cda3"} Dec 11 09:00:02 crc kubenswrapper[4860]: I1211 09:00:02.855524 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29424060-4dmn8" Dec 11 09:00:02 crc kubenswrapper[4860]: I1211 09:00:02.993138 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vfh4z\" (UniqueName: \"kubernetes.io/projected/548c8c0d-b078-4304-b561-e04d80a5806a-kube-api-access-vfh4z\") pod \"548c8c0d-b078-4304-b561-e04d80a5806a\" (UID: \"548c8c0d-b078-4304-b561-e04d80a5806a\") " Dec 11 09:00:02 crc kubenswrapper[4860]: I1211 09:00:02.993490 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/548c8c0d-b078-4304-b561-e04d80a5806a-secret-volume\") pod \"548c8c0d-b078-4304-b561-e04d80a5806a\" (UID: \"548c8c0d-b078-4304-b561-e04d80a5806a\") " Dec 11 09:00:02 crc kubenswrapper[4860]: I1211 09:00:02.993585 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/548c8c0d-b078-4304-b561-e04d80a5806a-config-volume\") pod \"548c8c0d-b078-4304-b561-e04d80a5806a\" (UID: \"548c8c0d-b078-4304-b561-e04d80a5806a\") " Dec 11 09:00:02 crc kubenswrapper[4860]: I1211 09:00:02.994010 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/548c8c0d-b078-4304-b561-e04d80a5806a-config-volume" (OuterVolumeSpecName: "config-volume") pod "548c8c0d-b078-4304-b561-e04d80a5806a" (UID: "548c8c0d-b078-4304-b561-e04d80a5806a"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 09:00:02 crc kubenswrapper[4860]: I1211 09:00:02.994170 4860 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/548c8c0d-b078-4304-b561-e04d80a5806a-config-volume\") on node \"crc\" DevicePath \"\"" Dec 11 09:00:03 crc kubenswrapper[4860]: I1211 09:00:03.000871 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/548c8c0d-b078-4304-b561-e04d80a5806a-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "548c8c0d-b078-4304-b561-e04d80a5806a" (UID: "548c8c0d-b078-4304-b561-e04d80a5806a"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 09:00:03 crc kubenswrapper[4860]: I1211 09:00:03.003891 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/548c8c0d-b078-4304-b561-e04d80a5806a-kube-api-access-vfh4z" (OuterVolumeSpecName: "kube-api-access-vfh4z") pod "548c8c0d-b078-4304-b561-e04d80a5806a" (UID: "548c8c0d-b078-4304-b561-e04d80a5806a"). InnerVolumeSpecName "kube-api-access-vfh4z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 09:00:03 crc kubenswrapper[4860]: I1211 09:00:03.098347 4860 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/548c8c0d-b078-4304-b561-e04d80a5806a-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 11 09:00:03 crc kubenswrapper[4860]: I1211 09:00:03.098401 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vfh4z\" (UniqueName: \"kubernetes.io/projected/548c8c0d-b078-4304-b561-e04d80a5806a-kube-api-access-vfh4z\") on node \"crc\" DevicePath \"\"" Dec 11 09:00:03 crc kubenswrapper[4860]: I1211 09:00:03.534144 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29424060-4dmn8" event={"ID":"548c8c0d-b078-4304-b561-e04d80a5806a","Type":"ContainerDied","Data":"583bdbbe4ff0357f329f13e31203e6f544cbc08a5cf23bd9c19376c63eb2cda3"} Dec 11 09:00:03 crc kubenswrapper[4860]: I1211 09:00:03.534470 4860 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="583bdbbe4ff0357f329f13e31203e6f544cbc08a5cf23bd9c19376c63eb2cda3" Dec 11 09:00:03 crc kubenswrapper[4860]: I1211 09:00:03.534536 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29424060-4dmn8" Dec 11 09:00:03 crc kubenswrapper[4860]: I1211 09:00:03.927304 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29424015-j4cd9"] Dec 11 09:00:03 crc kubenswrapper[4860]: I1211 09:00:03.935921 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29424015-j4cd9"] Dec 11 09:00:05 crc kubenswrapper[4860]: I1211 09:00:05.589457 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a32ff4fb-1fd8-4097-88cf-74c67cf6fa56" path="/var/lib/kubelet/pods/a32ff4fb-1fd8-4097-88cf-74c67cf6fa56/volumes" Dec 11 09:00:33 crc kubenswrapper[4860]: I1211 09:00:33.912791 4860 scope.go:117] "RemoveContainer" containerID="d0289548ad4f995ea757eb89d9adc13ae624e1e3565fb874661e50fd36d76e30" Dec 11 09:01:00 crc kubenswrapper[4860]: I1211 09:01:00.143356 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29424061-gq8tl"] Dec 11 09:01:00 crc kubenswrapper[4860]: E1211 09:01:00.144241 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="548c8c0d-b078-4304-b561-e04d80a5806a" containerName="collect-profiles" Dec 11 09:01:00 crc kubenswrapper[4860]: I1211 09:01:00.144254 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="548c8c0d-b078-4304-b561-e04d80a5806a" containerName="collect-profiles" Dec 11 09:01:00 crc kubenswrapper[4860]: I1211 09:01:00.144452 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="548c8c0d-b078-4304-b561-e04d80a5806a" containerName="collect-profiles" Dec 11 09:01:00 crc kubenswrapper[4860]: I1211 09:01:00.145139 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29424061-gq8tl" Dec 11 09:01:00 crc kubenswrapper[4860]: I1211 09:01:00.158405 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29424061-gq8tl"] Dec 11 09:01:00 crc kubenswrapper[4860]: I1211 09:01:00.363944 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8fb3c7c5-8c06-4e05-ad79-11f0671e16da-config-data\") pod \"keystone-cron-29424061-gq8tl\" (UID: \"8fb3c7c5-8c06-4e05-ad79-11f0671e16da\") " pod="openstack/keystone-cron-29424061-gq8tl" Dec 11 09:01:00 crc kubenswrapper[4860]: I1211 09:01:00.365664 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8fb3c7c5-8c06-4e05-ad79-11f0671e16da-fernet-keys\") pod \"keystone-cron-29424061-gq8tl\" (UID: \"8fb3c7c5-8c06-4e05-ad79-11f0671e16da\") " pod="openstack/keystone-cron-29424061-gq8tl" Dec 11 09:01:00 crc kubenswrapper[4860]: I1211 09:01:00.365811 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8fb3c7c5-8c06-4e05-ad79-11f0671e16da-combined-ca-bundle\") pod \"keystone-cron-29424061-gq8tl\" (UID: \"8fb3c7c5-8c06-4e05-ad79-11f0671e16da\") " pod="openstack/keystone-cron-29424061-gq8tl" Dec 11 09:01:00 crc kubenswrapper[4860]: I1211 09:01:00.365918 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xhkv6\" (UniqueName: \"kubernetes.io/projected/8fb3c7c5-8c06-4e05-ad79-11f0671e16da-kube-api-access-xhkv6\") pod \"keystone-cron-29424061-gq8tl\" (UID: \"8fb3c7c5-8c06-4e05-ad79-11f0671e16da\") " pod="openstack/keystone-cron-29424061-gq8tl" Dec 11 09:01:00 crc kubenswrapper[4860]: I1211 09:01:00.467087 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8fb3c7c5-8c06-4e05-ad79-11f0671e16da-combined-ca-bundle\") pod \"keystone-cron-29424061-gq8tl\" (UID: \"8fb3c7c5-8c06-4e05-ad79-11f0671e16da\") " pod="openstack/keystone-cron-29424061-gq8tl" Dec 11 09:01:00 crc kubenswrapper[4860]: I1211 09:01:00.467433 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xhkv6\" (UniqueName: \"kubernetes.io/projected/8fb3c7c5-8c06-4e05-ad79-11f0671e16da-kube-api-access-xhkv6\") pod \"keystone-cron-29424061-gq8tl\" (UID: \"8fb3c7c5-8c06-4e05-ad79-11f0671e16da\") " pod="openstack/keystone-cron-29424061-gq8tl" Dec 11 09:01:00 crc kubenswrapper[4860]: I1211 09:01:00.467532 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8fb3c7c5-8c06-4e05-ad79-11f0671e16da-config-data\") pod \"keystone-cron-29424061-gq8tl\" (UID: \"8fb3c7c5-8c06-4e05-ad79-11f0671e16da\") " pod="openstack/keystone-cron-29424061-gq8tl" Dec 11 09:01:00 crc kubenswrapper[4860]: I1211 09:01:00.467682 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8fb3c7c5-8c06-4e05-ad79-11f0671e16da-fernet-keys\") pod \"keystone-cron-29424061-gq8tl\" (UID: \"8fb3c7c5-8c06-4e05-ad79-11f0671e16da\") " pod="openstack/keystone-cron-29424061-gq8tl" Dec 11 09:01:00 crc kubenswrapper[4860]: I1211 09:01:00.473609 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8fb3c7c5-8c06-4e05-ad79-11f0671e16da-combined-ca-bundle\") pod \"keystone-cron-29424061-gq8tl\" (UID: \"8fb3c7c5-8c06-4e05-ad79-11f0671e16da\") " pod="openstack/keystone-cron-29424061-gq8tl" Dec 11 09:01:00 crc kubenswrapper[4860]: I1211 09:01:00.473730 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8fb3c7c5-8c06-4e05-ad79-11f0671e16da-fernet-keys\") pod \"keystone-cron-29424061-gq8tl\" (UID: \"8fb3c7c5-8c06-4e05-ad79-11f0671e16da\") " pod="openstack/keystone-cron-29424061-gq8tl" Dec 11 09:01:00 crc kubenswrapper[4860]: I1211 09:01:00.475072 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8fb3c7c5-8c06-4e05-ad79-11f0671e16da-config-data\") pod \"keystone-cron-29424061-gq8tl\" (UID: \"8fb3c7c5-8c06-4e05-ad79-11f0671e16da\") " pod="openstack/keystone-cron-29424061-gq8tl" Dec 11 09:01:00 crc kubenswrapper[4860]: I1211 09:01:00.485224 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xhkv6\" (UniqueName: \"kubernetes.io/projected/8fb3c7c5-8c06-4e05-ad79-11f0671e16da-kube-api-access-xhkv6\") pod \"keystone-cron-29424061-gq8tl\" (UID: \"8fb3c7c5-8c06-4e05-ad79-11f0671e16da\") " pod="openstack/keystone-cron-29424061-gq8tl" Dec 11 09:01:00 crc kubenswrapper[4860]: I1211 09:01:00.574967 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29424061-gq8tl" Dec 11 09:01:01 crc kubenswrapper[4860]: I1211 09:01:01.030911 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29424061-gq8tl"] Dec 11 09:01:01 crc kubenswrapper[4860]: I1211 09:01:01.070845 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29424061-gq8tl" event={"ID":"8fb3c7c5-8c06-4e05-ad79-11f0671e16da","Type":"ContainerStarted","Data":"76db0f50965afd707706771b96e46a4a08c93bc883789d8fc02df1623f3d323a"} Dec 11 09:01:02 crc kubenswrapper[4860]: I1211 09:01:02.080896 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29424061-gq8tl" event={"ID":"8fb3c7c5-8c06-4e05-ad79-11f0671e16da","Type":"ContainerStarted","Data":"6a4686f1c1df782fa096be519581d8a16b1b36943d9c7d57dd536c660cc1540a"} Dec 11 09:01:02 crc kubenswrapper[4860]: I1211 09:01:02.095610 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29424061-gq8tl" podStartSLOduration=2.095591256 podStartE2EDuration="2.095591256s" podCreationTimestamp="2025-12-11 09:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 09:01:02.094889264 +0000 UTC m=+2994.823408319" watchObservedRunningTime="2025-12-11 09:01:02.095591256 +0000 UTC m=+2994.824110311" Dec 11 09:01:04 crc kubenswrapper[4860]: I1211 09:01:04.098201 4860 generic.go:334] "Generic (PLEG): container finished" podID="8fb3c7c5-8c06-4e05-ad79-11f0671e16da" containerID="6a4686f1c1df782fa096be519581d8a16b1b36943d9c7d57dd536c660cc1540a" exitCode=0 Dec 11 09:01:04 crc kubenswrapper[4860]: I1211 09:01:04.098289 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29424061-gq8tl" event={"ID":"8fb3c7c5-8c06-4e05-ad79-11f0671e16da","Type":"ContainerDied","Data":"6a4686f1c1df782fa096be519581d8a16b1b36943d9c7d57dd536c660cc1540a"} Dec 11 09:01:05 crc kubenswrapper[4860]: I1211 09:01:05.465361 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29424061-gq8tl" Dec 11 09:01:05 crc kubenswrapper[4860]: I1211 09:01:05.472485 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8fb3c7c5-8c06-4e05-ad79-11f0671e16da-config-data\") pod \"8fb3c7c5-8c06-4e05-ad79-11f0671e16da\" (UID: \"8fb3c7c5-8c06-4e05-ad79-11f0671e16da\") " Dec 11 09:01:05 crc kubenswrapper[4860]: I1211 09:01:05.472662 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8fb3c7c5-8c06-4e05-ad79-11f0671e16da-fernet-keys\") pod \"8fb3c7c5-8c06-4e05-ad79-11f0671e16da\" (UID: \"8fb3c7c5-8c06-4e05-ad79-11f0671e16da\") " Dec 11 09:01:05 crc kubenswrapper[4860]: I1211 09:01:05.479826 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8fb3c7c5-8c06-4e05-ad79-11f0671e16da-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "8fb3c7c5-8c06-4e05-ad79-11f0671e16da" (UID: "8fb3c7c5-8c06-4e05-ad79-11f0671e16da"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 09:01:05 crc kubenswrapper[4860]: I1211 09:01:05.547009 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8fb3c7c5-8c06-4e05-ad79-11f0671e16da-config-data" (OuterVolumeSpecName: "config-data") pod "8fb3c7c5-8c06-4e05-ad79-11f0671e16da" (UID: "8fb3c7c5-8c06-4e05-ad79-11f0671e16da"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 09:01:05 crc kubenswrapper[4860]: I1211 09:01:05.574476 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xhkv6\" (UniqueName: \"kubernetes.io/projected/8fb3c7c5-8c06-4e05-ad79-11f0671e16da-kube-api-access-xhkv6\") pod \"8fb3c7c5-8c06-4e05-ad79-11f0671e16da\" (UID: \"8fb3c7c5-8c06-4e05-ad79-11f0671e16da\") " Dec 11 09:01:05 crc kubenswrapper[4860]: I1211 09:01:05.574534 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8fb3c7c5-8c06-4e05-ad79-11f0671e16da-combined-ca-bundle\") pod \"8fb3c7c5-8c06-4e05-ad79-11f0671e16da\" (UID: \"8fb3c7c5-8c06-4e05-ad79-11f0671e16da\") " Dec 11 09:01:05 crc kubenswrapper[4860]: I1211 09:01:05.575331 4860 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8fb3c7c5-8c06-4e05-ad79-11f0671e16da-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 09:01:05 crc kubenswrapper[4860]: I1211 09:01:05.575359 4860 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8fb3c7c5-8c06-4e05-ad79-11f0671e16da-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 11 09:01:05 crc kubenswrapper[4860]: I1211 09:01:05.577587 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8fb3c7c5-8c06-4e05-ad79-11f0671e16da-kube-api-access-xhkv6" (OuterVolumeSpecName: "kube-api-access-xhkv6") pod "8fb3c7c5-8c06-4e05-ad79-11f0671e16da" (UID: "8fb3c7c5-8c06-4e05-ad79-11f0671e16da"). InnerVolumeSpecName "kube-api-access-xhkv6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 09:01:05 crc kubenswrapper[4860]: I1211 09:01:05.604114 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8fb3c7c5-8c06-4e05-ad79-11f0671e16da-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8fb3c7c5-8c06-4e05-ad79-11f0671e16da" (UID: "8fb3c7c5-8c06-4e05-ad79-11f0671e16da"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 09:01:05 crc kubenswrapper[4860]: I1211 09:01:05.677708 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xhkv6\" (UniqueName: \"kubernetes.io/projected/8fb3c7c5-8c06-4e05-ad79-11f0671e16da-kube-api-access-xhkv6\") on node \"crc\" DevicePath \"\"" Dec 11 09:01:05 crc kubenswrapper[4860]: I1211 09:01:05.677744 4860 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8fb3c7c5-8c06-4e05-ad79-11f0671e16da-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 09:01:06 crc kubenswrapper[4860]: I1211 09:01:06.118027 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29424061-gq8tl" event={"ID":"8fb3c7c5-8c06-4e05-ad79-11f0671e16da","Type":"ContainerDied","Data":"76db0f50965afd707706771b96e46a4a08c93bc883789d8fc02df1623f3d323a"} Dec 11 09:01:06 crc kubenswrapper[4860]: I1211 09:01:06.118068 4860 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="76db0f50965afd707706771b96e46a4a08c93bc883789d8fc02df1623f3d323a" Dec 11 09:01:06 crc kubenswrapper[4860]: I1211 09:01:06.118136 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29424061-gq8tl" Dec 11 09:01:33 crc kubenswrapper[4860]: I1211 09:01:33.974190 4860 scope.go:117] "RemoveContainer" containerID="ecd79f18a97bd96fb9100822e84a8ad6d15595be73a89217eab927ef489ffc40" Dec 11 09:01:33 crc kubenswrapper[4860]: I1211 09:01:33.995761 4860 scope.go:117] "RemoveContainer" containerID="51bac1372f9d26b98b99f3a545a65774c8fb124c6d5ea2621e53c5ac9570767a" Dec 11 09:01:34 crc kubenswrapper[4860]: I1211 09:01:34.025698 4860 scope.go:117] "RemoveContainer" containerID="10b1058083dc1341da92b2116ff70a327f303e2595dd670f7741ae20dabecace" Dec 11 09:02:08 crc kubenswrapper[4860]: I1211 09:02:08.795759 4860 patch_prober.go:28] interesting pod/machine-config-daemon-99qgp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 09:02:08 crc kubenswrapper[4860]: I1211 09:02:08.796309 4860 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 09:02:38 crc kubenswrapper[4860]: I1211 09:02:38.796276 4860 patch_prober.go:28] interesting pod/machine-config-daemon-99qgp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 09:02:38 crc kubenswrapper[4860]: I1211 09:02:38.796948 4860 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 09:03:08 crc kubenswrapper[4860]: I1211 09:03:08.795811 4860 patch_prober.go:28] interesting pod/machine-config-daemon-99qgp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 09:03:08 crc kubenswrapper[4860]: I1211 09:03:08.796288 4860 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 09:03:08 crc kubenswrapper[4860]: I1211 09:03:08.796339 4860 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" Dec 11 09:03:08 crc kubenswrapper[4860]: I1211 09:03:08.797223 4860 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"76665d219a809fd146709b7837eb798329ce7fb1c53d5b15c247e7e654985e07"} pod="openshift-machine-config-operator/machine-config-daemon-99qgp" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 11 09:03:08 crc kubenswrapper[4860]: I1211 09:03:08.797282 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" containerName="machine-config-daemon" containerID="cri-o://76665d219a809fd146709b7837eb798329ce7fb1c53d5b15c247e7e654985e07" gracePeriod=600 Dec 11 09:03:08 crc kubenswrapper[4860]: E1211 09:03:08.927121 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-99qgp_openshift-machine-config-operator(31c30642-6e60-41b4-a477-0d802424e0aa)\"" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" Dec 11 09:03:09 crc kubenswrapper[4860]: I1211 09:03:09.287891 4860 generic.go:334] "Generic (PLEG): container finished" podID="31c30642-6e60-41b4-a477-0d802424e0aa" containerID="76665d219a809fd146709b7837eb798329ce7fb1c53d5b15c247e7e654985e07" exitCode=0 Dec 11 09:03:09 crc kubenswrapper[4860]: I1211 09:03:09.287961 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" event={"ID":"31c30642-6e60-41b4-a477-0d802424e0aa","Type":"ContainerDied","Data":"76665d219a809fd146709b7837eb798329ce7fb1c53d5b15c247e7e654985e07"} Dec 11 09:03:09 crc kubenswrapper[4860]: I1211 09:03:09.288029 4860 scope.go:117] "RemoveContainer" containerID="a784a9513182b229e3bc2bf76a6f95c5e4bcc9f5b196234f21cfe4fd910a02e2" Dec 11 09:03:09 crc kubenswrapper[4860]: I1211 09:03:09.292109 4860 scope.go:117] "RemoveContainer" containerID="76665d219a809fd146709b7837eb798329ce7fb1c53d5b15c247e7e654985e07" Dec 11 09:03:09 crc kubenswrapper[4860]: E1211 09:03:09.292881 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-99qgp_openshift-machine-config-operator(31c30642-6e60-41b4-a477-0d802424e0aa)\"" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" Dec 11 09:03:20 crc kubenswrapper[4860]: I1211 09:03:20.579094 4860 scope.go:117] "RemoveContainer" containerID="76665d219a809fd146709b7837eb798329ce7fb1c53d5b15c247e7e654985e07" Dec 11 09:03:20 crc kubenswrapper[4860]: E1211 09:03:20.581236 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-99qgp_openshift-machine-config-operator(31c30642-6e60-41b4-a477-0d802424e0aa)\"" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" Dec 11 09:03:32 crc kubenswrapper[4860]: I1211 09:03:32.578836 4860 scope.go:117] "RemoveContainer" containerID="76665d219a809fd146709b7837eb798329ce7fb1c53d5b15c247e7e654985e07" Dec 11 09:03:32 crc kubenswrapper[4860]: E1211 09:03:32.579603 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-99qgp_openshift-machine-config-operator(31c30642-6e60-41b4-a477-0d802424e0aa)\"" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" Dec 11 09:03:44 crc kubenswrapper[4860]: I1211 09:03:44.578820 4860 scope.go:117] "RemoveContainer" containerID="76665d219a809fd146709b7837eb798329ce7fb1c53d5b15c247e7e654985e07" Dec 11 09:03:44 crc kubenswrapper[4860]: E1211 09:03:44.579764 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-99qgp_openshift-machine-config-operator(31c30642-6e60-41b4-a477-0d802424e0aa)\"" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" Dec 11 09:03:59 crc kubenswrapper[4860]: I1211 09:03:59.579070 4860 scope.go:117] "RemoveContainer" containerID="76665d219a809fd146709b7837eb798329ce7fb1c53d5b15c247e7e654985e07" Dec 11 09:03:59 crc kubenswrapper[4860]: E1211 09:03:59.579853 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-99qgp_openshift-machine-config-operator(31c30642-6e60-41b4-a477-0d802424e0aa)\"" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" Dec 11 09:04:11 crc kubenswrapper[4860]: I1211 09:04:11.580042 4860 scope.go:117] "RemoveContainer" containerID="76665d219a809fd146709b7837eb798329ce7fb1c53d5b15c247e7e654985e07" Dec 11 09:04:11 crc kubenswrapper[4860]: E1211 09:04:11.581114 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-99qgp_openshift-machine-config-operator(31c30642-6e60-41b4-a477-0d802424e0aa)\"" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" Dec 11 09:04:22 crc kubenswrapper[4860]: I1211 09:04:22.579070 4860 scope.go:117] "RemoveContainer" containerID="76665d219a809fd146709b7837eb798329ce7fb1c53d5b15c247e7e654985e07" Dec 11 09:04:22 crc kubenswrapper[4860]: E1211 09:04:22.580012 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-99qgp_openshift-machine-config-operator(31c30642-6e60-41b4-a477-0d802424e0aa)\"" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" Dec 11 09:04:34 crc kubenswrapper[4860]: I1211 09:04:34.579084 4860 scope.go:117] "RemoveContainer" containerID="76665d219a809fd146709b7837eb798329ce7fb1c53d5b15c247e7e654985e07" Dec 11 09:04:34 crc kubenswrapper[4860]: E1211 09:04:34.579965 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-99qgp_openshift-machine-config-operator(31c30642-6e60-41b4-a477-0d802424e0aa)\"" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" Dec 11 09:04:49 crc kubenswrapper[4860]: I1211 09:04:49.578957 4860 scope.go:117] "RemoveContainer" containerID="76665d219a809fd146709b7837eb798329ce7fb1c53d5b15c247e7e654985e07" Dec 11 09:04:49 crc kubenswrapper[4860]: E1211 09:04:49.579787 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-99qgp_openshift-machine-config-operator(31c30642-6e60-41b4-a477-0d802424e0aa)\"" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" Dec 11 09:05:03 crc kubenswrapper[4860]: I1211 09:05:03.580583 4860 scope.go:117] "RemoveContainer" containerID="76665d219a809fd146709b7837eb798329ce7fb1c53d5b15c247e7e654985e07" Dec 11 09:05:03 crc kubenswrapper[4860]: E1211 09:05:03.581400 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-99qgp_openshift-machine-config-operator(31c30642-6e60-41b4-a477-0d802424e0aa)\"" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" Dec 11 09:05:15 crc kubenswrapper[4860]: I1211 09:05:15.579022 4860 scope.go:117] "RemoveContainer" containerID="76665d219a809fd146709b7837eb798329ce7fb1c53d5b15c247e7e654985e07" Dec 11 09:05:15 crc kubenswrapper[4860]: E1211 09:05:15.579829 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-99qgp_openshift-machine-config-operator(31c30642-6e60-41b4-a477-0d802424e0aa)\"" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" Dec 11 09:05:28 crc kubenswrapper[4860]: I1211 09:05:28.579547 4860 scope.go:117] "RemoveContainer" containerID="76665d219a809fd146709b7837eb798329ce7fb1c53d5b15c247e7e654985e07" Dec 11 09:05:28 crc kubenswrapper[4860]: E1211 09:05:28.580442 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-99qgp_openshift-machine-config-operator(31c30642-6e60-41b4-a477-0d802424e0aa)\"" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" Dec 11 09:05:40 crc kubenswrapper[4860]: I1211 09:05:40.579512 4860 scope.go:117] "RemoveContainer" containerID="76665d219a809fd146709b7837eb798329ce7fb1c53d5b15c247e7e654985e07" Dec 11 09:05:40 crc kubenswrapper[4860]: E1211 09:05:40.580419 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-99qgp_openshift-machine-config-operator(31c30642-6e60-41b4-a477-0d802424e0aa)\"" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" Dec 11 09:05:54 crc kubenswrapper[4860]: I1211 09:05:54.579221 4860 scope.go:117] "RemoveContainer" containerID="76665d219a809fd146709b7837eb798329ce7fb1c53d5b15c247e7e654985e07" Dec 11 09:05:54 crc kubenswrapper[4860]: E1211 09:05:54.579872 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-99qgp_openshift-machine-config-operator(31c30642-6e60-41b4-a477-0d802424e0aa)\"" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" Dec 11 09:06:06 crc kubenswrapper[4860]: I1211 09:06:06.579203 4860 scope.go:117] "RemoveContainer" containerID="76665d219a809fd146709b7837eb798329ce7fb1c53d5b15c247e7e654985e07" Dec 11 09:06:06 crc kubenswrapper[4860]: E1211 09:06:06.580147 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-99qgp_openshift-machine-config-operator(31c30642-6e60-41b4-a477-0d802424e0aa)\"" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" Dec 11 09:06:21 crc kubenswrapper[4860]: I1211 09:06:21.579675 4860 scope.go:117] "RemoveContainer" containerID="76665d219a809fd146709b7837eb798329ce7fb1c53d5b15c247e7e654985e07" Dec 11 09:06:21 crc kubenswrapper[4860]: E1211 09:06:21.580473 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-99qgp_openshift-machine-config-operator(31c30642-6e60-41b4-a477-0d802424e0aa)\"" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" Dec 11 09:06:24 crc kubenswrapper[4860]: I1211 09:06:24.973699 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-mmtt9"] Dec 11 09:06:24 crc kubenswrapper[4860]: E1211 09:06:24.974393 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fb3c7c5-8c06-4e05-ad79-11f0671e16da" containerName="keystone-cron" Dec 11 09:06:24 crc kubenswrapper[4860]: I1211 09:06:24.974414 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fb3c7c5-8c06-4e05-ad79-11f0671e16da" containerName="keystone-cron" Dec 11 09:06:24 crc kubenswrapper[4860]: I1211 09:06:24.974720 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="8fb3c7c5-8c06-4e05-ad79-11f0671e16da" containerName="keystone-cron" Dec 11 09:06:24 crc kubenswrapper[4860]: I1211 09:06:24.976181 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mmtt9" Dec 11 09:06:24 crc kubenswrapper[4860]: I1211 09:06:24.985230 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mmtt9"] Dec 11 09:06:25 crc kubenswrapper[4860]: I1211 09:06:25.077384 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hmw6t\" (UniqueName: \"kubernetes.io/projected/91f1321a-00c3-43a8-bd56-add4ca029780-kube-api-access-hmw6t\") pod \"certified-operators-mmtt9\" (UID: \"91f1321a-00c3-43a8-bd56-add4ca029780\") " pod="openshift-marketplace/certified-operators-mmtt9" Dec 11 09:06:25 crc kubenswrapper[4860]: I1211 09:06:25.077537 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91f1321a-00c3-43a8-bd56-add4ca029780-catalog-content\") pod \"certified-operators-mmtt9\" (UID: \"91f1321a-00c3-43a8-bd56-add4ca029780\") " pod="openshift-marketplace/certified-operators-mmtt9" Dec 11 09:06:25 crc kubenswrapper[4860]: I1211 09:06:25.077832 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91f1321a-00c3-43a8-bd56-add4ca029780-utilities\") pod \"certified-operators-mmtt9\" (UID: \"91f1321a-00c3-43a8-bd56-add4ca029780\") " pod="openshift-marketplace/certified-operators-mmtt9" Dec 11 09:06:25 crc kubenswrapper[4860]: I1211 09:06:25.179848 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91f1321a-00c3-43a8-bd56-add4ca029780-catalog-content\") pod \"certified-operators-mmtt9\" (UID: \"91f1321a-00c3-43a8-bd56-add4ca029780\") " pod="openshift-marketplace/certified-operators-mmtt9" Dec 11 09:06:25 crc kubenswrapper[4860]: I1211 09:06:25.179978 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91f1321a-00c3-43a8-bd56-add4ca029780-utilities\") pod \"certified-operators-mmtt9\" (UID: \"91f1321a-00c3-43a8-bd56-add4ca029780\") " pod="openshift-marketplace/certified-operators-mmtt9" Dec 11 09:06:25 crc kubenswrapper[4860]: I1211 09:06:25.180031 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hmw6t\" (UniqueName: \"kubernetes.io/projected/91f1321a-00c3-43a8-bd56-add4ca029780-kube-api-access-hmw6t\") pod \"certified-operators-mmtt9\" (UID: \"91f1321a-00c3-43a8-bd56-add4ca029780\") " pod="openshift-marketplace/certified-operators-mmtt9" Dec 11 09:06:25 crc kubenswrapper[4860]: I1211 09:06:25.180391 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91f1321a-00c3-43a8-bd56-add4ca029780-catalog-content\") pod \"certified-operators-mmtt9\" (UID: \"91f1321a-00c3-43a8-bd56-add4ca029780\") " pod="openshift-marketplace/certified-operators-mmtt9" Dec 11 09:06:25 crc kubenswrapper[4860]: I1211 09:06:25.180491 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91f1321a-00c3-43a8-bd56-add4ca029780-utilities\") pod \"certified-operators-mmtt9\" (UID: \"91f1321a-00c3-43a8-bd56-add4ca029780\") " pod="openshift-marketplace/certified-operators-mmtt9" Dec 11 09:06:25 crc kubenswrapper[4860]: I1211 09:06:25.214306 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hmw6t\" (UniqueName: \"kubernetes.io/projected/91f1321a-00c3-43a8-bd56-add4ca029780-kube-api-access-hmw6t\") pod \"certified-operators-mmtt9\" (UID: \"91f1321a-00c3-43a8-bd56-add4ca029780\") " pod="openshift-marketplace/certified-operators-mmtt9" Dec 11 09:06:25 crc kubenswrapper[4860]: I1211 09:06:25.312540 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mmtt9" Dec 11 09:06:25 crc kubenswrapper[4860]: I1211 09:06:25.818108 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mmtt9"] Dec 11 09:06:26 crc kubenswrapper[4860]: I1211 09:06:26.201091 4860 generic.go:334] "Generic (PLEG): container finished" podID="91f1321a-00c3-43a8-bd56-add4ca029780" containerID="81768794f2f872c0777df7a56b46523db04d2fb2246d0258f185edb17452da85" exitCode=0 Dec 11 09:06:26 crc kubenswrapper[4860]: I1211 09:06:26.201158 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mmtt9" event={"ID":"91f1321a-00c3-43a8-bd56-add4ca029780","Type":"ContainerDied","Data":"81768794f2f872c0777df7a56b46523db04d2fb2246d0258f185edb17452da85"} Dec 11 09:06:26 crc kubenswrapper[4860]: I1211 09:06:26.201232 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mmtt9" event={"ID":"91f1321a-00c3-43a8-bd56-add4ca029780","Type":"ContainerStarted","Data":"0b3612e5c421733472be8f8d179670ce42fbe97c95b551699ef98c104459b0e7"} Dec 11 09:06:26 crc kubenswrapper[4860]: I1211 09:06:26.202985 4860 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 11 09:06:28 crc kubenswrapper[4860]: I1211 09:06:28.222151 4860 generic.go:334] "Generic (PLEG): container finished" podID="91f1321a-00c3-43a8-bd56-add4ca029780" containerID="619dd1ba11a09a1196a4e956e2fe414ed7746b8f33710bf71c0d0768c234b4ce" exitCode=0 Dec 11 09:06:28 crc kubenswrapper[4860]: I1211 09:06:28.222513 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mmtt9" event={"ID":"91f1321a-00c3-43a8-bd56-add4ca029780","Type":"ContainerDied","Data":"619dd1ba11a09a1196a4e956e2fe414ed7746b8f33710bf71c0d0768c234b4ce"} Dec 11 09:06:29 crc kubenswrapper[4860]: I1211 09:06:29.234127 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mmtt9" event={"ID":"91f1321a-00c3-43a8-bd56-add4ca029780","Type":"ContainerStarted","Data":"a2484ecb14b06167fc67c85af14815f6d29a85589a426f77ae1d14bc6b809412"} Dec 11 09:06:29 crc kubenswrapper[4860]: I1211 09:06:29.253954 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-mmtt9" podStartSLOduration=2.49966793 podStartE2EDuration="5.253937101s" podCreationTimestamp="2025-12-11 09:06:24 +0000 UTC" firstStartedPulling="2025-12-11 09:06:26.202757507 +0000 UTC m=+3318.931276562" lastFinishedPulling="2025-12-11 09:06:28.957026678 +0000 UTC m=+3321.685545733" observedRunningTime="2025-12-11 09:06:29.252892138 +0000 UTC m=+3321.981411193" watchObservedRunningTime="2025-12-11 09:06:29.253937101 +0000 UTC m=+3321.982456156" Dec 11 09:06:30 crc kubenswrapper[4860]: I1211 09:06:30.944303 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-8bmlh"] Dec 11 09:06:30 crc kubenswrapper[4860]: I1211 09:06:30.950459 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8bmlh" Dec 11 09:06:30 crc kubenswrapper[4860]: I1211 09:06:30.960993 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8bmlh"] Dec 11 09:06:31 crc kubenswrapper[4860]: I1211 09:06:31.009470 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ca0c7e5-f1df-4ae9-b21c-ee41430e0e5c-utilities\") pod \"community-operators-8bmlh\" (UID: \"5ca0c7e5-f1df-4ae9-b21c-ee41430e0e5c\") " pod="openshift-marketplace/community-operators-8bmlh" Dec 11 09:06:31 crc kubenswrapper[4860]: I1211 09:06:31.009546 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9j69j\" (UniqueName: \"kubernetes.io/projected/5ca0c7e5-f1df-4ae9-b21c-ee41430e0e5c-kube-api-access-9j69j\") pod \"community-operators-8bmlh\" (UID: \"5ca0c7e5-f1df-4ae9-b21c-ee41430e0e5c\") " pod="openshift-marketplace/community-operators-8bmlh" Dec 11 09:06:31 crc kubenswrapper[4860]: I1211 09:06:31.009571 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ca0c7e5-f1df-4ae9-b21c-ee41430e0e5c-catalog-content\") pod \"community-operators-8bmlh\" (UID: \"5ca0c7e5-f1df-4ae9-b21c-ee41430e0e5c\") " pod="openshift-marketplace/community-operators-8bmlh" Dec 11 09:06:31 crc kubenswrapper[4860]: I1211 09:06:31.112364 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ca0c7e5-f1df-4ae9-b21c-ee41430e0e5c-utilities\") pod \"community-operators-8bmlh\" (UID: \"5ca0c7e5-f1df-4ae9-b21c-ee41430e0e5c\") " pod="openshift-marketplace/community-operators-8bmlh" Dec 11 09:06:31 crc kubenswrapper[4860]: I1211 09:06:31.112818 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9j69j\" (UniqueName: \"kubernetes.io/projected/5ca0c7e5-f1df-4ae9-b21c-ee41430e0e5c-kube-api-access-9j69j\") pod \"community-operators-8bmlh\" (UID: \"5ca0c7e5-f1df-4ae9-b21c-ee41430e0e5c\") " pod="openshift-marketplace/community-operators-8bmlh" Dec 11 09:06:31 crc kubenswrapper[4860]: I1211 09:06:31.112848 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ca0c7e5-f1df-4ae9-b21c-ee41430e0e5c-catalog-content\") pod \"community-operators-8bmlh\" (UID: \"5ca0c7e5-f1df-4ae9-b21c-ee41430e0e5c\") " pod="openshift-marketplace/community-operators-8bmlh" Dec 11 09:06:31 crc kubenswrapper[4860]: I1211 09:06:31.112842 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ca0c7e5-f1df-4ae9-b21c-ee41430e0e5c-utilities\") pod \"community-operators-8bmlh\" (UID: \"5ca0c7e5-f1df-4ae9-b21c-ee41430e0e5c\") " pod="openshift-marketplace/community-operators-8bmlh" Dec 11 09:06:31 crc kubenswrapper[4860]: I1211 09:06:31.113126 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ca0c7e5-f1df-4ae9-b21c-ee41430e0e5c-catalog-content\") pod \"community-operators-8bmlh\" (UID: \"5ca0c7e5-f1df-4ae9-b21c-ee41430e0e5c\") " pod="openshift-marketplace/community-operators-8bmlh" Dec 11 09:06:31 crc kubenswrapper[4860]: I1211 09:06:31.140721 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9j69j\" (UniqueName: \"kubernetes.io/projected/5ca0c7e5-f1df-4ae9-b21c-ee41430e0e5c-kube-api-access-9j69j\") pod \"community-operators-8bmlh\" (UID: \"5ca0c7e5-f1df-4ae9-b21c-ee41430e0e5c\") " pod="openshift-marketplace/community-operators-8bmlh" Dec 11 09:06:31 crc kubenswrapper[4860]: I1211 09:06:31.277967 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8bmlh" Dec 11 09:06:31 crc kubenswrapper[4860]: I1211 09:06:31.929239 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8bmlh"] Dec 11 09:06:32 crc kubenswrapper[4860]: I1211 09:06:32.262582 4860 generic.go:334] "Generic (PLEG): container finished" podID="5ca0c7e5-f1df-4ae9-b21c-ee41430e0e5c" containerID="9469ffccdae44f3dc733381c0b7ad283b8a12f92cd43ec194cacb29302b8e9e2" exitCode=0 Dec 11 09:06:32 crc kubenswrapper[4860]: I1211 09:06:32.262626 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8bmlh" event={"ID":"5ca0c7e5-f1df-4ae9-b21c-ee41430e0e5c","Type":"ContainerDied","Data":"9469ffccdae44f3dc733381c0b7ad283b8a12f92cd43ec194cacb29302b8e9e2"} Dec 11 09:06:32 crc kubenswrapper[4860]: I1211 09:06:32.262871 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8bmlh" event={"ID":"5ca0c7e5-f1df-4ae9-b21c-ee41430e0e5c","Type":"ContainerStarted","Data":"70920d26dfa5712c6f79fe0cffc4697bf28e30a6a1ae16552094f03ca4aeb1e0"} Dec 11 09:06:34 crc kubenswrapper[4860]: I1211 09:06:34.283966 4860 generic.go:334] "Generic (PLEG): container finished" podID="5ca0c7e5-f1df-4ae9-b21c-ee41430e0e5c" containerID="3ff1f50e5780bf40c7d70a28816acc9760269ab264646d9863c532b9891d1fb5" exitCode=0 Dec 11 09:06:34 crc kubenswrapper[4860]: I1211 09:06:34.284105 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8bmlh" event={"ID":"5ca0c7e5-f1df-4ae9-b21c-ee41430e0e5c","Type":"ContainerDied","Data":"3ff1f50e5780bf40c7d70a28816acc9760269ab264646d9863c532b9891d1fb5"} Dec 11 09:06:35 crc kubenswrapper[4860]: I1211 09:06:35.297669 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8bmlh" event={"ID":"5ca0c7e5-f1df-4ae9-b21c-ee41430e0e5c","Type":"ContainerStarted","Data":"35625c4270d350d7da63d604f302120284c558aa5a9e45e35ba428d490d53e4e"} Dec 11 09:06:35 crc kubenswrapper[4860]: I1211 09:06:35.313165 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-mmtt9" Dec 11 09:06:35 crc kubenswrapper[4860]: I1211 09:06:35.314908 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-mmtt9" Dec 11 09:06:35 crc kubenswrapper[4860]: I1211 09:06:35.321527 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-8bmlh" podStartSLOduration=2.7981458200000002 podStartE2EDuration="5.321509636s" podCreationTimestamp="2025-12-11 09:06:30 +0000 UTC" firstStartedPulling="2025-12-11 09:06:32.264626041 +0000 UTC m=+3324.993145096" lastFinishedPulling="2025-12-11 09:06:34.787989857 +0000 UTC m=+3327.516508912" observedRunningTime="2025-12-11 09:06:35.318795561 +0000 UTC m=+3328.047314636" watchObservedRunningTime="2025-12-11 09:06:35.321509636 +0000 UTC m=+3328.050028691" Dec 11 09:06:35 crc kubenswrapper[4860]: I1211 09:06:35.366214 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-mmtt9" Dec 11 09:06:36 crc kubenswrapper[4860]: I1211 09:06:36.353118 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-mmtt9" Dec 11 09:06:36 crc kubenswrapper[4860]: I1211 09:06:36.579053 4860 scope.go:117] "RemoveContainer" containerID="76665d219a809fd146709b7837eb798329ce7fb1c53d5b15c247e7e654985e07" Dec 11 09:06:36 crc kubenswrapper[4860]: E1211 09:06:36.579310 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-99qgp_openshift-machine-config-operator(31c30642-6e60-41b4-a477-0d802424e0aa)\"" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" Dec 11 09:06:37 crc kubenswrapper[4860]: I1211 09:06:37.535078 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mmtt9"] Dec 11 09:06:39 crc kubenswrapper[4860]: I1211 09:06:39.331928 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-mmtt9" podUID="91f1321a-00c3-43a8-bd56-add4ca029780" containerName="registry-server" containerID="cri-o://a2484ecb14b06167fc67c85af14815f6d29a85589a426f77ae1d14bc6b809412" gracePeriod=2 Dec 11 09:06:40 crc kubenswrapper[4860]: I1211 09:06:40.342634 4860 generic.go:334] "Generic (PLEG): container finished" podID="91f1321a-00c3-43a8-bd56-add4ca029780" containerID="a2484ecb14b06167fc67c85af14815f6d29a85589a426f77ae1d14bc6b809412" exitCode=0 Dec 11 09:06:40 crc kubenswrapper[4860]: I1211 09:06:40.342691 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mmtt9" event={"ID":"91f1321a-00c3-43a8-bd56-add4ca029780","Type":"ContainerDied","Data":"a2484ecb14b06167fc67c85af14815f6d29a85589a426f77ae1d14bc6b809412"} Dec 11 09:06:40 crc kubenswrapper[4860]: I1211 09:06:40.342985 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mmtt9" event={"ID":"91f1321a-00c3-43a8-bd56-add4ca029780","Type":"ContainerDied","Data":"0b3612e5c421733472be8f8d179670ce42fbe97c95b551699ef98c104459b0e7"} Dec 11 09:06:40 crc kubenswrapper[4860]: I1211 09:06:40.343002 4860 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0b3612e5c421733472be8f8d179670ce42fbe97c95b551699ef98c104459b0e7" Dec 11 09:06:40 crc kubenswrapper[4860]: I1211 09:06:40.349343 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mmtt9" Dec 11 09:06:40 crc kubenswrapper[4860]: I1211 09:06:40.434689 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91f1321a-00c3-43a8-bd56-add4ca029780-utilities\") pod \"91f1321a-00c3-43a8-bd56-add4ca029780\" (UID: \"91f1321a-00c3-43a8-bd56-add4ca029780\") " Dec 11 09:06:40 crc kubenswrapper[4860]: I1211 09:06:40.434774 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hmw6t\" (UniqueName: \"kubernetes.io/projected/91f1321a-00c3-43a8-bd56-add4ca029780-kube-api-access-hmw6t\") pod \"91f1321a-00c3-43a8-bd56-add4ca029780\" (UID: \"91f1321a-00c3-43a8-bd56-add4ca029780\") " Dec 11 09:06:40 crc kubenswrapper[4860]: I1211 09:06:40.434877 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91f1321a-00c3-43a8-bd56-add4ca029780-catalog-content\") pod \"91f1321a-00c3-43a8-bd56-add4ca029780\" (UID: \"91f1321a-00c3-43a8-bd56-add4ca029780\") " Dec 11 09:06:40 crc kubenswrapper[4860]: I1211 09:06:40.435545 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/91f1321a-00c3-43a8-bd56-add4ca029780-utilities" (OuterVolumeSpecName: "utilities") pod "91f1321a-00c3-43a8-bd56-add4ca029780" (UID: "91f1321a-00c3-43a8-bd56-add4ca029780"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 09:06:40 crc kubenswrapper[4860]: I1211 09:06:40.442479 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/91f1321a-00c3-43a8-bd56-add4ca029780-kube-api-access-hmw6t" (OuterVolumeSpecName: "kube-api-access-hmw6t") pod "91f1321a-00c3-43a8-bd56-add4ca029780" (UID: "91f1321a-00c3-43a8-bd56-add4ca029780"). InnerVolumeSpecName "kube-api-access-hmw6t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 09:06:40 crc kubenswrapper[4860]: I1211 09:06:40.482003 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/91f1321a-00c3-43a8-bd56-add4ca029780-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "91f1321a-00c3-43a8-bd56-add4ca029780" (UID: "91f1321a-00c3-43a8-bd56-add4ca029780"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 09:06:40 crc kubenswrapper[4860]: I1211 09:06:40.538228 4860 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91f1321a-00c3-43a8-bd56-add4ca029780-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 09:06:40 crc kubenswrapper[4860]: I1211 09:06:40.538277 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hmw6t\" (UniqueName: \"kubernetes.io/projected/91f1321a-00c3-43a8-bd56-add4ca029780-kube-api-access-hmw6t\") on node \"crc\" DevicePath \"\"" Dec 11 09:06:40 crc kubenswrapper[4860]: I1211 09:06:40.538292 4860 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91f1321a-00c3-43a8-bd56-add4ca029780-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 09:06:41 crc kubenswrapper[4860]: I1211 09:06:41.278855 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-8bmlh" Dec 11 09:06:41 crc kubenswrapper[4860]: I1211 09:06:41.279290 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-8bmlh" Dec 11 09:06:41 crc kubenswrapper[4860]: I1211 09:06:41.328055 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-8bmlh" Dec 11 09:06:41 crc kubenswrapper[4860]: I1211 09:06:41.350484 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mmtt9" Dec 11 09:06:41 crc kubenswrapper[4860]: I1211 09:06:41.388670 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mmtt9"] Dec 11 09:06:41 crc kubenswrapper[4860]: I1211 09:06:41.397910 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-mmtt9"] Dec 11 09:06:41 crc kubenswrapper[4860]: I1211 09:06:41.404844 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-8bmlh" Dec 11 09:06:41 crc kubenswrapper[4860]: I1211 09:06:41.595820 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="91f1321a-00c3-43a8-bd56-add4ca029780" path="/var/lib/kubelet/pods/91f1321a-00c3-43a8-bd56-add4ca029780/volumes" Dec 11 09:06:43 crc kubenswrapper[4860]: I1211 09:06:43.735903 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8bmlh"] Dec 11 09:06:43 crc kubenswrapper[4860]: I1211 09:06:43.736636 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-8bmlh" podUID="5ca0c7e5-f1df-4ae9-b21c-ee41430e0e5c" containerName="registry-server" containerID="cri-o://35625c4270d350d7da63d604f302120284c558aa5a9e45e35ba428d490d53e4e" gracePeriod=2 Dec 11 09:06:44 crc kubenswrapper[4860]: I1211 09:06:44.283233 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8bmlh" Dec 11 09:06:44 crc kubenswrapper[4860]: I1211 09:06:44.307086 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ca0c7e5-f1df-4ae9-b21c-ee41430e0e5c-catalog-content\") pod \"5ca0c7e5-f1df-4ae9-b21c-ee41430e0e5c\" (UID: \"5ca0c7e5-f1df-4ae9-b21c-ee41430e0e5c\") " Dec 11 09:06:44 crc kubenswrapper[4860]: I1211 09:06:44.307201 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9j69j\" (UniqueName: \"kubernetes.io/projected/5ca0c7e5-f1df-4ae9-b21c-ee41430e0e5c-kube-api-access-9j69j\") pod \"5ca0c7e5-f1df-4ae9-b21c-ee41430e0e5c\" (UID: \"5ca0c7e5-f1df-4ae9-b21c-ee41430e0e5c\") " Dec 11 09:06:44 crc kubenswrapper[4860]: I1211 09:06:44.307355 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ca0c7e5-f1df-4ae9-b21c-ee41430e0e5c-utilities\") pod \"5ca0c7e5-f1df-4ae9-b21c-ee41430e0e5c\" (UID: \"5ca0c7e5-f1df-4ae9-b21c-ee41430e0e5c\") " Dec 11 09:06:44 crc kubenswrapper[4860]: I1211 09:06:44.308857 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5ca0c7e5-f1df-4ae9-b21c-ee41430e0e5c-utilities" (OuterVolumeSpecName: "utilities") pod "5ca0c7e5-f1df-4ae9-b21c-ee41430e0e5c" (UID: "5ca0c7e5-f1df-4ae9-b21c-ee41430e0e5c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 09:06:44 crc kubenswrapper[4860]: I1211 09:06:44.314868 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ca0c7e5-f1df-4ae9-b21c-ee41430e0e5c-kube-api-access-9j69j" (OuterVolumeSpecName: "kube-api-access-9j69j") pod "5ca0c7e5-f1df-4ae9-b21c-ee41430e0e5c" (UID: "5ca0c7e5-f1df-4ae9-b21c-ee41430e0e5c"). InnerVolumeSpecName "kube-api-access-9j69j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 09:06:44 crc kubenswrapper[4860]: I1211 09:06:44.365239 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5ca0c7e5-f1df-4ae9-b21c-ee41430e0e5c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5ca0c7e5-f1df-4ae9-b21c-ee41430e0e5c" (UID: "5ca0c7e5-f1df-4ae9-b21c-ee41430e0e5c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 09:06:44 crc kubenswrapper[4860]: I1211 09:06:44.383445 4860 generic.go:334] "Generic (PLEG): container finished" podID="5ca0c7e5-f1df-4ae9-b21c-ee41430e0e5c" containerID="35625c4270d350d7da63d604f302120284c558aa5a9e45e35ba428d490d53e4e" exitCode=0 Dec 11 09:06:44 crc kubenswrapper[4860]: I1211 09:06:44.383493 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8bmlh" event={"ID":"5ca0c7e5-f1df-4ae9-b21c-ee41430e0e5c","Type":"ContainerDied","Data":"35625c4270d350d7da63d604f302120284c558aa5a9e45e35ba428d490d53e4e"} Dec 11 09:06:44 crc kubenswrapper[4860]: I1211 09:06:44.383530 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8bmlh" event={"ID":"5ca0c7e5-f1df-4ae9-b21c-ee41430e0e5c","Type":"ContainerDied","Data":"70920d26dfa5712c6f79fe0cffc4697bf28e30a6a1ae16552094f03ca4aeb1e0"} Dec 11 09:06:44 crc kubenswrapper[4860]: I1211 09:06:44.383559 4860 scope.go:117] "RemoveContainer" containerID="35625c4270d350d7da63d604f302120284c558aa5a9e45e35ba428d490d53e4e" Dec 11 09:06:44 crc kubenswrapper[4860]: I1211 09:06:44.383560 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8bmlh" Dec 11 09:06:44 crc kubenswrapper[4860]: I1211 09:06:44.408420 4860 scope.go:117] "RemoveContainer" containerID="3ff1f50e5780bf40c7d70a28816acc9760269ab264646d9863c532b9891d1fb5" Dec 11 09:06:44 crc kubenswrapper[4860]: I1211 09:06:44.411098 4860 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ca0c7e5-f1df-4ae9-b21c-ee41430e0e5c-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 09:06:44 crc kubenswrapper[4860]: I1211 09:06:44.411134 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9j69j\" (UniqueName: \"kubernetes.io/projected/5ca0c7e5-f1df-4ae9-b21c-ee41430e0e5c-kube-api-access-9j69j\") on node \"crc\" DevicePath \"\"" Dec 11 09:06:44 crc kubenswrapper[4860]: I1211 09:06:44.411146 4860 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ca0c7e5-f1df-4ae9-b21c-ee41430e0e5c-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 09:06:44 crc kubenswrapper[4860]: I1211 09:06:44.423753 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8bmlh"] Dec 11 09:06:44 crc kubenswrapper[4860]: I1211 09:06:44.432022 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-8bmlh"] Dec 11 09:06:44 crc kubenswrapper[4860]: I1211 09:06:44.455891 4860 scope.go:117] "RemoveContainer" containerID="9469ffccdae44f3dc733381c0b7ad283b8a12f92cd43ec194cacb29302b8e9e2" Dec 11 09:06:44 crc kubenswrapper[4860]: I1211 09:06:44.483356 4860 scope.go:117] "RemoveContainer" containerID="35625c4270d350d7da63d604f302120284c558aa5a9e45e35ba428d490d53e4e" Dec 11 09:06:44 crc kubenswrapper[4860]: E1211 09:06:44.483750 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"35625c4270d350d7da63d604f302120284c558aa5a9e45e35ba428d490d53e4e\": container with ID starting with 35625c4270d350d7da63d604f302120284c558aa5a9e45e35ba428d490d53e4e not found: ID does not exist" containerID="35625c4270d350d7da63d604f302120284c558aa5a9e45e35ba428d490d53e4e" Dec 11 09:06:44 crc kubenswrapper[4860]: I1211 09:06:44.483777 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"35625c4270d350d7da63d604f302120284c558aa5a9e45e35ba428d490d53e4e"} err="failed to get container status \"35625c4270d350d7da63d604f302120284c558aa5a9e45e35ba428d490d53e4e\": rpc error: code = NotFound desc = could not find container \"35625c4270d350d7da63d604f302120284c558aa5a9e45e35ba428d490d53e4e\": container with ID starting with 35625c4270d350d7da63d604f302120284c558aa5a9e45e35ba428d490d53e4e not found: ID does not exist" Dec 11 09:06:44 crc kubenswrapper[4860]: I1211 09:06:44.483797 4860 scope.go:117] "RemoveContainer" containerID="3ff1f50e5780bf40c7d70a28816acc9760269ab264646d9863c532b9891d1fb5" Dec 11 09:06:44 crc kubenswrapper[4860]: E1211 09:06:44.484006 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3ff1f50e5780bf40c7d70a28816acc9760269ab264646d9863c532b9891d1fb5\": container with ID starting with 3ff1f50e5780bf40c7d70a28816acc9760269ab264646d9863c532b9891d1fb5 not found: ID does not exist" containerID="3ff1f50e5780bf40c7d70a28816acc9760269ab264646d9863c532b9891d1fb5" Dec 11 09:06:44 crc kubenswrapper[4860]: I1211 09:06:44.484026 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3ff1f50e5780bf40c7d70a28816acc9760269ab264646d9863c532b9891d1fb5"} err="failed to get container status \"3ff1f50e5780bf40c7d70a28816acc9760269ab264646d9863c532b9891d1fb5\": rpc error: code = NotFound desc = could not find container \"3ff1f50e5780bf40c7d70a28816acc9760269ab264646d9863c532b9891d1fb5\": container with ID starting with 3ff1f50e5780bf40c7d70a28816acc9760269ab264646d9863c532b9891d1fb5 not found: ID does not exist" Dec 11 09:06:44 crc kubenswrapper[4860]: I1211 09:06:44.484037 4860 scope.go:117] "RemoveContainer" containerID="9469ffccdae44f3dc733381c0b7ad283b8a12f92cd43ec194cacb29302b8e9e2" Dec 11 09:06:44 crc kubenswrapper[4860]: E1211 09:06:44.484200 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9469ffccdae44f3dc733381c0b7ad283b8a12f92cd43ec194cacb29302b8e9e2\": container with ID starting with 9469ffccdae44f3dc733381c0b7ad283b8a12f92cd43ec194cacb29302b8e9e2 not found: ID does not exist" containerID="9469ffccdae44f3dc733381c0b7ad283b8a12f92cd43ec194cacb29302b8e9e2" Dec 11 09:06:44 crc kubenswrapper[4860]: I1211 09:06:44.484219 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9469ffccdae44f3dc733381c0b7ad283b8a12f92cd43ec194cacb29302b8e9e2"} err="failed to get container status \"9469ffccdae44f3dc733381c0b7ad283b8a12f92cd43ec194cacb29302b8e9e2\": rpc error: code = NotFound desc = could not find container \"9469ffccdae44f3dc733381c0b7ad283b8a12f92cd43ec194cacb29302b8e9e2\": container with ID starting with 9469ffccdae44f3dc733381c0b7ad283b8a12f92cd43ec194cacb29302b8e9e2 not found: ID does not exist" Dec 11 09:06:45 crc kubenswrapper[4860]: I1211 09:06:45.590279 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ca0c7e5-f1df-4ae9-b21c-ee41430e0e5c" path="/var/lib/kubelet/pods/5ca0c7e5-f1df-4ae9-b21c-ee41430e0e5c/volumes" Dec 11 09:06:47 crc kubenswrapper[4860]: I1211 09:06:47.579496 4860 scope.go:117] "RemoveContainer" containerID="76665d219a809fd146709b7837eb798329ce7fb1c53d5b15c247e7e654985e07" Dec 11 09:06:47 crc kubenswrapper[4860]: E1211 09:06:47.581831 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-99qgp_openshift-machine-config-operator(31c30642-6e60-41b4-a477-0d802424e0aa)\"" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" Dec 11 09:06:58 crc kubenswrapper[4860]: I1211 09:06:58.579281 4860 scope.go:117] "RemoveContainer" containerID="76665d219a809fd146709b7837eb798329ce7fb1c53d5b15c247e7e654985e07" Dec 11 09:06:58 crc kubenswrapper[4860]: E1211 09:06:58.580013 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-99qgp_openshift-machine-config-operator(31c30642-6e60-41b4-a477-0d802424e0aa)\"" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" Dec 11 09:07:01 crc kubenswrapper[4860]: I1211 09:07:01.567064 4860 generic.go:334] "Generic (PLEG): container finished" podID="3b659d05-76c9-4b04-a05e-a7dd488e9852" containerID="e8b2b959a1b90a4db8fbd7a58476c9c82fb1f896125d0763a74f9579fa5726c2" exitCode=0 Dec 11 09:07:01 crc kubenswrapper[4860]: I1211 09:07:01.567158 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"3b659d05-76c9-4b04-a05e-a7dd488e9852","Type":"ContainerDied","Data":"e8b2b959a1b90a4db8fbd7a58476c9c82fb1f896125d0763a74f9579fa5726c2"} Dec 11 09:07:02 crc kubenswrapper[4860]: I1211 09:07:02.946490 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 11 09:07:03 crc kubenswrapper[4860]: I1211 09:07:03.058565 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3b659d05-76c9-4b04-a05e-a7dd488e9852-config-data\") pod \"3b659d05-76c9-4b04-a05e-a7dd488e9852\" (UID: \"3b659d05-76c9-4b04-a05e-a7dd488e9852\") " Dec 11 09:07:03 crc kubenswrapper[4860]: I1211 09:07:03.058950 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3b659d05-76c9-4b04-a05e-a7dd488e9852-ssh-key\") pod \"3b659d05-76c9-4b04-a05e-a7dd488e9852\" (UID: \"3b659d05-76c9-4b04-a05e-a7dd488e9852\") " Dec 11 09:07:03 crc kubenswrapper[4860]: I1211 09:07:03.058974 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/3b659d05-76c9-4b04-a05e-a7dd488e9852-openstack-config-secret\") pod \"3b659d05-76c9-4b04-a05e-a7dd488e9852\" (UID: \"3b659d05-76c9-4b04-a05e-a7dd488e9852\") " Dec 11 09:07:03 crc kubenswrapper[4860]: I1211 09:07:03.059024 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/3b659d05-76c9-4b04-a05e-a7dd488e9852-test-operator-ephemeral-temporary\") pod \"3b659d05-76c9-4b04-a05e-a7dd488e9852\" (UID: \"3b659d05-76c9-4b04-a05e-a7dd488e9852\") " Dec 11 09:07:03 crc kubenswrapper[4860]: I1211 09:07:03.059098 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"3b659d05-76c9-4b04-a05e-a7dd488e9852\" (UID: \"3b659d05-76c9-4b04-a05e-a7dd488e9852\") " Dec 11 09:07:03 crc kubenswrapper[4860]: I1211 09:07:03.059133 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q4rtg\" (UniqueName: \"kubernetes.io/projected/3b659d05-76c9-4b04-a05e-a7dd488e9852-kube-api-access-q4rtg\") pod \"3b659d05-76c9-4b04-a05e-a7dd488e9852\" (UID: \"3b659d05-76c9-4b04-a05e-a7dd488e9852\") " Dec 11 09:07:03 crc kubenswrapper[4860]: I1211 09:07:03.059149 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/3b659d05-76c9-4b04-a05e-a7dd488e9852-openstack-config\") pod \"3b659d05-76c9-4b04-a05e-a7dd488e9852\" (UID: \"3b659d05-76c9-4b04-a05e-a7dd488e9852\") " Dec 11 09:07:03 crc kubenswrapper[4860]: I1211 09:07:03.059163 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/3b659d05-76c9-4b04-a05e-a7dd488e9852-ca-certs\") pod \"3b659d05-76c9-4b04-a05e-a7dd488e9852\" (UID: \"3b659d05-76c9-4b04-a05e-a7dd488e9852\") " Dec 11 09:07:03 crc kubenswrapper[4860]: I1211 09:07:03.059189 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/3b659d05-76c9-4b04-a05e-a7dd488e9852-test-operator-ephemeral-workdir\") pod \"3b659d05-76c9-4b04-a05e-a7dd488e9852\" (UID: \"3b659d05-76c9-4b04-a05e-a7dd488e9852\") " Dec 11 09:07:03 crc kubenswrapper[4860]: I1211 09:07:03.059490 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3b659d05-76c9-4b04-a05e-a7dd488e9852-config-data" (OuterVolumeSpecName: "config-data") pod "3b659d05-76c9-4b04-a05e-a7dd488e9852" (UID: "3b659d05-76c9-4b04-a05e-a7dd488e9852"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 09:07:03 crc kubenswrapper[4860]: I1211 09:07:03.059917 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3b659d05-76c9-4b04-a05e-a7dd488e9852-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "3b659d05-76c9-4b04-a05e-a7dd488e9852" (UID: "3b659d05-76c9-4b04-a05e-a7dd488e9852"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 09:07:03 crc kubenswrapper[4860]: I1211 09:07:03.060023 4860 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3b659d05-76c9-4b04-a05e-a7dd488e9852-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 09:07:03 crc kubenswrapper[4860]: I1211 09:07:03.064619 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "test-operator-logs") pod "3b659d05-76c9-4b04-a05e-a7dd488e9852" (UID: "3b659d05-76c9-4b04-a05e-a7dd488e9852"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 11 09:07:03 crc kubenswrapper[4860]: I1211 09:07:03.065840 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b659d05-76c9-4b04-a05e-a7dd488e9852-kube-api-access-q4rtg" (OuterVolumeSpecName: "kube-api-access-q4rtg") pod "3b659d05-76c9-4b04-a05e-a7dd488e9852" (UID: "3b659d05-76c9-4b04-a05e-a7dd488e9852"). InnerVolumeSpecName "kube-api-access-q4rtg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 09:07:03 crc kubenswrapper[4860]: I1211 09:07:03.066518 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3b659d05-76c9-4b04-a05e-a7dd488e9852-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "3b659d05-76c9-4b04-a05e-a7dd488e9852" (UID: "3b659d05-76c9-4b04-a05e-a7dd488e9852"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 09:07:03 crc kubenswrapper[4860]: I1211 09:07:03.087296 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b659d05-76c9-4b04-a05e-a7dd488e9852-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "3b659d05-76c9-4b04-a05e-a7dd488e9852" (UID: "3b659d05-76c9-4b04-a05e-a7dd488e9852"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 09:07:03 crc kubenswrapper[4860]: I1211 09:07:03.087963 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b659d05-76c9-4b04-a05e-a7dd488e9852-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "3b659d05-76c9-4b04-a05e-a7dd488e9852" (UID: "3b659d05-76c9-4b04-a05e-a7dd488e9852"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 09:07:03 crc kubenswrapper[4860]: I1211 09:07:03.089722 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b659d05-76c9-4b04-a05e-a7dd488e9852-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "3b659d05-76c9-4b04-a05e-a7dd488e9852" (UID: "3b659d05-76c9-4b04-a05e-a7dd488e9852"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 09:07:03 crc kubenswrapper[4860]: I1211 09:07:03.108690 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3b659d05-76c9-4b04-a05e-a7dd488e9852-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "3b659d05-76c9-4b04-a05e-a7dd488e9852" (UID: "3b659d05-76c9-4b04-a05e-a7dd488e9852"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 09:07:03 crc kubenswrapper[4860]: I1211 09:07:03.162056 4860 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/3b659d05-76c9-4b04-a05e-a7dd488e9852-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Dec 11 09:07:03 crc kubenswrapper[4860]: I1211 09:07:03.162100 4860 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3b659d05-76c9-4b04-a05e-a7dd488e9852-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 11 09:07:03 crc kubenswrapper[4860]: I1211 09:07:03.162117 4860 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/3b659d05-76c9-4b04-a05e-a7dd488e9852-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 11 09:07:03 crc kubenswrapper[4860]: I1211 09:07:03.162131 4860 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/3b659d05-76c9-4b04-a05e-a7dd488e9852-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Dec 11 09:07:03 crc kubenswrapper[4860]: I1211 09:07:03.162173 4860 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Dec 11 09:07:03 crc kubenswrapper[4860]: I1211 09:07:03.162189 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q4rtg\" (UniqueName: \"kubernetes.io/projected/3b659d05-76c9-4b04-a05e-a7dd488e9852-kube-api-access-q4rtg\") on node \"crc\" DevicePath \"\"" Dec 11 09:07:03 crc kubenswrapper[4860]: I1211 09:07:03.162202 4860 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/3b659d05-76c9-4b04-a05e-a7dd488e9852-openstack-config\") on node \"crc\" DevicePath \"\"" Dec 11 09:07:03 crc kubenswrapper[4860]: I1211 09:07:03.162215 4860 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/3b659d05-76c9-4b04-a05e-a7dd488e9852-ca-certs\") on node \"crc\" DevicePath \"\"" Dec 11 09:07:03 crc kubenswrapper[4860]: I1211 09:07:03.184054 4860 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Dec 11 09:07:03 crc kubenswrapper[4860]: I1211 09:07:03.263761 4860 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Dec 11 09:07:03 crc kubenswrapper[4860]: I1211 09:07:03.585960 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 11 09:07:03 crc kubenswrapper[4860]: I1211 09:07:03.593202 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"3b659d05-76c9-4b04-a05e-a7dd488e9852","Type":"ContainerDied","Data":"bd540183ed09af10fb87bc2ab950c566a34bdf06b893dac04f7da1ec42044907"} Dec 11 09:07:03 crc kubenswrapper[4860]: I1211 09:07:03.593249 4860 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bd540183ed09af10fb87bc2ab950c566a34bdf06b893dac04f7da1ec42044907" Dec 11 09:07:09 crc kubenswrapper[4860]: I1211 09:07:09.485063 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-wtznh"] Dec 11 09:07:09 crc kubenswrapper[4860]: E1211 09:07:09.485984 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91f1321a-00c3-43a8-bd56-add4ca029780" containerName="registry-server" Dec 11 09:07:09 crc kubenswrapper[4860]: I1211 09:07:09.485999 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="91f1321a-00c3-43a8-bd56-add4ca029780" containerName="registry-server" Dec 11 09:07:09 crc kubenswrapper[4860]: E1211 09:07:09.486018 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91f1321a-00c3-43a8-bd56-add4ca029780" containerName="extract-content" Dec 11 09:07:09 crc kubenswrapper[4860]: I1211 09:07:09.486024 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="91f1321a-00c3-43a8-bd56-add4ca029780" containerName="extract-content" Dec 11 09:07:09 crc kubenswrapper[4860]: E1211 09:07:09.486035 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ca0c7e5-f1df-4ae9-b21c-ee41430e0e5c" containerName="extract-content" Dec 11 09:07:09 crc kubenswrapper[4860]: I1211 09:07:09.486042 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ca0c7e5-f1df-4ae9-b21c-ee41430e0e5c" containerName="extract-content" Dec 11 09:07:09 crc kubenswrapper[4860]: E1211 09:07:09.486056 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91f1321a-00c3-43a8-bd56-add4ca029780" containerName="extract-utilities" Dec 11 09:07:09 crc kubenswrapper[4860]: I1211 09:07:09.486062 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="91f1321a-00c3-43a8-bd56-add4ca029780" containerName="extract-utilities" Dec 11 09:07:09 crc kubenswrapper[4860]: E1211 09:07:09.486080 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b659d05-76c9-4b04-a05e-a7dd488e9852" containerName="tempest-tests-tempest-tests-runner" Dec 11 09:07:09 crc kubenswrapper[4860]: I1211 09:07:09.486086 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b659d05-76c9-4b04-a05e-a7dd488e9852" containerName="tempest-tests-tempest-tests-runner" Dec 11 09:07:09 crc kubenswrapper[4860]: E1211 09:07:09.486105 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ca0c7e5-f1df-4ae9-b21c-ee41430e0e5c" containerName="extract-utilities" Dec 11 09:07:09 crc kubenswrapper[4860]: I1211 09:07:09.486111 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ca0c7e5-f1df-4ae9-b21c-ee41430e0e5c" containerName="extract-utilities" Dec 11 09:07:09 crc kubenswrapper[4860]: E1211 09:07:09.486121 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ca0c7e5-f1df-4ae9-b21c-ee41430e0e5c" containerName="registry-server" Dec 11 09:07:09 crc kubenswrapper[4860]: I1211 09:07:09.486126 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ca0c7e5-f1df-4ae9-b21c-ee41430e0e5c" containerName="registry-server" Dec 11 09:07:09 crc kubenswrapper[4860]: I1211 09:07:09.486518 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="91f1321a-00c3-43a8-bd56-add4ca029780" containerName="registry-server" Dec 11 09:07:09 crc kubenswrapper[4860]: I1211 09:07:09.486540 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ca0c7e5-f1df-4ae9-b21c-ee41430e0e5c" containerName="registry-server" Dec 11 09:07:09 crc kubenswrapper[4860]: I1211 09:07:09.486562 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b659d05-76c9-4b04-a05e-a7dd488e9852" containerName="tempest-tests-tempest-tests-runner" Dec 11 09:07:09 crc kubenswrapper[4860]: I1211 09:07:09.489033 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wtznh" Dec 11 09:07:09 crc kubenswrapper[4860]: I1211 09:07:09.510884 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wtznh"] Dec 11 09:07:09 crc kubenswrapper[4860]: I1211 09:07:09.592800 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1a75cf4b-8d95-43ae-b738-c3fe0346e0f1-utilities\") pod \"redhat-operators-wtznh\" (UID: \"1a75cf4b-8d95-43ae-b738-c3fe0346e0f1\") " pod="openshift-marketplace/redhat-operators-wtznh" Dec 11 09:07:09 crc kubenswrapper[4860]: I1211 09:07:09.592869 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bh99t\" (UniqueName: \"kubernetes.io/projected/1a75cf4b-8d95-43ae-b738-c3fe0346e0f1-kube-api-access-bh99t\") pod \"redhat-operators-wtznh\" (UID: \"1a75cf4b-8d95-43ae-b738-c3fe0346e0f1\") " pod="openshift-marketplace/redhat-operators-wtznh" Dec 11 09:07:09 crc kubenswrapper[4860]: I1211 09:07:09.593021 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1a75cf4b-8d95-43ae-b738-c3fe0346e0f1-catalog-content\") pod \"redhat-operators-wtznh\" (UID: \"1a75cf4b-8d95-43ae-b738-c3fe0346e0f1\") " pod="openshift-marketplace/redhat-operators-wtznh" Dec 11 09:07:09 crc kubenswrapper[4860]: I1211 09:07:09.696013 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1a75cf4b-8d95-43ae-b738-c3fe0346e0f1-utilities\") pod \"redhat-operators-wtznh\" (UID: \"1a75cf4b-8d95-43ae-b738-c3fe0346e0f1\") " pod="openshift-marketplace/redhat-operators-wtznh" Dec 11 09:07:09 crc kubenswrapper[4860]: I1211 09:07:09.696344 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bh99t\" (UniqueName: \"kubernetes.io/projected/1a75cf4b-8d95-43ae-b738-c3fe0346e0f1-kube-api-access-bh99t\") pod \"redhat-operators-wtznh\" (UID: \"1a75cf4b-8d95-43ae-b738-c3fe0346e0f1\") " pod="openshift-marketplace/redhat-operators-wtznh" Dec 11 09:07:09 crc kubenswrapper[4860]: I1211 09:07:09.696511 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1a75cf4b-8d95-43ae-b738-c3fe0346e0f1-catalog-content\") pod \"redhat-operators-wtznh\" (UID: \"1a75cf4b-8d95-43ae-b738-c3fe0346e0f1\") " pod="openshift-marketplace/redhat-operators-wtznh" Dec 11 09:07:09 crc kubenswrapper[4860]: I1211 09:07:09.697407 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1a75cf4b-8d95-43ae-b738-c3fe0346e0f1-catalog-content\") pod \"redhat-operators-wtznh\" (UID: \"1a75cf4b-8d95-43ae-b738-c3fe0346e0f1\") " pod="openshift-marketplace/redhat-operators-wtznh" Dec 11 09:07:09 crc kubenswrapper[4860]: I1211 09:07:09.697475 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1a75cf4b-8d95-43ae-b738-c3fe0346e0f1-utilities\") pod \"redhat-operators-wtznh\" (UID: \"1a75cf4b-8d95-43ae-b738-c3fe0346e0f1\") " pod="openshift-marketplace/redhat-operators-wtznh" Dec 11 09:07:09 crc kubenswrapper[4860]: I1211 09:07:09.721207 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bh99t\" (UniqueName: \"kubernetes.io/projected/1a75cf4b-8d95-43ae-b738-c3fe0346e0f1-kube-api-access-bh99t\") pod \"redhat-operators-wtznh\" (UID: \"1a75cf4b-8d95-43ae-b738-c3fe0346e0f1\") " pod="openshift-marketplace/redhat-operators-wtznh" Dec 11 09:07:09 crc kubenswrapper[4860]: I1211 09:07:09.806557 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wtznh" Dec 11 09:07:10 crc kubenswrapper[4860]: I1211 09:07:10.337420 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wtznh"] Dec 11 09:07:10 crc kubenswrapper[4860]: I1211 09:07:10.649185 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wtznh" event={"ID":"1a75cf4b-8d95-43ae-b738-c3fe0346e0f1","Type":"ContainerStarted","Data":"f7af66d535caa99c000b5077fecdd281bf9e92bd886d4cc32b18e20e208337ff"} Dec 11 09:07:10 crc kubenswrapper[4860]: I1211 09:07:10.649447 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wtznh" event={"ID":"1a75cf4b-8d95-43ae-b738-c3fe0346e0f1","Type":"ContainerStarted","Data":"0a7fed6aab76718a90c1551d7cefa4137e625197602100213cfd8b72a091c7f3"} Dec 11 09:07:11 crc kubenswrapper[4860]: I1211 09:07:11.582079 4860 scope.go:117] "RemoveContainer" containerID="76665d219a809fd146709b7837eb798329ce7fb1c53d5b15c247e7e654985e07" Dec 11 09:07:11 crc kubenswrapper[4860]: E1211 09:07:11.582665 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-99qgp_openshift-machine-config-operator(31c30642-6e60-41b4-a477-0d802424e0aa)\"" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" Dec 11 09:07:11 crc kubenswrapper[4860]: I1211 09:07:11.662722 4860 generic.go:334] "Generic (PLEG): container finished" podID="1a75cf4b-8d95-43ae-b738-c3fe0346e0f1" containerID="f7af66d535caa99c000b5077fecdd281bf9e92bd886d4cc32b18e20e208337ff" exitCode=0 Dec 11 09:07:11 crc kubenswrapper[4860]: I1211 09:07:11.662774 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wtznh" event={"ID":"1a75cf4b-8d95-43ae-b738-c3fe0346e0f1","Type":"ContainerDied","Data":"f7af66d535caa99c000b5077fecdd281bf9e92bd886d4cc32b18e20e208337ff"} Dec 11 09:07:12 crc kubenswrapper[4860]: I1211 09:07:12.614968 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 11 09:07:12 crc kubenswrapper[4860]: I1211 09:07:12.616562 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 11 09:07:12 crc kubenswrapper[4860]: I1211 09:07:12.630136 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 11 09:07:12 crc kubenswrapper[4860]: I1211 09:07:12.638830 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-hlscw" Dec 11 09:07:12 crc kubenswrapper[4860]: I1211 09:07:12.673380 4860 generic.go:334] "Generic (PLEG): container finished" podID="1a75cf4b-8d95-43ae-b738-c3fe0346e0f1" containerID="5157603512a8bf14cca12e8fed6962249fa644c0b0b632f6e5f0befb0e521c6a" exitCode=0 Dec 11 09:07:12 crc kubenswrapper[4860]: I1211 09:07:12.673429 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wtznh" event={"ID":"1a75cf4b-8d95-43ae-b738-c3fe0346e0f1","Type":"ContainerDied","Data":"5157603512a8bf14cca12e8fed6962249fa644c0b0b632f6e5f0befb0e521c6a"} Dec 11 09:07:12 crc kubenswrapper[4860]: I1211 09:07:12.755619 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jr7mv\" (UniqueName: \"kubernetes.io/projected/b658df4a-886e-49de-b953-87539b28329b-kube-api-access-jr7mv\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"b658df4a-886e-49de-b953-87539b28329b\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 11 09:07:12 crc kubenswrapper[4860]: I1211 09:07:12.756029 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"b658df4a-886e-49de-b953-87539b28329b\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 11 09:07:12 crc kubenswrapper[4860]: I1211 09:07:12.858519 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jr7mv\" (UniqueName: \"kubernetes.io/projected/b658df4a-886e-49de-b953-87539b28329b-kube-api-access-jr7mv\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"b658df4a-886e-49de-b953-87539b28329b\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 11 09:07:12 crc kubenswrapper[4860]: I1211 09:07:12.858655 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"b658df4a-886e-49de-b953-87539b28329b\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 11 09:07:12 crc kubenswrapper[4860]: I1211 09:07:12.859545 4860 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"b658df4a-886e-49de-b953-87539b28329b\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 11 09:07:12 crc kubenswrapper[4860]: I1211 09:07:12.890473 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jr7mv\" (UniqueName: \"kubernetes.io/projected/b658df4a-886e-49de-b953-87539b28329b-kube-api-access-jr7mv\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"b658df4a-886e-49de-b953-87539b28329b\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 11 09:07:12 crc kubenswrapper[4860]: I1211 09:07:12.897783 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"b658df4a-886e-49de-b953-87539b28329b\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 11 09:07:12 crc kubenswrapper[4860]: I1211 09:07:12.956102 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 11 09:07:13 crc kubenswrapper[4860]: I1211 09:07:13.388834 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 11 09:07:13 crc kubenswrapper[4860]: I1211 09:07:13.685497 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"b658df4a-886e-49de-b953-87539b28329b","Type":"ContainerStarted","Data":"ee7b9e0437f5f9c8f00ac3b7dd22f0b1d5e0aa155a05e98a2dda0379fd347688"} Dec 11 09:07:14 crc kubenswrapper[4860]: I1211 09:07:14.701263 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"b658df4a-886e-49de-b953-87539b28329b","Type":"ContainerStarted","Data":"082d2d5f52fde32a63fdfcb34172d27a5fc78cabf6d374019859f386ba6dc067"} Dec 11 09:07:14 crc kubenswrapper[4860]: I1211 09:07:14.705241 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wtznh" event={"ID":"1a75cf4b-8d95-43ae-b738-c3fe0346e0f1","Type":"ContainerStarted","Data":"2e80a524939e00c7820aba6b79e50d94e27ca6a8688ebc0319683c6dad7d217f"} Dec 11 09:07:14 crc kubenswrapper[4860]: I1211 09:07:14.721416 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=1.922326404 podStartE2EDuration="2.72139536s" podCreationTimestamp="2025-12-11 09:07:12 +0000 UTC" firstStartedPulling="2025-12-11 09:07:13.394074952 +0000 UTC m=+3366.122594007" lastFinishedPulling="2025-12-11 09:07:14.193143908 +0000 UTC m=+3366.921662963" observedRunningTime="2025-12-11 09:07:14.716145555 +0000 UTC m=+3367.444664630" watchObservedRunningTime="2025-12-11 09:07:14.72139536 +0000 UTC m=+3367.449914415" Dec 11 09:07:14 crc kubenswrapper[4860]: I1211 09:07:14.748138 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-wtznh" podStartSLOduration=2.784767978 podStartE2EDuration="5.74811559s" podCreationTimestamp="2025-12-11 09:07:09 +0000 UTC" firstStartedPulling="2025-12-11 09:07:10.650694643 +0000 UTC m=+3363.379213698" lastFinishedPulling="2025-12-11 09:07:13.614042245 +0000 UTC m=+3366.342561310" observedRunningTime="2025-12-11 09:07:14.742091022 +0000 UTC m=+3367.470610077" watchObservedRunningTime="2025-12-11 09:07:14.74811559 +0000 UTC m=+3367.476634645" Dec 11 09:07:19 crc kubenswrapper[4860]: I1211 09:07:19.807671 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-wtznh" Dec 11 09:07:19 crc kubenswrapper[4860]: I1211 09:07:19.808285 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-wtznh" Dec 11 09:07:19 crc kubenswrapper[4860]: I1211 09:07:19.853181 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-wtznh" Dec 11 09:07:20 crc kubenswrapper[4860]: I1211 09:07:20.807817 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-wtznh" Dec 11 09:07:20 crc kubenswrapper[4860]: I1211 09:07:20.863782 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wtznh"] Dec 11 09:07:22 crc kubenswrapper[4860]: I1211 09:07:22.778108 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-wtznh" podUID="1a75cf4b-8d95-43ae-b738-c3fe0346e0f1" containerName="registry-server" containerID="cri-o://2e80a524939e00c7820aba6b79e50d94e27ca6a8688ebc0319683c6dad7d217f" gracePeriod=2 Dec 11 09:07:24 crc kubenswrapper[4860]: I1211 09:07:24.578828 4860 scope.go:117] "RemoveContainer" containerID="76665d219a809fd146709b7837eb798329ce7fb1c53d5b15c247e7e654985e07" Dec 11 09:07:24 crc kubenswrapper[4860]: E1211 09:07:24.579878 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-99qgp_openshift-machine-config-operator(31c30642-6e60-41b4-a477-0d802424e0aa)\"" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" Dec 11 09:07:24 crc kubenswrapper[4860]: I1211 09:07:24.799480 4860 generic.go:334] "Generic (PLEG): container finished" podID="1a75cf4b-8d95-43ae-b738-c3fe0346e0f1" containerID="2e80a524939e00c7820aba6b79e50d94e27ca6a8688ebc0319683c6dad7d217f" exitCode=0 Dec 11 09:07:24 crc kubenswrapper[4860]: I1211 09:07:24.799583 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wtznh" event={"ID":"1a75cf4b-8d95-43ae-b738-c3fe0346e0f1","Type":"ContainerDied","Data":"2e80a524939e00c7820aba6b79e50d94e27ca6a8688ebc0319683c6dad7d217f"} Dec 11 09:07:25 crc kubenswrapper[4860]: I1211 09:07:25.202184 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wtznh" Dec 11 09:07:25 crc kubenswrapper[4860]: I1211 09:07:25.305724 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1a75cf4b-8d95-43ae-b738-c3fe0346e0f1-utilities\") pod \"1a75cf4b-8d95-43ae-b738-c3fe0346e0f1\" (UID: \"1a75cf4b-8d95-43ae-b738-c3fe0346e0f1\") " Dec 11 09:07:25 crc kubenswrapper[4860]: I1211 09:07:25.306105 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1a75cf4b-8d95-43ae-b738-c3fe0346e0f1-catalog-content\") pod \"1a75cf4b-8d95-43ae-b738-c3fe0346e0f1\" (UID: \"1a75cf4b-8d95-43ae-b738-c3fe0346e0f1\") " Dec 11 09:07:25 crc kubenswrapper[4860]: I1211 09:07:25.306338 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bh99t\" (UniqueName: \"kubernetes.io/projected/1a75cf4b-8d95-43ae-b738-c3fe0346e0f1-kube-api-access-bh99t\") pod \"1a75cf4b-8d95-43ae-b738-c3fe0346e0f1\" (UID: \"1a75cf4b-8d95-43ae-b738-c3fe0346e0f1\") " Dec 11 09:07:25 crc kubenswrapper[4860]: I1211 09:07:25.307919 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1a75cf4b-8d95-43ae-b738-c3fe0346e0f1-utilities" (OuterVolumeSpecName: "utilities") pod "1a75cf4b-8d95-43ae-b738-c3fe0346e0f1" (UID: "1a75cf4b-8d95-43ae-b738-c3fe0346e0f1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 09:07:25 crc kubenswrapper[4860]: I1211 09:07:25.313851 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a75cf4b-8d95-43ae-b738-c3fe0346e0f1-kube-api-access-bh99t" (OuterVolumeSpecName: "kube-api-access-bh99t") pod "1a75cf4b-8d95-43ae-b738-c3fe0346e0f1" (UID: "1a75cf4b-8d95-43ae-b738-c3fe0346e0f1"). InnerVolumeSpecName "kube-api-access-bh99t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 09:07:25 crc kubenswrapper[4860]: I1211 09:07:25.408430 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bh99t\" (UniqueName: \"kubernetes.io/projected/1a75cf4b-8d95-43ae-b738-c3fe0346e0f1-kube-api-access-bh99t\") on node \"crc\" DevicePath \"\"" Dec 11 09:07:25 crc kubenswrapper[4860]: I1211 09:07:25.408466 4860 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1a75cf4b-8d95-43ae-b738-c3fe0346e0f1-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 09:07:25 crc kubenswrapper[4860]: I1211 09:07:25.433836 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1a75cf4b-8d95-43ae-b738-c3fe0346e0f1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1a75cf4b-8d95-43ae-b738-c3fe0346e0f1" (UID: "1a75cf4b-8d95-43ae-b738-c3fe0346e0f1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 09:07:25 crc kubenswrapper[4860]: I1211 09:07:25.510112 4860 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1a75cf4b-8d95-43ae-b738-c3fe0346e0f1-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 09:07:25 crc kubenswrapper[4860]: I1211 09:07:25.811288 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wtznh" event={"ID":"1a75cf4b-8d95-43ae-b738-c3fe0346e0f1","Type":"ContainerDied","Data":"0a7fed6aab76718a90c1551d7cefa4137e625197602100213cfd8b72a091c7f3"} Dec 11 09:07:25 crc kubenswrapper[4860]: I1211 09:07:25.811585 4860 scope.go:117] "RemoveContainer" containerID="2e80a524939e00c7820aba6b79e50d94e27ca6a8688ebc0319683c6dad7d217f" Dec 11 09:07:25 crc kubenswrapper[4860]: I1211 09:07:25.811331 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wtznh" Dec 11 09:07:25 crc kubenswrapper[4860]: I1211 09:07:25.833267 4860 scope.go:117] "RemoveContainer" containerID="5157603512a8bf14cca12e8fed6962249fa644c0b0b632f6e5f0befb0e521c6a" Dec 11 09:07:25 crc kubenswrapper[4860]: I1211 09:07:25.837080 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wtznh"] Dec 11 09:07:25 crc kubenswrapper[4860]: I1211 09:07:25.847240 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-wtznh"] Dec 11 09:07:25 crc kubenswrapper[4860]: I1211 09:07:25.857202 4860 scope.go:117] "RemoveContainer" containerID="f7af66d535caa99c000b5077fecdd281bf9e92bd886d4cc32b18e20e208337ff" Dec 11 09:07:27 crc kubenswrapper[4860]: I1211 09:07:27.591500 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1a75cf4b-8d95-43ae-b738-c3fe0346e0f1" path="/var/lib/kubelet/pods/1a75cf4b-8d95-43ae-b738-c3fe0346e0f1/volumes" Dec 11 09:07:37 crc kubenswrapper[4860]: I1211 09:07:37.604690 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-n96c5/must-gather-6gnzt"] Dec 11 09:07:37 crc kubenswrapper[4860]: E1211 09:07:37.605928 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a75cf4b-8d95-43ae-b738-c3fe0346e0f1" containerName="extract-utilities" Dec 11 09:07:37 crc kubenswrapper[4860]: I1211 09:07:37.605955 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a75cf4b-8d95-43ae-b738-c3fe0346e0f1" containerName="extract-utilities" Dec 11 09:07:37 crc kubenswrapper[4860]: E1211 09:07:37.605978 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a75cf4b-8d95-43ae-b738-c3fe0346e0f1" containerName="registry-server" Dec 11 09:07:37 crc kubenswrapper[4860]: I1211 09:07:37.605991 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a75cf4b-8d95-43ae-b738-c3fe0346e0f1" containerName="registry-server" Dec 11 09:07:37 crc kubenswrapper[4860]: E1211 09:07:37.606044 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a75cf4b-8d95-43ae-b738-c3fe0346e0f1" containerName="extract-content" Dec 11 09:07:37 crc kubenswrapper[4860]: I1211 09:07:37.606058 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a75cf4b-8d95-43ae-b738-c3fe0346e0f1" containerName="extract-content" Dec 11 09:07:37 crc kubenswrapper[4860]: I1211 09:07:37.606441 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a75cf4b-8d95-43ae-b738-c3fe0346e0f1" containerName="registry-server" Dec 11 09:07:37 crc kubenswrapper[4860]: I1211 09:07:37.608611 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-n96c5/must-gather-6gnzt" Dec 11 09:07:37 crc kubenswrapper[4860]: I1211 09:07:37.610499 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-n96c5"/"kube-root-ca.crt" Dec 11 09:07:37 crc kubenswrapper[4860]: I1211 09:07:37.610677 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-n96c5"/"openshift-service-ca.crt" Dec 11 09:07:37 crc kubenswrapper[4860]: I1211 09:07:37.610825 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-n96c5"/"default-dockercfg-tt7x8" Dec 11 09:07:37 crc kubenswrapper[4860]: I1211 09:07:37.638855 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-n96c5/must-gather-6gnzt"] Dec 11 09:07:37 crc kubenswrapper[4860]: I1211 09:07:37.787173 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/11670ddb-0c7a-4ac0-b039-a5254f80b5f8-must-gather-output\") pod \"must-gather-6gnzt\" (UID: \"11670ddb-0c7a-4ac0-b039-a5254f80b5f8\") " pod="openshift-must-gather-n96c5/must-gather-6gnzt" Dec 11 09:07:37 crc kubenswrapper[4860]: I1211 09:07:37.787245 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j9jwp\" (UniqueName: \"kubernetes.io/projected/11670ddb-0c7a-4ac0-b039-a5254f80b5f8-kube-api-access-j9jwp\") pod \"must-gather-6gnzt\" (UID: \"11670ddb-0c7a-4ac0-b039-a5254f80b5f8\") " pod="openshift-must-gather-n96c5/must-gather-6gnzt" Dec 11 09:07:37 crc kubenswrapper[4860]: I1211 09:07:37.889487 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/11670ddb-0c7a-4ac0-b039-a5254f80b5f8-must-gather-output\") pod \"must-gather-6gnzt\" (UID: \"11670ddb-0c7a-4ac0-b039-a5254f80b5f8\") " pod="openshift-must-gather-n96c5/must-gather-6gnzt" Dec 11 09:07:37 crc kubenswrapper[4860]: I1211 09:07:37.889567 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j9jwp\" (UniqueName: \"kubernetes.io/projected/11670ddb-0c7a-4ac0-b039-a5254f80b5f8-kube-api-access-j9jwp\") pod \"must-gather-6gnzt\" (UID: \"11670ddb-0c7a-4ac0-b039-a5254f80b5f8\") " pod="openshift-must-gather-n96c5/must-gather-6gnzt" Dec 11 09:07:37 crc kubenswrapper[4860]: I1211 09:07:37.890131 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/11670ddb-0c7a-4ac0-b039-a5254f80b5f8-must-gather-output\") pod \"must-gather-6gnzt\" (UID: \"11670ddb-0c7a-4ac0-b039-a5254f80b5f8\") " pod="openshift-must-gather-n96c5/must-gather-6gnzt" Dec 11 09:07:37 crc kubenswrapper[4860]: I1211 09:07:37.910090 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j9jwp\" (UniqueName: \"kubernetes.io/projected/11670ddb-0c7a-4ac0-b039-a5254f80b5f8-kube-api-access-j9jwp\") pod \"must-gather-6gnzt\" (UID: \"11670ddb-0c7a-4ac0-b039-a5254f80b5f8\") " pod="openshift-must-gather-n96c5/must-gather-6gnzt" Dec 11 09:07:37 crc kubenswrapper[4860]: I1211 09:07:37.940673 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-n96c5/must-gather-6gnzt" Dec 11 09:07:38 crc kubenswrapper[4860]: I1211 09:07:38.562744 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-n96c5/must-gather-6gnzt"] Dec 11 09:07:38 crc kubenswrapper[4860]: I1211 09:07:38.925265 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-n96c5/must-gather-6gnzt" event={"ID":"11670ddb-0c7a-4ac0-b039-a5254f80b5f8","Type":"ContainerStarted","Data":"fea1bc57dbe21af00dfd9d80609ad18279e622b40086cc7d83280a5b814bafaf"} Dec 11 09:07:39 crc kubenswrapper[4860]: I1211 09:07:39.578901 4860 scope.go:117] "RemoveContainer" containerID="76665d219a809fd146709b7837eb798329ce7fb1c53d5b15c247e7e654985e07" Dec 11 09:07:39 crc kubenswrapper[4860]: E1211 09:07:39.579212 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-99qgp_openshift-machine-config-operator(31c30642-6e60-41b4-a477-0d802424e0aa)\"" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" Dec 11 09:07:46 crc kubenswrapper[4860]: I1211 09:07:46.009236 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-n96c5/must-gather-6gnzt" event={"ID":"11670ddb-0c7a-4ac0-b039-a5254f80b5f8","Type":"ContainerStarted","Data":"c5b8630a7391dc3185f8f8b32bc1a078cd3af6b7c3b10143b59bca15218a1732"} Dec 11 09:07:46 crc kubenswrapper[4860]: I1211 09:07:46.009778 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-n96c5/must-gather-6gnzt" event={"ID":"11670ddb-0c7a-4ac0-b039-a5254f80b5f8","Type":"ContainerStarted","Data":"4cefe6af6c5989c2d95f88a602fdcd910e7ecf41908dc35b81db7ef0bb0d72e7"} Dec 11 09:07:46 crc kubenswrapper[4860]: I1211 09:07:46.030444 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-n96c5/must-gather-6gnzt" podStartSLOduration=3.281299809 podStartE2EDuration="9.030420443s" podCreationTimestamp="2025-12-11 09:07:37 +0000 UTC" firstStartedPulling="2025-12-11 09:07:38.567478328 +0000 UTC m=+3391.295997383" lastFinishedPulling="2025-12-11 09:07:44.316598962 +0000 UTC m=+3397.045118017" observedRunningTime="2025-12-11 09:07:46.023019309 +0000 UTC m=+3398.751538364" watchObservedRunningTime="2025-12-11 09:07:46.030420443 +0000 UTC m=+3398.758939498" Dec 11 09:07:48 crc kubenswrapper[4860]: I1211 09:07:48.761367 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-n96c5/crc-debug-7xzt7"] Dec 11 09:07:48 crc kubenswrapper[4860]: I1211 09:07:48.763134 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-n96c5/crc-debug-7xzt7" Dec 11 09:07:48 crc kubenswrapper[4860]: I1211 09:07:48.783110 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-65t7r\" (UniqueName: \"kubernetes.io/projected/fcd6d214-92c3-48e8-ad46-a1f5e7f6f71b-kube-api-access-65t7r\") pod \"crc-debug-7xzt7\" (UID: \"fcd6d214-92c3-48e8-ad46-a1f5e7f6f71b\") " pod="openshift-must-gather-n96c5/crc-debug-7xzt7" Dec 11 09:07:48 crc kubenswrapper[4860]: I1211 09:07:48.783178 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fcd6d214-92c3-48e8-ad46-a1f5e7f6f71b-host\") pod \"crc-debug-7xzt7\" (UID: \"fcd6d214-92c3-48e8-ad46-a1f5e7f6f71b\") " pod="openshift-must-gather-n96c5/crc-debug-7xzt7" Dec 11 09:07:48 crc kubenswrapper[4860]: I1211 09:07:48.885917 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-65t7r\" (UniqueName: \"kubernetes.io/projected/fcd6d214-92c3-48e8-ad46-a1f5e7f6f71b-kube-api-access-65t7r\") pod \"crc-debug-7xzt7\" (UID: \"fcd6d214-92c3-48e8-ad46-a1f5e7f6f71b\") " pod="openshift-must-gather-n96c5/crc-debug-7xzt7" Dec 11 09:07:48 crc kubenswrapper[4860]: I1211 09:07:48.885969 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fcd6d214-92c3-48e8-ad46-a1f5e7f6f71b-host\") pod \"crc-debug-7xzt7\" (UID: \"fcd6d214-92c3-48e8-ad46-a1f5e7f6f71b\") " pod="openshift-must-gather-n96c5/crc-debug-7xzt7" Dec 11 09:07:48 crc kubenswrapper[4860]: I1211 09:07:48.886077 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fcd6d214-92c3-48e8-ad46-a1f5e7f6f71b-host\") pod \"crc-debug-7xzt7\" (UID: \"fcd6d214-92c3-48e8-ad46-a1f5e7f6f71b\") " pod="openshift-must-gather-n96c5/crc-debug-7xzt7" Dec 11 09:07:48 crc kubenswrapper[4860]: I1211 09:07:48.908845 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-65t7r\" (UniqueName: \"kubernetes.io/projected/fcd6d214-92c3-48e8-ad46-a1f5e7f6f71b-kube-api-access-65t7r\") pod \"crc-debug-7xzt7\" (UID: \"fcd6d214-92c3-48e8-ad46-a1f5e7f6f71b\") " pod="openshift-must-gather-n96c5/crc-debug-7xzt7" Dec 11 09:07:49 crc kubenswrapper[4860]: I1211 09:07:49.080785 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-n96c5/crc-debug-7xzt7" Dec 11 09:07:49 crc kubenswrapper[4860]: W1211 09:07:49.109768 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfcd6d214_92c3_48e8_ad46_a1f5e7f6f71b.slice/crio-8a65190b6fb8aa95a356587f718aa12ee8d0272010af054d4aef0016911fcaba WatchSource:0}: Error finding container 8a65190b6fb8aa95a356587f718aa12ee8d0272010af054d4aef0016911fcaba: Status 404 returned error can't find the container with id 8a65190b6fb8aa95a356587f718aa12ee8d0272010af054d4aef0016911fcaba Dec 11 09:07:50 crc kubenswrapper[4860]: I1211 09:07:50.075293 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-n96c5/crc-debug-7xzt7" event={"ID":"fcd6d214-92c3-48e8-ad46-a1f5e7f6f71b","Type":"ContainerStarted","Data":"8a65190b6fb8aa95a356587f718aa12ee8d0272010af054d4aef0016911fcaba"} Dec 11 09:07:51 crc kubenswrapper[4860]: I1211 09:07:51.580150 4860 scope.go:117] "RemoveContainer" containerID="76665d219a809fd146709b7837eb798329ce7fb1c53d5b15c247e7e654985e07" Dec 11 09:07:51 crc kubenswrapper[4860]: E1211 09:07:51.580740 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-99qgp_openshift-machine-config-operator(31c30642-6e60-41b4-a477-0d802424e0aa)\"" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" Dec 11 09:08:02 crc kubenswrapper[4860]: I1211 09:08:02.224441 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-n96c5/crc-debug-7xzt7" event={"ID":"fcd6d214-92c3-48e8-ad46-a1f5e7f6f71b","Type":"ContainerStarted","Data":"f91e589f8b1373f3ceaf3e69ffa16298ef25dba9f371cd94e520c9c8902d3d0c"} Dec 11 09:08:02 crc kubenswrapper[4860]: I1211 09:08:02.239881 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-n96c5/crc-debug-7xzt7" podStartSLOduration=1.9762620800000001 podStartE2EDuration="14.239861331s" podCreationTimestamp="2025-12-11 09:07:48 +0000 UTC" firstStartedPulling="2025-12-11 09:07:49.11476997 +0000 UTC m=+3401.843289025" lastFinishedPulling="2025-12-11 09:08:01.378369211 +0000 UTC m=+3414.106888276" observedRunningTime="2025-12-11 09:08:02.23733035 +0000 UTC m=+3414.965849405" watchObservedRunningTime="2025-12-11 09:08:02.239861331 +0000 UTC m=+3414.968380396" Dec 11 09:08:06 crc kubenswrapper[4860]: I1211 09:08:06.579184 4860 scope.go:117] "RemoveContainer" containerID="76665d219a809fd146709b7837eb798329ce7fb1c53d5b15c247e7e654985e07" Dec 11 09:08:06 crc kubenswrapper[4860]: E1211 09:08:06.579969 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-99qgp_openshift-machine-config-operator(31c30642-6e60-41b4-a477-0d802424e0aa)\"" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" Dec 11 09:08:17 crc kubenswrapper[4860]: I1211 09:08:17.588251 4860 scope.go:117] "RemoveContainer" containerID="76665d219a809fd146709b7837eb798329ce7fb1c53d5b15c247e7e654985e07" Dec 11 09:08:19 crc kubenswrapper[4860]: I1211 09:08:19.378340 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" event={"ID":"31c30642-6e60-41b4-a477-0d802424e0aa","Type":"ContainerStarted","Data":"8eb6bd97d2b0921d3b9d3592fa7a8bac43db52ba5376a780109acc0b8111e2bf"} Dec 11 09:08:37 crc kubenswrapper[4860]: I1211 09:08:37.692627 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-x5v4z"] Dec 11 09:08:37 crc kubenswrapper[4860]: I1211 09:08:37.696129 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x5v4z" Dec 11 09:08:37 crc kubenswrapper[4860]: I1211 09:08:37.722014 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-x5v4z"] Dec 11 09:08:37 crc kubenswrapper[4860]: I1211 09:08:37.879191 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zks2k\" (UniqueName: \"kubernetes.io/projected/4d3489b0-ae39-422e-a7f0-7969aa384f6f-kube-api-access-zks2k\") pod \"redhat-marketplace-x5v4z\" (UID: \"4d3489b0-ae39-422e-a7f0-7969aa384f6f\") " pod="openshift-marketplace/redhat-marketplace-x5v4z" Dec 11 09:08:37 crc kubenswrapper[4860]: I1211 09:08:37.879711 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d3489b0-ae39-422e-a7f0-7969aa384f6f-catalog-content\") pod \"redhat-marketplace-x5v4z\" (UID: \"4d3489b0-ae39-422e-a7f0-7969aa384f6f\") " pod="openshift-marketplace/redhat-marketplace-x5v4z" Dec 11 09:08:37 crc kubenswrapper[4860]: I1211 09:08:37.879898 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d3489b0-ae39-422e-a7f0-7969aa384f6f-utilities\") pod \"redhat-marketplace-x5v4z\" (UID: \"4d3489b0-ae39-422e-a7f0-7969aa384f6f\") " pod="openshift-marketplace/redhat-marketplace-x5v4z" Dec 11 09:08:37 crc kubenswrapper[4860]: I1211 09:08:37.981171 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d3489b0-ae39-422e-a7f0-7969aa384f6f-catalog-content\") pod \"redhat-marketplace-x5v4z\" (UID: \"4d3489b0-ae39-422e-a7f0-7969aa384f6f\") " pod="openshift-marketplace/redhat-marketplace-x5v4z" Dec 11 09:08:37 crc kubenswrapper[4860]: I1211 09:08:37.981248 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d3489b0-ae39-422e-a7f0-7969aa384f6f-utilities\") pod \"redhat-marketplace-x5v4z\" (UID: \"4d3489b0-ae39-422e-a7f0-7969aa384f6f\") " pod="openshift-marketplace/redhat-marketplace-x5v4z" Dec 11 09:08:37 crc kubenswrapper[4860]: I1211 09:08:37.981292 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zks2k\" (UniqueName: \"kubernetes.io/projected/4d3489b0-ae39-422e-a7f0-7969aa384f6f-kube-api-access-zks2k\") pod \"redhat-marketplace-x5v4z\" (UID: \"4d3489b0-ae39-422e-a7f0-7969aa384f6f\") " pod="openshift-marketplace/redhat-marketplace-x5v4z" Dec 11 09:08:37 crc kubenswrapper[4860]: I1211 09:08:37.982315 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d3489b0-ae39-422e-a7f0-7969aa384f6f-catalog-content\") pod \"redhat-marketplace-x5v4z\" (UID: \"4d3489b0-ae39-422e-a7f0-7969aa384f6f\") " pod="openshift-marketplace/redhat-marketplace-x5v4z" Dec 11 09:08:37 crc kubenswrapper[4860]: I1211 09:08:37.982548 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d3489b0-ae39-422e-a7f0-7969aa384f6f-utilities\") pod \"redhat-marketplace-x5v4z\" (UID: \"4d3489b0-ae39-422e-a7f0-7969aa384f6f\") " pod="openshift-marketplace/redhat-marketplace-x5v4z" Dec 11 09:08:38 crc kubenswrapper[4860]: I1211 09:08:38.011090 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zks2k\" (UniqueName: \"kubernetes.io/projected/4d3489b0-ae39-422e-a7f0-7969aa384f6f-kube-api-access-zks2k\") pod \"redhat-marketplace-x5v4z\" (UID: \"4d3489b0-ae39-422e-a7f0-7969aa384f6f\") " pod="openshift-marketplace/redhat-marketplace-x5v4z" Dec 11 09:08:38 crc kubenswrapper[4860]: I1211 09:08:38.021664 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x5v4z" Dec 11 09:08:38 crc kubenswrapper[4860]: I1211 09:08:38.526622 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-x5v4z"] Dec 11 09:08:38 crc kubenswrapper[4860]: I1211 09:08:38.571665 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x5v4z" event={"ID":"4d3489b0-ae39-422e-a7f0-7969aa384f6f","Type":"ContainerStarted","Data":"bb267d377a2d0e38a62f6a17ea9aec087f93dcf1b78c006449ae9e88a500bda7"} Dec 11 09:08:39 crc kubenswrapper[4860]: I1211 09:08:39.587147 4860 generic.go:334] "Generic (PLEG): container finished" podID="4d3489b0-ae39-422e-a7f0-7969aa384f6f" containerID="cad1c90e0656df477f82e528c3f7d1472ab829672d3031592b9a66cc9a4d58bb" exitCode=0 Dec 11 09:08:39 crc kubenswrapper[4860]: I1211 09:08:39.593310 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x5v4z" event={"ID":"4d3489b0-ae39-422e-a7f0-7969aa384f6f","Type":"ContainerDied","Data":"cad1c90e0656df477f82e528c3f7d1472ab829672d3031592b9a66cc9a4d58bb"} Dec 11 09:08:41 crc kubenswrapper[4860]: I1211 09:08:41.613784 4860 generic.go:334] "Generic (PLEG): container finished" podID="4d3489b0-ae39-422e-a7f0-7969aa384f6f" containerID="9297cdae67d53b32424d70b00183a5905c33a98d2d3d5b9e1d1afe6315d7ba79" exitCode=0 Dec 11 09:08:41 crc kubenswrapper[4860]: I1211 09:08:41.613852 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x5v4z" event={"ID":"4d3489b0-ae39-422e-a7f0-7969aa384f6f","Type":"ContainerDied","Data":"9297cdae67d53b32424d70b00183a5905c33a98d2d3d5b9e1d1afe6315d7ba79"} Dec 11 09:08:42 crc kubenswrapper[4860]: I1211 09:08:42.624254 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x5v4z" event={"ID":"4d3489b0-ae39-422e-a7f0-7969aa384f6f","Type":"ContainerStarted","Data":"26e3c2e1dbdbe9d7a87e448e371b3d36f69407775139d1f8cab613246a6a6df1"} Dec 11 09:08:42 crc kubenswrapper[4860]: I1211 09:08:42.649102 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-x5v4z" podStartSLOduration=2.966814191 podStartE2EDuration="5.649085987s" podCreationTimestamp="2025-12-11 09:08:37 +0000 UTC" firstStartedPulling="2025-12-11 09:08:39.590852371 +0000 UTC m=+3452.319371426" lastFinishedPulling="2025-12-11 09:08:42.273124167 +0000 UTC m=+3455.001643222" observedRunningTime="2025-12-11 09:08:42.642573003 +0000 UTC m=+3455.371092058" watchObservedRunningTime="2025-12-11 09:08:42.649085987 +0000 UTC m=+3455.377605042" Dec 11 09:08:43 crc kubenswrapper[4860]: I1211 09:08:43.635427 4860 generic.go:334] "Generic (PLEG): container finished" podID="fcd6d214-92c3-48e8-ad46-a1f5e7f6f71b" containerID="f91e589f8b1373f3ceaf3e69ffa16298ef25dba9f371cd94e520c9c8902d3d0c" exitCode=0 Dec 11 09:08:43 crc kubenswrapper[4860]: I1211 09:08:43.635519 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-n96c5/crc-debug-7xzt7" event={"ID":"fcd6d214-92c3-48e8-ad46-a1f5e7f6f71b","Type":"ContainerDied","Data":"f91e589f8b1373f3ceaf3e69ffa16298ef25dba9f371cd94e520c9c8902d3d0c"} Dec 11 09:08:44 crc kubenswrapper[4860]: I1211 09:08:44.763823 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-n96c5/crc-debug-7xzt7" Dec 11 09:08:44 crc kubenswrapper[4860]: I1211 09:08:44.798996 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-n96c5/crc-debug-7xzt7"] Dec 11 09:08:44 crc kubenswrapper[4860]: I1211 09:08:44.807107 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-n96c5/crc-debug-7xzt7"] Dec 11 09:08:44 crc kubenswrapper[4860]: I1211 09:08:44.851604 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-65t7r\" (UniqueName: \"kubernetes.io/projected/fcd6d214-92c3-48e8-ad46-a1f5e7f6f71b-kube-api-access-65t7r\") pod \"fcd6d214-92c3-48e8-ad46-a1f5e7f6f71b\" (UID: \"fcd6d214-92c3-48e8-ad46-a1f5e7f6f71b\") " Dec 11 09:08:44 crc kubenswrapper[4860]: I1211 09:08:44.852225 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fcd6d214-92c3-48e8-ad46-a1f5e7f6f71b-host\") pod \"fcd6d214-92c3-48e8-ad46-a1f5e7f6f71b\" (UID: \"fcd6d214-92c3-48e8-ad46-a1f5e7f6f71b\") " Dec 11 09:08:44 crc kubenswrapper[4860]: I1211 09:08:44.852572 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fcd6d214-92c3-48e8-ad46-a1f5e7f6f71b-host" (OuterVolumeSpecName: "host") pod "fcd6d214-92c3-48e8-ad46-a1f5e7f6f71b" (UID: "fcd6d214-92c3-48e8-ad46-a1f5e7f6f71b"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 09:08:44 crc kubenswrapper[4860]: I1211 09:08:44.854634 4860 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fcd6d214-92c3-48e8-ad46-a1f5e7f6f71b-host\") on node \"crc\" DevicePath \"\"" Dec 11 09:08:44 crc kubenswrapper[4860]: I1211 09:08:44.860284 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fcd6d214-92c3-48e8-ad46-a1f5e7f6f71b-kube-api-access-65t7r" (OuterVolumeSpecName: "kube-api-access-65t7r") pod "fcd6d214-92c3-48e8-ad46-a1f5e7f6f71b" (UID: "fcd6d214-92c3-48e8-ad46-a1f5e7f6f71b"). InnerVolumeSpecName "kube-api-access-65t7r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 09:08:44 crc kubenswrapper[4860]: I1211 09:08:44.956396 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-65t7r\" (UniqueName: \"kubernetes.io/projected/fcd6d214-92c3-48e8-ad46-a1f5e7f6f71b-kube-api-access-65t7r\") on node \"crc\" DevicePath \"\"" Dec 11 09:08:45 crc kubenswrapper[4860]: I1211 09:08:45.589440 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fcd6d214-92c3-48e8-ad46-a1f5e7f6f71b" path="/var/lib/kubelet/pods/fcd6d214-92c3-48e8-ad46-a1f5e7f6f71b/volumes" Dec 11 09:08:45 crc kubenswrapper[4860]: I1211 09:08:45.657476 4860 scope.go:117] "RemoveContainer" containerID="f91e589f8b1373f3ceaf3e69ffa16298ef25dba9f371cd94e520c9c8902d3d0c" Dec 11 09:08:45 crc kubenswrapper[4860]: I1211 09:08:45.657703 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-n96c5/crc-debug-7xzt7" Dec 11 09:08:46 crc kubenswrapper[4860]: I1211 09:08:46.010882 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-n96c5/crc-debug-6tpfw"] Dec 11 09:08:46 crc kubenswrapper[4860]: E1211 09:08:46.011908 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fcd6d214-92c3-48e8-ad46-a1f5e7f6f71b" containerName="container-00" Dec 11 09:08:46 crc kubenswrapper[4860]: I1211 09:08:46.011924 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="fcd6d214-92c3-48e8-ad46-a1f5e7f6f71b" containerName="container-00" Dec 11 09:08:46 crc kubenswrapper[4860]: I1211 09:08:46.012148 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="fcd6d214-92c3-48e8-ad46-a1f5e7f6f71b" containerName="container-00" Dec 11 09:08:46 crc kubenswrapper[4860]: I1211 09:08:46.012861 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-n96c5/crc-debug-6tpfw" Dec 11 09:08:46 crc kubenswrapper[4860]: I1211 09:08:46.079312 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/50096672-384b-44d0-88b4-4cc900d4897e-host\") pod \"crc-debug-6tpfw\" (UID: \"50096672-384b-44d0-88b4-4cc900d4897e\") " pod="openshift-must-gather-n96c5/crc-debug-6tpfw" Dec 11 09:08:46 crc kubenswrapper[4860]: I1211 09:08:46.079358 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9f7n8\" (UniqueName: \"kubernetes.io/projected/50096672-384b-44d0-88b4-4cc900d4897e-kube-api-access-9f7n8\") pod \"crc-debug-6tpfw\" (UID: \"50096672-384b-44d0-88b4-4cc900d4897e\") " pod="openshift-must-gather-n96c5/crc-debug-6tpfw" Dec 11 09:08:46 crc kubenswrapper[4860]: I1211 09:08:46.180966 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/50096672-384b-44d0-88b4-4cc900d4897e-host\") pod \"crc-debug-6tpfw\" (UID: \"50096672-384b-44d0-88b4-4cc900d4897e\") " pod="openshift-must-gather-n96c5/crc-debug-6tpfw" Dec 11 09:08:46 crc kubenswrapper[4860]: I1211 09:08:46.181013 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9f7n8\" (UniqueName: \"kubernetes.io/projected/50096672-384b-44d0-88b4-4cc900d4897e-kube-api-access-9f7n8\") pod \"crc-debug-6tpfw\" (UID: \"50096672-384b-44d0-88b4-4cc900d4897e\") " pod="openshift-must-gather-n96c5/crc-debug-6tpfw" Dec 11 09:08:46 crc kubenswrapper[4860]: I1211 09:08:46.181166 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/50096672-384b-44d0-88b4-4cc900d4897e-host\") pod \"crc-debug-6tpfw\" (UID: \"50096672-384b-44d0-88b4-4cc900d4897e\") " pod="openshift-must-gather-n96c5/crc-debug-6tpfw" Dec 11 09:08:46 crc kubenswrapper[4860]: I1211 09:08:46.216411 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9f7n8\" (UniqueName: \"kubernetes.io/projected/50096672-384b-44d0-88b4-4cc900d4897e-kube-api-access-9f7n8\") pod \"crc-debug-6tpfw\" (UID: \"50096672-384b-44d0-88b4-4cc900d4897e\") " pod="openshift-must-gather-n96c5/crc-debug-6tpfw" Dec 11 09:08:46 crc kubenswrapper[4860]: I1211 09:08:46.335450 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-n96c5/crc-debug-6tpfw" Dec 11 09:08:46 crc kubenswrapper[4860]: W1211 09:08:46.370472 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod50096672_384b_44d0_88b4_4cc900d4897e.slice/crio-5984cd9c2bdec316d1b977dc3357d2d901cfdfa3719e6defedacd71a57839b9a WatchSource:0}: Error finding container 5984cd9c2bdec316d1b977dc3357d2d901cfdfa3719e6defedacd71a57839b9a: Status 404 returned error can't find the container with id 5984cd9c2bdec316d1b977dc3357d2d901cfdfa3719e6defedacd71a57839b9a Dec 11 09:08:46 crc kubenswrapper[4860]: I1211 09:08:46.670980 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-n96c5/crc-debug-6tpfw" event={"ID":"50096672-384b-44d0-88b4-4cc900d4897e","Type":"ContainerStarted","Data":"5984cd9c2bdec316d1b977dc3357d2d901cfdfa3719e6defedacd71a57839b9a"} Dec 11 09:08:47 crc kubenswrapper[4860]: I1211 09:08:47.683097 4860 generic.go:334] "Generic (PLEG): container finished" podID="50096672-384b-44d0-88b4-4cc900d4897e" containerID="29ca94a7fa6de4e56c737d88271b4ad02d0623e1249a4176ee84a46232427b93" exitCode=0 Dec 11 09:08:47 crc kubenswrapper[4860]: I1211 09:08:47.683196 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-n96c5/crc-debug-6tpfw" event={"ID":"50096672-384b-44d0-88b4-4cc900d4897e","Type":"ContainerDied","Data":"29ca94a7fa6de4e56c737d88271b4ad02d0623e1249a4176ee84a46232427b93"} Dec 11 09:08:48 crc kubenswrapper[4860]: I1211 09:08:48.021845 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-x5v4z" Dec 11 09:08:48 crc kubenswrapper[4860]: I1211 09:08:48.021926 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-x5v4z" Dec 11 09:08:48 crc kubenswrapper[4860]: I1211 09:08:48.066149 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-x5v4z" Dec 11 09:08:48 crc kubenswrapper[4860]: I1211 09:08:48.265657 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-n96c5/crc-debug-6tpfw"] Dec 11 09:08:48 crc kubenswrapper[4860]: I1211 09:08:48.278972 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-n96c5/crc-debug-6tpfw"] Dec 11 09:08:48 crc kubenswrapper[4860]: I1211 09:08:48.748624 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-x5v4z" Dec 11 09:08:48 crc kubenswrapper[4860]: I1211 09:08:48.789167 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-n96c5/crc-debug-6tpfw" Dec 11 09:08:48 crc kubenswrapper[4860]: I1211 09:08:48.792314 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-x5v4z"] Dec 11 09:08:48 crc kubenswrapper[4860]: I1211 09:08:48.834889 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9f7n8\" (UniqueName: \"kubernetes.io/projected/50096672-384b-44d0-88b4-4cc900d4897e-kube-api-access-9f7n8\") pod \"50096672-384b-44d0-88b4-4cc900d4897e\" (UID: \"50096672-384b-44d0-88b4-4cc900d4897e\") " Dec 11 09:08:48 crc kubenswrapper[4860]: I1211 09:08:48.835033 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/50096672-384b-44d0-88b4-4cc900d4897e-host\") pod \"50096672-384b-44d0-88b4-4cc900d4897e\" (UID: \"50096672-384b-44d0-88b4-4cc900d4897e\") " Dec 11 09:08:48 crc kubenswrapper[4860]: I1211 09:08:48.835083 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/50096672-384b-44d0-88b4-4cc900d4897e-host" (OuterVolumeSpecName: "host") pod "50096672-384b-44d0-88b4-4cc900d4897e" (UID: "50096672-384b-44d0-88b4-4cc900d4897e"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 09:08:48 crc kubenswrapper[4860]: I1211 09:08:48.835429 4860 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/50096672-384b-44d0-88b4-4cc900d4897e-host\") on node \"crc\" DevicePath \"\"" Dec 11 09:08:48 crc kubenswrapper[4860]: I1211 09:08:48.841446 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/50096672-384b-44d0-88b4-4cc900d4897e-kube-api-access-9f7n8" (OuterVolumeSpecName: "kube-api-access-9f7n8") pod "50096672-384b-44d0-88b4-4cc900d4897e" (UID: "50096672-384b-44d0-88b4-4cc900d4897e"). InnerVolumeSpecName "kube-api-access-9f7n8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 09:08:48 crc kubenswrapper[4860]: I1211 09:08:48.937363 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9f7n8\" (UniqueName: \"kubernetes.io/projected/50096672-384b-44d0-88b4-4cc900d4897e-kube-api-access-9f7n8\") on node \"crc\" DevicePath \"\"" Dec 11 09:08:49 crc kubenswrapper[4860]: I1211 09:08:49.478590 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-n96c5/crc-debug-69wsm"] Dec 11 09:08:49 crc kubenswrapper[4860]: E1211 09:08:49.479105 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50096672-384b-44d0-88b4-4cc900d4897e" containerName="container-00" Dec 11 09:08:49 crc kubenswrapper[4860]: I1211 09:08:49.479128 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="50096672-384b-44d0-88b4-4cc900d4897e" containerName="container-00" Dec 11 09:08:49 crc kubenswrapper[4860]: I1211 09:08:49.479338 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="50096672-384b-44d0-88b4-4cc900d4897e" containerName="container-00" Dec 11 09:08:49 crc kubenswrapper[4860]: I1211 09:08:49.480048 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-n96c5/crc-debug-69wsm" Dec 11 09:08:49 crc kubenswrapper[4860]: I1211 09:08:49.548387 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/dbe01adc-96a1-49b3-8541-7032fbc5d8ae-host\") pod \"crc-debug-69wsm\" (UID: \"dbe01adc-96a1-49b3-8541-7032fbc5d8ae\") " pod="openshift-must-gather-n96c5/crc-debug-69wsm" Dec 11 09:08:49 crc kubenswrapper[4860]: I1211 09:08:49.548850 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4tx5r\" (UniqueName: \"kubernetes.io/projected/dbe01adc-96a1-49b3-8541-7032fbc5d8ae-kube-api-access-4tx5r\") pod \"crc-debug-69wsm\" (UID: \"dbe01adc-96a1-49b3-8541-7032fbc5d8ae\") " pod="openshift-must-gather-n96c5/crc-debug-69wsm" Dec 11 09:08:49 crc kubenswrapper[4860]: I1211 09:08:49.589957 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="50096672-384b-44d0-88b4-4cc900d4897e" path="/var/lib/kubelet/pods/50096672-384b-44d0-88b4-4cc900d4897e/volumes" Dec 11 09:08:49 crc kubenswrapper[4860]: I1211 09:08:49.650846 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/dbe01adc-96a1-49b3-8541-7032fbc5d8ae-host\") pod \"crc-debug-69wsm\" (UID: \"dbe01adc-96a1-49b3-8541-7032fbc5d8ae\") " pod="openshift-must-gather-n96c5/crc-debug-69wsm" Dec 11 09:08:49 crc kubenswrapper[4860]: I1211 09:08:49.650905 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4tx5r\" (UniqueName: \"kubernetes.io/projected/dbe01adc-96a1-49b3-8541-7032fbc5d8ae-kube-api-access-4tx5r\") pod \"crc-debug-69wsm\" (UID: \"dbe01adc-96a1-49b3-8541-7032fbc5d8ae\") " pod="openshift-must-gather-n96c5/crc-debug-69wsm" Dec 11 09:08:49 crc kubenswrapper[4860]: I1211 09:08:49.651693 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/dbe01adc-96a1-49b3-8541-7032fbc5d8ae-host\") pod \"crc-debug-69wsm\" (UID: \"dbe01adc-96a1-49b3-8541-7032fbc5d8ae\") " pod="openshift-must-gather-n96c5/crc-debug-69wsm" Dec 11 09:08:49 crc kubenswrapper[4860]: I1211 09:08:49.668507 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4tx5r\" (UniqueName: \"kubernetes.io/projected/dbe01adc-96a1-49b3-8541-7032fbc5d8ae-kube-api-access-4tx5r\") pod \"crc-debug-69wsm\" (UID: \"dbe01adc-96a1-49b3-8541-7032fbc5d8ae\") " pod="openshift-must-gather-n96c5/crc-debug-69wsm" Dec 11 09:08:49 crc kubenswrapper[4860]: I1211 09:08:49.704032 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-n96c5/crc-debug-6tpfw" Dec 11 09:08:49 crc kubenswrapper[4860]: I1211 09:08:49.704428 4860 scope.go:117] "RemoveContainer" containerID="29ca94a7fa6de4e56c737d88271b4ad02d0623e1249a4176ee84a46232427b93" Dec 11 09:08:49 crc kubenswrapper[4860]: I1211 09:08:49.797804 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-n96c5/crc-debug-69wsm" Dec 11 09:08:49 crc kubenswrapper[4860]: W1211 09:08:49.823250 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddbe01adc_96a1_49b3_8541_7032fbc5d8ae.slice/crio-77ea08563bd165ab8da0c40f751691b233160326ab2ac6e4d0c66e3f9b584140 WatchSource:0}: Error finding container 77ea08563bd165ab8da0c40f751691b233160326ab2ac6e4d0c66e3f9b584140: Status 404 returned error can't find the container with id 77ea08563bd165ab8da0c40f751691b233160326ab2ac6e4d0c66e3f9b584140 Dec 11 09:08:50 crc kubenswrapper[4860]: I1211 09:08:50.721658 4860 generic.go:334] "Generic (PLEG): container finished" podID="dbe01adc-96a1-49b3-8541-7032fbc5d8ae" containerID="363faa69bf3223a656b3e3466163cf28ddb34300526467c54b499a8441d7c670" exitCode=0 Dec 11 09:08:50 crc kubenswrapper[4860]: I1211 09:08:50.721680 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-n96c5/crc-debug-69wsm" event={"ID":"dbe01adc-96a1-49b3-8541-7032fbc5d8ae","Type":"ContainerDied","Data":"363faa69bf3223a656b3e3466163cf28ddb34300526467c54b499a8441d7c670"} Dec 11 09:08:50 crc kubenswrapper[4860]: I1211 09:08:50.722010 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-n96c5/crc-debug-69wsm" event={"ID":"dbe01adc-96a1-49b3-8541-7032fbc5d8ae","Type":"ContainerStarted","Data":"77ea08563bd165ab8da0c40f751691b233160326ab2ac6e4d0c66e3f9b584140"} Dec 11 09:08:50 crc kubenswrapper[4860]: I1211 09:08:50.722134 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-x5v4z" podUID="4d3489b0-ae39-422e-a7f0-7969aa384f6f" containerName="registry-server" containerID="cri-o://26e3c2e1dbdbe9d7a87e448e371b3d36f69407775139d1f8cab613246a6a6df1" gracePeriod=2 Dec 11 09:08:50 crc kubenswrapper[4860]: I1211 09:08:50.767568 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-n96c5/crc-debug-69wsm"] Dec 11 09:08:50 crc kubenswrapper[4860]: I1211 09:08:50.777552 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-n96c5/crc-debug-69wsm"] Dec 11 09:08:51 crc kubenswrapper[4860]: I1211 09:08:51.208919 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x5v4z" Dec 11 09:08:51 crc kubenswrapper[4860]: I1211 09:08:51.283552 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d3489b0-ae39-422e-a7f0-7969aa384f6f-utilities\") pod \"4d3489b0-ae39-422e-a7f0-7969aa384f6f\" (UID: \"4d3489b0-ae39-422e-a7f0-7969aa384f6f\") " Dec 11 09:08:51 crc kubenswrapper[4860]: I1211 09:08:51.283771 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zks2k\" (UniqueName: \"kubernetes.io/projected/4d3489b0-ae39-422e-a7f0-7969aa384f6f-kube-api-access-zks2k\") pod \"4d3489b0-ae39-422e-a7f0-7969aa384f6f\" (UID: \"4d3489b0-ae39-422e-a7f0-7969aa384f6f\") " Dec 11 09:08:51 crc kubenswrapper[4860]: I1211 09:08:51.283957 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d3489b0-ae39-422e-a7f0-7969aa384f6f-catalog-content\") pod \"4d3489b0-ae39-422e-a7f0-7969aa384f6f\" (UID: \"4d3489b0-ae39-422e-a7f0-7969aa384f6f\") " Dec 11 09:08:51 crc kubenswrapper[4860]: I1211 09:08:51.285055 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4d3489b0-ae39-422e-a7f0-7969aa384f6f-utilities" (OuterVolumeSpecName: "utilities") pod "4d3489b0-ae39-422e-a7f0-7969aa384f6f" (UID: "4d3489b0-ae39-422e-a7f0-7969aa384f6f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 09:08:51 crc kubenswrapper[4860]: I1211 09:08:51.300814 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d3489b0-ae39-422e-a7f0-7969aa384f6f-kube-api-access-zks2k" (OuterVolumeSpecName: "kube-api-access-zks2k") pod "4d3489b0-ae39-422e-a7f0-7969aa384f6f" (UID: "4d3489b0-ae39-422e-a7f0-7969aa384f6f"). InnerVolumeSpecName "kube-api-access-zks2k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 09:08:51 crc kubenswrapper[4860]: I1211 09:08:51.304575 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4d3489b0-ae39-422e-a7f0-7969aa384f6f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4d3489b0-ae39-422e-a7f0-7969aa384f6f" (UID: "4d3489b0-ae39-422e-a7f0-7969aa384f6f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 09:08:51 crc kubenswrapper[4860]: I1211 09:08:51.386852 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zks2k\" (UniqueName: \"kubernetes.io/projected/4d3489b0-ae39-422e-a7f0-7969aa384f6f-kube-api-access-zks2k\") on node \"crc\" DevicePath \"\"" Dec 11 09:08:51 crc kubenswrapper[4860]: I1211 09:08:51.386909 4860 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d3489b0-ae39-422e-a7f0-7969aa384f6f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 09:08:51 crc kubenswrapper[4860]: I1211 09:08:51.386936 4860 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d3489b0-ae39-422e-a7f0-7969aa384f6f-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 09:08:51 crc kubenswrapper[4860]: I1211 09:08:51.733585 4860 generic.go:334] "Generic (PLEG): container finished" podID="4d3489b0-ae39-422e-a7f0-7969aa384f6f" containerID="26e3c2e1dbdbe9d7a87e448e371b3d36f69407775139d1f8cab613246a6a6df1" exitCode=0 Dec 11 09:08:51 crc kubenswrapper[4860]: I1211 09:08:51.733711 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x5v4z" Dec 11 09:08:51 crc kubenswrapper[4860]: I1211 09:08:51.733698 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x5v4z" event={"ID":"4d3489b0-ae39-422e-a7f0-7969aa384f6f","Type":"ContainerDied","Data":"26e3c2e1dbdbe9d7a87e448e371b3d36f69407775139d1f8cab613246a6a6df1"} Dec 11 09:08:51 crc kubenswrapper[4860]: I1211 09:08:51.734071 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x5v4z" event={"ID":"4d3489b0-ae39-422e-a7f0-7969aa384f6f","Type":"ContainerDied","Data":"bb267d377a2d0e38a62f6a17ea9aec087f93dcf1b78c006449ae9e88a500bda7"} Dec 11 09:08:51 crc kubenswrapper[4860]: I1211 09:08:51.734114 4860 scope.go:117] "RemoveContainer" containerID="26e3c2e1dbdbe9d7a87e448e371b3d36f69407775139d1f8cab613246a6a6df1" Dec 11 09:08:51 crc kubenswrapper[4860]: I1211 09:08:51.804390 4860 scope.go:117] "RemoveContainer" containerID="9297cdae67d53b32424d70b00183a5905c33a98d2d3d5b9e1d1afe6315d7ba79" Dec 11 09:08:51 crc kubenswrapper[4860]: I1211 09:08:51.811613 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-n96c5/crc-debug-69wsm" Dec 11 09:08:51 crc kubenswrapper[4860]: I1211 09:08:51.829396 4860 scope.go:117] "RemoveContainer" containerID="cad1c90e0656df477f82e528c3f7d1472ab829672d3031592b9a66cc9a4d58bb" Dec 11 09:08:51 crc kubenswrapper[4860]: I1211 09:08:51.844931 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-x5v4z"] Dec 11 09:08:51 crc kubenswrapper[4860]: I1211 09:08:51.858004 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-x5v4z"] Dec 11 09:08:51 crc kubenswrapper[4860]: I1211 09:08:51.892071 4860 scope.go:117] "RemoveContainer" containerID="26e3c2e1dbdbe9d7a87e448e371b3d36f69407775139d1f8cab613246a6a6df1" Dec 11 09:08:51 crc kubenswrapper[4860]: E1211 09:08:51.892710 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"26e3c2e1dbdbe9d7a87e448e371b3d36f69407775139d1f8cab613246a6a6df1\": container with ID starting with 26e3c2e1dbdbe9d7a87e448e371b3d36f69407775139d1f8cab613246a6a6df1 not found: ID does not exist" containerID="26e3c2e1dbdbe9d7a87e448e371b3d36f69407775139d1f8cab613246a6a6df1" Dec 11 09:08:51 crc kubenswrapper[4860]: I1211 09:08:51.892887 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"26e3c2e1dbdbe9d7a87e448e371b3d36f69407775139d1f8cab613246a6a6df1"} err="failed to get container status \"26e3c2e1dbdbe9d7a87e448e371b3d36f69407775139d1f8cab613246a6a6df1\": rpc error: code = NotFound desc = could not find container \"26e3c2e1dbdbe9d7a87e448e371b3d36f69407775139d1f8cab613246a6a6df1\": container with ID starting with 26e3c2e1dbdbe9d7a87e448e371b3d36f69407775139d1f8cab613246a6a6df1 not found: ID does not exist" Dec 11 09:08:51 crc kubenswrapper[4860]: I1211 09:08:51.892927 4860 scope.go:117] "RemoveContainer" containerID="9297cdae67d53b32424d70b00183a5905c33a98d2d3d5b9e1d1afe6315d7ba79" Dec 11 09:08:51 crc kubenswrapper[4860]: E1211 09:08:51.893388 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9297cdae67d53b32424d70b00183a5905c33a98d2d3d5b9e1d1afe6315d7ba79\": container with ID starting with 9297cdae67d53b32424d70b00183a5905c33a98d2d3d5b9e1d1afe6315d7ba79 not found: ID does not exist" containerID="9297cdae67d53b32424d70b00183a5905c33a98d2d3d5b9e1d1afe6315d7ba79" Dec 11 09:08:51 crc kubenswrapper[4860]: I1211 09:08:51.893427 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9297cdae67d53b32424d70b00183a5905c33a98d2d3d5b9e1d1afe6315d7ba79"} err="failed to get container status \"9297cdae67d53b32424d70b00183a5905c33a98d2d3d5b9e1d1afe6315d7ba79\": rpc error: code = NotFound desc = could not find container \"9297cdae67d53b32424d70b00183a5905c33a98d2d3d5b9e1d1afe6315d7ba79\": container with ID starting with 9297cdae67d53b32424d70b00183a5905c33a98d2d3d5b9e1d1afe6315d7ba79 not found: ID does not exist" Dec 11 09:08:51 crc kubenswrapper[4860]: I1211 09:08:51.893451 4860 scope.go:117] "RemoveContainer" containerID="cad1c90e0656df477f82e528c3f7d1472ab829672d3031592b9a66cc9a4d58bb" Dec 11 09:08:51 crc kubenswrapper[4860]: E1211 09:08:51.893719 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cad1c90e0656df477f82e528c3f7d1472ab829672d3031592b9a66cc9a4d58bb\": container with ID starting with cad1c90e0656df477f82e528c3f7d1472ab829672d3031592b9a66cc9a4d58bb not found: ID does not exist" containerID="cad1c90e0656df477f82e528c3f7d1472ab829672d3031592b9a66cc9a4d58bb" Dec 11 09:08:51 crc kubenswrapper[4860]: I1211 09:08:51.893754 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cad1c90e0656df477f82e528c3f7d1472ab829672d3031592b9a66cc9a4d58bb"} err="failed to get container status \"cad1c90e0656df477f82e528c3f7d1472ab829672d3031592b9a66cc9a4d58bb\": rpc error: code = NotFound desc = could not find container \"cad1c90e0656df477f82e528c3f7d1472ab829672d3031592b9a66cc9a4d58bb\": container with ID starting with cad1c90e0656df477f82e528c3f7d1472ab829672d3031592b9a66cc9a4d58bb not found: ID does not exist" Dec 11 09:08:51 crc kubenswrapper[4860]: I1211 09:08:51.897753 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/dbe01adc-96a1-49b3-8541-7032fbc5d8ae-host\") pod \"dbe01adc-96a1-49b3-8541-7032fbc5d8ae\" (UID: \"dbe01adc-96a1-49b3-8541-7032fbc5d8ae\") " Dec 11 09:08:51 crc kubenswrapper[4860]: I1211 09:08:51.897847 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/dbe01adc-96a1-49b3-8541-7032fbc5d8ae-host" (OuterVolumeSpecName: "host") pod "dbe01adc-96a1-49b3-8541-7032fbc5d8ae" (UID: "dbe01adc-96a1-49b3-8541-7032fbc5d8ae"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 09:08:51 crc kubenswrapper[4860]: I1211 09:08:51.897893 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4tx5r\" (UniqueName: \"kubernetes.io/projected/dbe01adc-96a1-49b3-8541-7032fbc5d8ae-kube-api-access-4tx5r\") pod \"dbe01adc-96a1-49b3-8541-7032fbc5d8ae\" (UID: \"dbe01adc-96a1-49b3-8541-7032fbc5d8ae\") " Dec 11 09:08:51 crc kubenswrapper[4860]: I1211 09:08:51.898890 4860 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/dbe01adc-96a1-49b3-8541-7032fbc5d8ae-host\") on node \"crc\" DevicePath \"\"" Dec 11 09:08:51 crc kubenswrapper[4860]: I1211 09:08:51.902124 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dbe01adc-96a1-49b3-8541-7032fbc5d8ae-kube-api-access-4tx5r" (OuterVolumeSpecName: "kube-api-access-4tx5r") pod "dbe01adc-96a1-49b3-8541-7032fbc5d8ae" (UID: "dbe01adc-96a1-49b3-8541-7032fbc5d8ae"). InnerVolumeSpecName "kube-api-access-4tx5r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 09:08:52 crc kubenswrapper[4860]: I1211 09:08:52.001350 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4tx5r\" (UniqueName: \"kubernetes.io/projected/dbe01adc-96a1-49b3-8541-7032fbc5d8ae-kube-api-access-4tx5r\") on node \"crc\" DevicePath \"\"" Dec 11 09:08:52 crc kubenswrapper[4860]: I1211 09:08:52.747182 4860 scope.go:117] "RemoveContainer" containerID="363faa69bf3223a656b3e3466163cf28ddb34300526467c54b499a8441d7c670" Dec 11 09:08:52 crc kubenswrapper[4860]: I1211 09:08:52.747210 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-n96c5/crc-debug-69wsm" Dec 11 09:08:53 crc kubenswrapper[4860]: I1211 09:08:53.589888 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4d3489b0-ae39-422e-a7f0-7969aa384f6f" path="/var/lib/kubelet/pods/4d3489b0-ae39-422e-a7f0-7969aa384f6f/volumes" Dec 11 09:08:53 crc kubenswrapper[4860]: I1211 09:08:53.590543 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dbe01adc-96a1-49b3-8541-7032fbc5d8ae" path="/var/lib/kubelet/pods/dbe01adc-96a1-49b3-8541-7032fbc5d8ae/volumes" Dec 11 09:09:05 crc kubenswrapper[4860]: I1211 09:09:05.881846 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-6bbf49f946-sw5n4_4a941157-652e-4efa-9165-ba55161bfb51/barbican-api/0.log" Dec 11 09:09:05 crc kubenswrapper[4860]: I1211 09:09:05.946425 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-6bbf49f946-sw5n4_4a941157-652e-4efa-9165-ba55161bfb51/barbican-api-log/0.log" Dec 11 09:09:06 crc kubenswrapper[4860]: I1211 09:09:06.079014 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-684c9ddd98-t29bd_d8954df2-1f06-4bbe-bb33-a5ac8fdf9675/barbican-keystone-listener/0.log" Dec 11 09:09:06 crc kubenswrapper[4860]: I1211 09:09:06.129448 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-684c9ddd98-t29bd_d8954df2-1f06-4bbe-bb33-a5ac8fdf9675/barbican-keystone-listener-log/0.log" Dec 11 09:09:06 crc kubenswrapper[4860]: I1211 09:09:06.238423 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-568d665c75-4vptt_a7e1def2-f417-40f5-a7dc-c63148858073/barbican-worker/0.log" Dec 11 09:09:06 crc kubenswrapper[4860]: I1211 09:09:06.265124 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-568d665c75-4vptt_a7e1def2-f417-40f5-a7dc-c63148858073/barbican-worker-log/0.log" Dec 11 09:09:06 crc kubenswrapper[4860]: I1211 09:09:06.385181 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-tm5lx_aa177563-2177-4361-b45a-7ffe6f88da96/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Dec 11 09:09:06 crc kubenswrapper[4860]: I1211 09:09:06.537859 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_f0d70280-9b14-4d9d-9b7a-7818f388c4fc/ceilometer-central-agent/0.log" Dec 11 09:09:06 crc kubenswrapper[4860]: I1211 09:09:06.606058 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_f0d70280-9b14-4d9d-9b7a-7818f388c4fc/ceilometer-notification-agent/0.log" Dec 11 09:09:06 crc kubenswrapper[4860]: I1211 09:09:06.648268 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_f0d70280-9b14-4d9d-9b7a-7818f388c4fc/proxy-httpd/0.log" Dec 11 09:09:06 crc kubenswrapper[4860]: I1211 09:09:06.752482 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_f0d70280-9b14-4d9d-9b7a-7818f388c4fc/sg-core/0.log" Dec 11 09:09:06 crc kubenswrapper[4860]: I1211 09:09:06.846770 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_8a6d65b6-5393-4edc-a9c3-3660c8860fd1/cinder-api/0.log" Dec 11 09:09:06 crc kubenswrapper[4860]: I1211 09:09:06.878954 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_8a6d65b6-5393-4edc-a9c3-3660c8860fd1/cinder-api-log/0.log" Dec 11 09:09:07 crc kubenswrapper[4860]: I1211 09:09:07.003611 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_5dc4bc3d-458e-4251-81f7-68b5b47a7275/cinder-scheduler/0.log" Dec 11 09:09:07 crc kubenswrapper[4860]: I1211 09:09:07.069342 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_5dc4bc3d-458e-4251-81f7-68b5b47a7275/probe/0.log" Dec 11 09:09:07 crc kubenswrapper[4860]: I1211 09:09:07.231246 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-7zg5t_155181af-f5ab-41fb-9c0f-820c03b91b23/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 11 09:09:07 crc kubenswrapper[4860]: I1211 09:09:07.274902 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-bv6xw_2651e591-78fd-410d-b8a0-bcb0da2066bb/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 11 09:09:07 crc kubenswrapper[4860]: I1211 09:09:07.430422 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-8c6f6df99-zx6zh_2e567b4f-94b9-4662-825a-045e124c7948/init/0.log" Dec 11 09:09:07 crc kubenswrapper[4860]: I1211 09:09:07.611786 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-8c6f6df99-zx6zh_2e567b4f-94b9-4662-825a-045e124c7948/dnsmasq-dns/0.log" Dec 11 09:09:07 crc kubenswrapper[4860]: I1211 09:09:07.620635 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-8c6f6df99-zx6zh_2e567b4f-94b9-4662-825a-045e124c7948/init/0.log" Dec 11 09:09:07 crc kubenswrapper[4860]: I1211 09:09:07.676896 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-zvzpf_1fb47321-3b60-4961-abf2-87ee1d57a58a/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Dec 11 09:09:07 crc kubenswrapper[4860]: I1211 09:09:07.825044 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_f51c05dd-1881-4644-9a40-7e6e2174c591/glance-httpd/0.log" Dec 11 09:09:07 crc kubenswrapper[4860]: I1211 09:09:07.858412 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_f51c05dd-1881-4644-9a40-7e6e2174c591/glance-log/0.log" Dec 11 09:09:08 crc kubenswrapper[4860]: I1211 09:09:08.044896 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_b7bf9af0-6051-4d90-8918-9fd00cb869cd/glance-log/0.log" Dec 11 09:09:08 crc kubenswrapper[4860]: I1211 09:09:08.045537 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_b7bf9af0-6051-4d90-8918-9fd00cb869cd/glance-httpd/0.log" Dec 11 09:09:08 crc kubenswrapper[4860]: I1211 09:09:08.261233 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-6b4697fc68-vnd26_78b0f75e-d447-4541-be11-9afe4549d2e1/horizon/0.log" Dec 11 09:09:08 crc kubenswrapper[4860]: I1211 09:09:08.391447 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-lwc56_ddaa83e7-d6db-4169-abdb-226ec9f50204/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Dec 11 09:09:08 crc kubenswrapper[4860]: I1211 09:09:08.532920 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-6b4697fc68-vnd26_78b0f75e-d447-4541-be11-9afe4549d2e1/horizon-log/0.log" Dec 11 09:09:08 crc kubenswrapper[4860]: I1211 09:09:08.579448 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-j8pcd_f9f3c692-bca8-49fb-a602-cc747b566849/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 11 09:09:08 crc kubenswrapper[4860]: I1211 09:09:08.801170 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29424061-gq8tl_8fb3c7c5-8c06-4e05-ad79-11f0671e16da/keystone-cron/0.log" Dec 11 09:09:08 crc kubenswrapper[4860]: I1211 09:09:08.867689 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-748d5bd978-4sz6v_e208d74e-f50f-443c-aa2d-eccdb584413d/keystone-api/0.log" Dec 11 09:09:09 crc kubenswrapper[4860]: I1211 09:09:09.046516 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_eceff086-19de-4aed-8ac3-5242d9b949ab/kube-state-metrics/0.log" Dec 11 09:09:09 crc kubenswrapper[4860]: I1211 09:09:09.104788 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-5z2qw_b0eed689-a827-47b6-81ba-91e76db6f35b/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Dec 11 09:09:09 crc kubenswrapper[4860]: I1211 09:09:09.501725 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-5fccb6555c-xwc9b_7739fee0-ccf5-49e9-bc21-4acd841daf55/neutron-api/0.log" Dec 11 09:09:09 crc kubenswrapper[4860]: I1211 09:09:09.505983 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-5fccb6555c-xwc9b_7739fee0-ccf5-49e9-bc21-4acd841daf55/neutron-httpd/0.log" Dec 11 09:09:09 crc kubenswrapper[4860]: I1211 09:09:09.824355 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-gfrj6_70691461-ffb3-4a01-aa11-bbe665e77b35/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Dec 11 09:09:10 crc kubenswrapper[4860]: I1211 09:09:10.248625 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_581da06a-7d8d-48dd-a5d9-d95ec1539a9f/nova-cell0-conductor-conductor/0.log" Dec 11 09:09:10 crc kubenswrapper[4860]: I1211 09:09:10.333656 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_fcbd0284-56fd-4cb0-895e-78e6a571868a/nova-api-log/0.log" Dec 11 09:09:10 crc kubenswrapper[4860]: I1211 09:09:10.554064 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_773668a5-7449-4ced-b8d0-509faab0eabb/nova-cell1-conductor-conductor/0.log" Dec 11 09:09:10 crc kubenswrapper[4860]: I1211 09:09:10.556652 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_fcbd0284-56fd-4cb0-895e-78e6a571868a/nova-api-api/0.log" Dec 11 09:09:10 crc kubenswrapper[4860]: I1211 09:09:10.599307 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_f4725b54-31da-4b85-a4f7-200b1d6a3069/nova-cell1-novncproxy-novncproxy/0.log" Dec 11 09:09:10 crc kubenswrapper[4860]: I1211 09:09:10.823199 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-x42ls_fac07a15-8332-4374-aaca-6577650e3183/nova-edpm-deployment-openstack-edpm-ipam/0.log" Dec 11 09:09:10 crc kubenswrapper[4860]: I1211 09:09:10.944001 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_4b75f4be-3810-4273-a18e-381ab6c6ef2b/nova-metadata-log/0.log" Dec 11 09:09:11 crc kubenswrapper[4860]: I1211 09:09:11.219035 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_08a3b710-84c1-4bbc-9547-a350c0d76318/nova-scheduler-scheduler/0.log" Dec 11 09:09:11 crc kubenswrapper[4860]: I1211 09:09:11.326301 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_a0838f24-14b5-4018-ad26-8cd0394b4488/mysql-bootstrap/0.log" Dec 11 09:09:11 crc kubenswrapper[4860]: I1211 09:09:11.482426 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_a0838f24-14b5-4018-ad26-8cd0394b4488/galera/0.log" Dec 11 09:09:11 crc kubenswrapper[4860]: I1211 09:09:11.520949 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_a0838f24-14b5-4018-ad26-8cd0394b4488/mysql-bootstrap/0.log" Dec 11 09:09:11 crc kubenswrapper[4860]: I1211 09:09:11.707142 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_c0d5e719-9d28-4e31-9349-f882c4b72567/mysql-bootstrap/0.log" Dec 11 09:09:11 crc kubenswrapper[4860]: I1211 09:09:11.915422 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_c0d5e719-9d28-4e31-9349-f882c4b72567/mysql-bootstrap/0.log" Dec 11 09:09:11 crc kubenswrapper[4860]: I1211 09:09:11.919148 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_c0d5e719-9d28-4e31-9349-f882c4b72567/galera/0.log" Dec 11 09:09:12 crc kubenswrapper[4860]: I1211 09:09:12.091925 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_6f6e1905-c5e6-495b-9076-fa5edd9a01ac/openstackclient/0.log" Dec 11 09:09:12 crc kubenswrapper[4860]: I1211 09:09:12.125772 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_4b75f4be-3810-4273-a18e-381ab6c6ef2b/nova-metadata-metadata/0.log" Dec 11 09:09:12 crc kubenswrapper[4860]: I1211 09:09:12.298042 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-dbdtz_e834572f-9650-45d3-b978-c39141b44b74/ovn-controller/0.log" Dec 11 09:09:12 crc kubenswrapper[4860]: I1211 09:09:12.320946 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-lgzmd_06395cef-744f-402e-844e-275a5025b81f/openstack-network-exporter/0.log" Dec 11 09:09:12 crc kubenswrapper[4860]: I1211 09:09:12.527564 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-q86ch_e73beb7d-c40c-491b-99f8-fabe037df9ba/ovsdb-server-init/0.log" Dec 11 09:09:12 crc kubenswrapper[4860]: I1211 09:09:12.684034 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-q86ch_e73beb7d-c40c-491b-99f8-fabe037df9ba/ovsdb-server-init/0.log" Dec 11 09:09:12 crc kubenswrapper[4860]: I1211 09:09:12.762762 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-q86ch_e73beb7d-c40c-491b-99f8-fabe037df9ba/ovs-vswitchd/0.log" Dec 11 09:09:13 crc kubenswrapper[4860]: I1211 09:09:13.054998 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-q86ch_e73beb7d-c40c-491b-99f8-fabe037df9ba/ovsdb-server/0.log" Dec 11 09:09:13 crc kubenswrapper[4860]: I1211 09:09:13.110769 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-8ptks_c2dda62c-b6df-481f-a388-a6bfc3b02c9e/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Dec 11 09:09:13 crc kubenswrapper[4860]: I1211 09:09:13.130252 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_ad53aa7d-2d02-4201-a90b-05bc8cca2f68/openstack-network-exporter/0.log" Dec 11 09:09:13 crc kubenswrapper[4860]: I1211 09:09:13.300758 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_ad53aa7d-2d02-4201-a90b-05bc8cca2f68/ovn-northd/0.log" Dec 11 09:09:13 crc kubenswrapper[4860]: I1211 09:09:13.309243 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_c017f36f-48a3-4919-ac22-6a85a5c73530/openstack-network-exporter/0.log" Dec 11 09:09:13 crc kubenswrapper[4860]: I1211 09:09:13.373483 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_c017f36f-48a3-4919-ac22-6a85a5c73530/ovsdbserver-nb/0.log" Dec 11 09:09:13 crc kubenswrapper[4860]: I1211 09:09:13.508464 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_9657082d-ff9c-4a54-8b4d-54e6ea411590/ovsdbserver-sb/0.log" Dec 11 09:09:13 crc kubenswrapper[4860]: I1211 09:09:13.544294 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_9657082d-ff9c-4a54-8b4d-54e6ea411590/openstack-network-exporter/0.log" Dec 11 09:09:13 crc kubenswrapper[4860]: I1211 09:09:13.733439 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-55ccdb6bbd-grdt6_78683923-6a60-4a9a-bb81-51ae61f55376/placement-api/0.log" Dec 11 09:09:13 crc kubenswrapper[4860]: I1211 09:09:13.857363 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-55ccdb6bbd-grdt6_78683923-6a60-4a9a-bb81-51ae61f55376/placement-log/0.log" Dec 11 09:09:13 crc kubenswrapper[4860]: I1211 09:09:13.884670 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_9cfe9bd2-a1fe-4cae-8f48-c007bfd95961/setup-container/0.log" Dec 11 09:09:14 crc kubenswrapper[4860]: I1211 09:09:14.339343 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_9cfe9bd2-a1fe-4cae-8f48-c007bfd95961/setup-container/0.log" Dec 11 09:09:14 crc kubenswrapper[4860]: I1211 09:09:14.506910 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_a1d839a1-4174-4abc-9e3d-5579603687e6/setup-container/0.log" Dec 11 09:09:14 crc kubenswrapper[4860]: I1211 09:09:14.519789 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_9cfe9bd2-a1fe-4cae-8f48-c007bfd95961/rabbitmq/0.log" Dec 11 09:09:14 crc kubenswrapper[4860]: I1211 09:09:14.713867 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_a1d839a1-4174-4abc-9e3d-5579603687e6/setup-container/0.log" Dec 11 09:09:14 crc kubenswrapper[4860]: I1211 09:09:14.727512 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_a1d839a1-4174-4abc-9e3d-5579603687e6/rabbitmq/0.log" Dec 11 09:09:14 crc kubenswrapper[4860]: I1211 09:09:14.736309 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-g9pmk_07a4b9f9-a1a7-416d-89a5-e8a57c2484cc/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 11 09:09:14 crc kubenswrapper[4860]: I1211 09:09:14.888347 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-zb8rn_a84f5e60-25bb-4412-a38c-fa4e7a32c177/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Dec 11 09:09:15 crc kubenswrapper[4860]: I1211 09:09:15.069113 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-2vqvr_b0ccaf6b-d322-440b-87c4-560b61547f52/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Dec 11 09:09:15 crc kubenswrapper[4860]: I1211 09:09:15.238095 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-rl5hl_6a0f9071-e3f7-43f2-a0ec-8ab22fd9c785/ssh-known-hosts-edpm-deployment/0.log" Dec 11 09:09:15 crc kubenswrapper[4860]: I1211 09:09:15.270608 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-c6n8q_bc86a63a-f36d-4893-a55e-df029aa586ea/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 11 09:09:15 crc kubenswrapper[4860]: I1211 09:09:15.546497 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-865bbb47ff-66xpd_ca8cd60f-8a69-44f4-9512-387dd82af334/proxy-server/0.log" Dec 11 09:09:15 crc kubenswrapper[4860]: I1211 09:09:15.580532 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-865bbb47ff-66xpd_ca8cd60f-8a69-44f4-9512-387dd82af334/proxy-httpd/0.log" Dec 11 09:09:15 crc kubenswrapper[4860]: I1211 09:09:15.725259 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-bphmx_79bcab38-b87b-44ff-ba2b-6027fd20719d/swift-ring-rebalance/0.log" Dec 11 09:09:15 crc kubenswrapper[4860]: I1211 09:09:15.784983 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a672e32c-a924-4080-a7d1-bf7012b6725b/account-auditor/0.log" Dec 11 09:09:15 crc kubenswrapper[4860]: I1211 09:09:15.875812 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a672e32c-a924-4080-a7d1-bf7012b6725b/account-reaper/0.log" Dec 11 09:09:15 crc kubenswrapper[4860]: I1211 09:09:15.983016 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a672e32c-a924-4080-a7d1-bf7012b6725b/account-replicator/0.log" Dec 11 09:09:16 crc kubenswrapper[4860]: I1211 09:09:16.042997 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a672e32c-a924-4080-a7d1-bf7012b6725b/account-server/0.log" Dec 11 09:09:16 crc kubenswrapper[4860]: I1211 09:09:16.147816 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a672e32c-a924-4080-a7d1-bf7012b6725b/container-replicator/0.log" Dec 11 09:09:16 crc kubenswrapper[4860]: I1211 09:09:16.152765 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a672e32c-a924-4080-a7d1-bf7012b6725b/container-auditor/0.log" Dec 11 09:09:16 crc kubenswrapper[4860]: I1211 09:09:16.198924 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a672e32c-a924-4080-a7d1-bf7012b6725b/container-server/0.log" Dec 11 09:09:16 crc kubenswrapper[4860]: I1211 09:09:16.267630 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a672e32c-a924-4080-a7d1-bf7012b6725b/container-updater/0.log" Dec 11 09:09:16 crc kubenswrapper[4860]: I1211 09:09:16.361139 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a672e32c-a924-4080-a7d1-bf7012b6725b/object-auditor/0.log" Dec 11 09:09:16 crc kubenswrapper[4860]: I1211 09:09:16.427270 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a672e32c-a924-4080-a7d1-bf7012b6725b/object-expirer/0.log" Dec 11 09:09:16 crc kubenswrapper[4860]: I1211 09:09:16.445118 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a672e32c-a924-4080-a7d1-bf7012b6725b/object-replicator/0.log" Dec 11 09:09:16 crc kubenswrapper[4860]: I1211 09:09:16.496868 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a672e32c-a924-4080-a7d1-bf7012b6725b/object-server/0.log" Dec 11 09:09:16 crc kubenswrapper[4860]: I1211 09:09:16.545121 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a672e32c-a924-4080-a7d1-bf7012b6725b/object-updater/0.log" Dec 11 09:09:16 crc kubenswrapper[4860]: I1211 09:09:16.666680 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a672e32c-a924-4080-a7d1-bf7012b6725b/rsync/0.log" Dec 11 09:09:16 crc kubenswrapper[4860]: I1211 09:09:16.672182 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a672e32c-a924-4080-a7d1-bf7012b6725b/swift-recon-cron/0.log" Dec 11 09:09:16 crc kubenswrapper[4860]: I1211 09:09:16.863516 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-wmwhw_beb319f7-9b48-413d-9561-e6c0dc62238f/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Dec 11 09:09:16 crc kubenswrapper[4860]: I1211 09:09:16.992142 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_3b659d05-76c9-4b04-a05e-a7dd488e9852/tempest-tests-tempest-tests-runner/0.log" Dec 11 09:09:17 crc kubenswrapper[4860]: I1211 09:09:17.072807 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_b658df4a-886e-49de-b953-87539b28329b/test-operator-logs-container/0.log" Dec 11 09:09:17 crc kubenswrapper[4860]: I1211 09:09:17.238712 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-ljr6l_27ee41d4-1afd-4700-828c-7388caf498c5/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 11 09:09:23 crc kubenswrapper[4860]: I1211 09:09:23.548723 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_1752934c-842c-45b6-88ec-0fc804374532/memcached/0.log" Dec 11 09:09:42 crc kubenswrapper[4860]: I1211 09:09:42.085691 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_941ec9b01b7a72127ba64306553d6766ab71ba59f32f9ca01c76932a62n7v8f_bf285af3-cbc9-4e8f-bb3a-bc1d16a3ea54/util/0.log" Dec 11 09:09:42 crc kubenswrapper[4860]: I1211 09:09:42.298262 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_941ec9b01b7a72127ba64306553d6766ab71ba59f32f9ca01c76932a62n7v8f_bf285af3-cbc9-4e8f-bb3a-bc1d16a3ea54/util/0.log" Dec 11 09:09:42 crc kubenswrapper[4860]: I1211 09:09:42.299463 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_941ec9b01b7a72127ba64306553d6766ab71ba59f32f9ca01c76932a62n7v8f_bf285af3-cbc9-4e8f-bb3a-bc1d16a3ea54/pull/0.log" Dec 11 09:09:42 crc kubenswrapper[4860]: I1211 09:09:42.299460 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_941ec9b01b7a72127ba64306553d6766ab71ba59f32f9ca01c76932a62n7v8f_bf285af3-cbc9-4e8f-bb3a-bc1d16a3ea54/pull/0.log" Dec 11 09:09:42 crc kubenswrapper[4860]: I1211 09:09:42.510126 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_941ec9b01b7a72127ba64306553d6766ab71ba59f32f9ca01c76932a62n7v8f_bf285af3-cbc9-4e8f-bb3a-bc1d16a3ea54/util/0.log" Dec 11 09:09:42 crc kubenswrapper[4860]: I1211 09:09:42.538828 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_941ec9b01b7a72127ba64306553d6766ab71ba59f32f9ca01c76932a62n7v8f_bf285af3-cbc9-4e8f-bb3a-bc1d16a3ea54/extract/0.log" Dec 11 09:09:42 crc kubenswrapper[4860]: I1211 09:09:42.553977 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_941ec9b01b7a72127ba64306553d6766ab71ba59f32f9ca01c76932a62n7v8f_bf285af3-cbc9-4e8f-bb3a-bc1d16a3ea54/pull/0.log" Dec 11 09:09:42 crc kubenswrapper[4860]: I1211 09:09:42.680951 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-bshq2_64f09777-fca3-412f-98f8-5cd189cb9fbd/kube-rbac-proxy/0.log" Dec 11 09:09:42 crc kubenswrapper[4860]: I1211 09:09:42.779793 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-bshq2_64f09777-fca3-412f-98f8-5cd189cb9fbd/manager/0.log" Dec 11 09:09:42 crc kubenswrapper[4860]: I1211 09:09:42.818039 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6c677c69b-zrrf6_323ad77f-a0d6-461d-b483-2ffd74ebb9b7/kube-rbac-proxy/0.log" Dec 11 09:09:43 crc kubenswrapper[4860]: I1211 09:09:43.027043 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6c677c69b-zrrf6_323ad77f-a0d6-461d-b483-2ffd74ebb9b7/manager/0.log" Dec 11 09:09:43 crc kubenswrapper[4860]: I1211 09:09:43.041802 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-697fb699cf-6t74n_a3bf483c-6438-4b7e-abb8-447b9b146644/manager/0.log" Dec 11 09:09:43 crc kubenswrapper[4860]: I1211 09:09:43.060173 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-697fb699cf-6t74n_a3bf483c-6438-4b7e-abb8-447b9b146644/kube-rbac-proxy/0.log" Dec 11 09:09:43 crc kubenswrapper[4860]: I1211 09:09:43.222052 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-5697bb5779-kgzwt_e00c360d-20a4-4e2e-81ea-e4ffa1bc62fc/kube-rbac-proxy/0.log" Dec 11 09:09:43 crc kubenswrapper[4860]: I1211 09:09:43.310273 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-5697bb5779-kgzwt_e00c360d-20a4-4e2e-81ea-e4ffa1bc62fc/manager/0.log" Dec 11 09:09:43 crc kubenswrapper[4860]: I1211 09:09:43.441870 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-lrhhk_505d980f-cfcb-42b2-876e-a4730abe7ea6/kube-rbac-proxy/0.log" Dec 11 09:09:43 crc kubenswrapper[4860]: I1211 09:09:43.457189 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-lrhhk_505d980f-cfcb-42b2-876e-a4730abe7ea6/manager/0.log" Dec 11 09:09:43 crc kubenswrapper[4860]: I1211 09:09:43.541340 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-65fv4_bb231e77-e19b-41ed-bf7a-ae354e1089ec/kube-rbac-proxy/0.log" Dec 11 09:09:43 crc kubenswrapper[4860]: I1211 09:09:43.662925 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-65fv4_bb231e77-e19b-41ed-bf7a-ae354e1089ec/manager/0.log" Dec 11 09:09:43 crc kubenswrapper[4860]: I1211 09:09:43.737324 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-78d48bff9d-sq5gd_7f1d9cd2-1e0c-4c4d-86a6-b91a38eadba2/kube-rbac-proxy/0.log" Dec 11 09:09:43 crc kubenswrapper[4860]: I1211 09:09:43.967904 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-78d48bff9d-sq5gd_7f1d9cd2-1e0c-4c4d-86a6-b91a38eadba2/manager/0.log" Dec 11 09:09:43 crc kubenswrapper[4860]: I1211 09:09:43.993634 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-967d97867-5wvld_91bb3ff9-fd19-4c79-bac4-621cb114c783/manager/0.log" Dec 11 09:09:43 crc kubenswrapper[4860]: I1211 09:09:43.996587 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-967d97867-5wvld_91bb3ff9-fd19-4c79-bac4-621cb114c783/kube-rbac-proxy/0.log" Dec 11 09:09:44 crc kubenswrapper[4860]: I1211 09:09:44.231360 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-6bpb9_d5e0079f-f500-407f-b06f-a610a19c741b/manager/0.log" Dec 11 09:09:44 crc kubenswrapper[4860]: I1211 09:09:44.239361 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-6bpb9_d5e0079f-f500-407f-b06f-a610a19c741b/kube-rbac-proxy/0.log" Dec 11 09:09:44 crc kubenswrapper[4860]: I1211 09:09:44.347418 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5b5fd79c9c-lpm8h_aae650b9-dcbc-482b-b49f-0a45651bb4b4/kube-rbac-proxy/0.log" Dec 11 09:09:44 crc kubenswrapper[4860]: I1211 09:09:44.446222 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5b5fd79c9c-lpm8h_aae650b9-dcbc-482b-b49f-0a45651bb4b4/manager/0.log" Dec 11 09:09:44 crc kubenswrapper[4860]: I1211 09:09:44.468928 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-79c8c4686c-4tw58_5a589b13-7f29-40ae-9804-7c5c2bc3d745/kube-rbac-proxy/0.log" Dec 11 09:09:44 crc kubenswrapper[4860]: I1211 09:09:44.634207 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-79c8c4686c-4tw58_5a589b13-7f29-40ae-9804-7c5c2bc3d745/manager/0.log" Dec 11 09:09:44 crc kubenswrapper[4860]: I1211 09:09:44.721692 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-ht7db_dc79abc2-7b6c-46c4-b9b4-a9c76696a8b2/manager/0.log" Dec 11 09:09:44 crc kubenswrapper[4860]: I1211 09:09:44.722582 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-ht7db_dc79abc2-7b6c-46c4-b9b4-a9c76696a8b2/kube-rbac-proxy/0.log" Dec 11 09:09:44 crc kubenswrapper[4860]: I1211 09:09:44.914949 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-zrhjm_5aff136d-8f48-4cb7-8ddc-72ce966607b7/kube-rbac-proxy/0.log" Dec 11 09:09:44 crc kubenswrapper[4860]: I1211 09:09:44.971840 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-zrhjm_5aff136d-8f48-4cb7-8ddc-72ce966607b7/manager/0.log" Dec 11 09:09:45 crc kubenswrapper[4860]: I1211 09:09:45.116829 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-bvs87_689923af-60de-49e9-bbed-bebaa63adc4b/kube-rbac-proxy/0.log" Dec 11 09:09:45 crc kubenswrapper[4860]: I1211 09:09:45.139315 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-bvs87_689923af-60de-49e9-bbed-bebaa63adc4b/manager/0.log" Dec 11 09:09:45 crc kubenswrapper[4860]: I1211 09:09:45.205623 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-84b575879fgp9c8_04044c38-9aa4-4439-bad4-cf687e00ed5a/kube-rbac-proxy/0.log" Dec 11 09:09:45 crc kubenswrapper[4860]: I1211 09:09:45.303923 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-84b575879fgp9c8_04044c38-9aa4-4439-bad4-cf687e00ed5a/manager/0.log" Dec 11 09:09:45 crc kubenswrapper[4860]: I1211 09:09:45.718158 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-7d57466ffb-mg466_f57d6313-d688-4a8d-a9cb-face8d0e0ef6/operator/0.log" Dec 11 09:09:45 crc kubenswrapper[4860]: I1211 09:09:45.901546 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-hm8hr_aec175c6-0b2d-40d8-855a-6132fb5e66a0/registry-server/0.log" Dec 11 09:09:46 crc kubenswrapper[4860]: I1211 09:09:46.172548 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-nvt56_99f7e97f-7cea-4785-b95e-133adf238ac3/kube-rbac-proxy/0.log" Dec 11 09:09:46 crc kubenswrapper[4860]: I1211 09:09:46.226293 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-nvt56_99f7e97f-7cea-4785-b95e-133adf238ac3/manager/0.log" Dec 11 09:09:46 crc kubenswrapper[4860]: I1211 09:09:46.304495 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-2qmxt_a52f8df6-7c28-4776-aca5-bd8b47e82fe8/kube-rbac-proxy/0.log" Dec 11 09:09:46 crc kubenswrapper[4860]: I1211 09:09:46.403955 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-6d784644c6-68pll_4bbd2122-dfb6-486b-8196-375ffeb93527/manager/0.log" Dec 11 09:09:46 crc kubenswrapper[4860]: I1211 09:09:46.437680 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-2qmxt_a52f8df6-7c28-4776-aca5-bd8b47e82fe8/manager/0.log" Dec 11 09:09:46 crc kubenswrapper[4860]: I1211 09:09:46.496747 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-jsftr_22f3c8f5-1781-4ba2-8610-c9ada5539d92/operator/0.log" Dec 11 09:09:46 crc kubenswrapper[4860]: I1211 09:09:46.605933 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-9d58d64bc-p7dkp_211d9c9c-f584-43ca-8db1-7b81f6307c21/kube-rbac-proxy/0.log" Dec 11 09:09:46 crc kubenswrapper[4860]: I1211 09:09:46.666246 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-9d58d64bc-p7dkp_211d9c9c-f584-43ca-8db1-7b81f6307c21/manager/0.log" Dec 11 09:09:46 crc kubenswrapper[4860]: I1211 09:09:46.698102 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-58d5ff84df-hrpn8_e36550af-98a6-49b3-9769-bd52d0da2838/kube-rbac-proxy/0.log" Dec 11 09:09:46 crc kubenswrapper[4860]: I1211 09:09:46.806837 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-58d5ff84df-hrpn8_e36550af-98a6-49b3-9769-bd52d0da2838/manager/0.log" Dec 11 09:09:46 crc kubenswrapper[4860]: I1211 09:09:46.904520 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-fh5vc_b6c38bbc-80d0-449a-aab0-291078361ebd/manager/0.log" Dec 11 09:09:46 crc kubenswrapper[4860]: I1211 09:09:46.945835 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-fh5vc_b6c38bbc-80d0-449a-aab0-291078361ebd/kube-rbac-proxy/0.log" Dec 11 09:09:47 crc kubenswrapper[4860]: I1211 09:09:47.010447 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-75944c9b7-ck9nc_8462b48d-cc3d-4714-9558-22818db99c08/kube-rbac-proxy/0.log" Dec 11 09:09:47 crc kubenswrapper[4860]: I1211 09:09:47.075822 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-75944c9b7-ck9nc_8462b48d-cc3d-4714-9558-22818db99c08/manager/0.log" Dec 11 09:10:05 crc kubenswrapper[4860]: I1211 09:10:05.512127 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-76lnq_7a764867-5216-47fb-8211-cea78b79f707/control-plane-machine-set-operator/0.log" Dec 11 09:10:05 crc kubenswrapper[4860]: I1211 09:10:05.702114 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-nc65d_cb55e751-9913-4a3c-a0b4-3e681129d052/kube-rbac-proxy/0.log" Dec 11 09:10:05 crc kubenswrapper[4860]: I1211 09:10:05.717395 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-nc65d_cb55e751-9913-4a3c-a0b4-3e681129d052/machine-api-operator/0.log" Dec 11 09:10:17 crc kubenswrapper[4860]: I1211 09:10:17.733622 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-7bszh_5198d79a-f8ab-4a3a-9859-0925545321a4/cert-manager-controller/0.log" Dec 11 09:10:17 crc kubenswrapper[4860]: I1211 09:10:17.923552 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-qhfrs_29ede248-cc17-41b1-a1c4-4feaa6010cf2/cert-manager-cainjector/0.log" Dec 11 09:10:17 crc kubenswrapper[4860]: I1211 09:10:17.964667 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-rvtht_7ea21c5c-320d-4bb0-9bea-186a528a61f4/cert-manager-webhook/0.log" Dec 11 09:10:30 crc kubenswrapper[4860]: I1211 09:10:30.489000 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-6ff7998486-2q4xw_4cc66f42-2ab8-4a37-be52-5eb9be9fdd4b/nmstate-console-plugin/0.log" Dec 11 09:10:30 crc kubenswrapper[4860]: I1211 09:10:30.615059 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-kv4zf_f17c9e82-c0f3-408d-a871-d24992f28860/nmstate-handler/0.log" Dec 11 09:10:30 crc kubenswrapper[4860]: I1211 09:10:30.705448 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f7f7578db-qv9ws_e18c8adc-cdba-4f7f-8b5d-f15a73397c2b/nmstate-metrics/0.log" Dec 11 09:10:30 crc kubenswrapper[4860]: I1211 09:10:30.768388 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f7f7578db-qv9ws_e18c8adc-cdba-4f7f-8b5d-f15a73397c2b/kube-rbac-proxy/0.log" Dec 11 09:10:30 crc kubenswrapper[4860]: I1211 09:10:30.942359 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-6769fb99d-49jqb_4d7eadcc-525c-4503-a8c1-7c0b0c04a9e8/nmstate-operator/0.log" Dec 11 09:10:30 crc kubenswrapper[4860]: I1211 09:10:30.967397 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-f8fb84555-nnjsk_295f8d4b-cf0e-41e9-a9a6-8ff38c589ecc/nmstate-webhook/0.log" Dec 11 09:10:38 crc kubenswrapper[4860]: I1211 09:10:38.795183 4860 patch_prober.go:28] interesting pod/machine-config-daemon-99qgp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 09:10:38 crc kubenswrapper[4860]: I1211 09:10:38.795816 4860 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 09:10:46 crc kubenswrapper[4860]: I1211 09:10:46.051489 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-5bddd4b946-96p7v_31da3aab-3e08-4b87-a866-d49a4224cf29/kube-rbac-proxy/0.log" Dec 11 09:10:46 crc kubenswrapper[4860]: I1211 09:10:46.179887 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-5bddd4b946-96p7v_31da3aab-3e08-4b87-a866-d49a4224cf29/controller/0.log" Dec 11 09:10:46 crc kubenswrapper[4860]: I1211 09:10:46.314497 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c6crs_130f6554-c476-49a0-8270-593008c1ffc6/cp-frr-files/0.log" Dec 11 09:10:46 crc kubenswrapper[4860]: I1211 09:10:46.468144 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c6crs_130f6554-c476-49a0-8270-593008c1ffc6/cp-frr-files/0.log" Dec 11 09:10:46 crc kubenswrapper[4860]: I1211 09:10:46.475569 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c6crs_130f6554-c476-49a0-8270-593008c1ffc6/cp-metrics/0.log" Dec 11 09:10:46 crc kubenswrapper[4860]: I1211 09:10:46.502361 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c6crs_130f6554-c476-49a0-8270-593008c1ffc6/cp-reloader/0.log" Dec 11 09:10:46 crc kubenswrapper[4860]: I1211 09:10:46.513803 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c6crs_130f6554-c476-49a0-8270-593008c1ffc6/cp-reloader/0.log" Dec 11 09:10:46 crc kubenswrapper[4860]: I1211 09:10:46.919737 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c6crs_130f6554-c476-49a0-8270-593008c1ffc6/cp-frr-files/0.log" Dec 11 09:10:46 crc kubenswrapper[4860]: I1211 09:10:46.941251 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c6crs_130f6554-c476-49a0-8270-593008c1ffc6/cp-metrics/0.log" Dec 11 09:10:46 crc kubenswrapper[4860]: I1211 09:10:46.942406 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c6crs_130f6554-c476-49a0-8270-593008c1ffc6/cp-reloader/0.log" Dec 11 09:10:46 crc kubenswrapper[4860]: I1211 09:10:46.976753 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c6crs_130f6554-c476-49a0-8270-593008c1ffc6/cp-metrics/0.log" Dec 11 09:10:47 crc kubenswrapper[4860]: I1211 09:10:47.146015 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c6crs_130f6554-c476-49a0-8270-593008c1ffc6/cp-frr-files/0.log" Dec 11 09:10:47 crc kubenswrapper[4860]: I1211 09:10:47.182707 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c6crs_130f6554-c476-49a0-8270-593008c1ffc6/cp-reloader/0.log" Dec 11 09:10:47 crc kubenswrapper[4860]: I1211 09:10:47.185588 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c6crs_130f6554-c476-49a0-8270-593008c1ffc6/cp-metrics/0.log" Dec 11 09:10:47 crc kubenswrapper[4860]: I1211 09:10:47.240907 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c6crs_130f6554-c476-49a0-8270-593008c1ffc6/controller/0.log" Dec 11 09:10:47 crc kubenswrapper[4860]: I1211 09:10:47.367622 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c6crs_130f6554-c476-49a0-8270-593008c1ffc6/frr-metrics/0.log" Dec 11 09:10:47 crc kubenswrapper[4860]: I1211 09:10:47.396669 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c6crs_130f6554-c476-49a0-8270-593008c1ffc6/kube-rbac-proxy/0.log" Dec 11 09:10:47 crc kubenswrapper[4860]: I1211 09:10:47.438754 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c6crs_130f6554-c476-49a0-8270-593008c1ffc6/kube-rbac-proxy-frr/0.log" Dec 11 09:10:47 crc kubenswrapper[4860]: I1211 09:10:47.601463 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c6crs_130f6554-c476-49a0-8270-593008c1ffc6/reloader/0.log" Dec 11 09:10:47 crc kubenswrapper[4860]: I1211 09:10:47.690344 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7784b6fcf-2548h_98afb3e0-d363-472c-886e-2b2981cb13a4/frr-k8s-webhook-server/0.log" Dec 11 09:10:47 crc kubenswrapper[4860]: I1211 09:10:47.948552 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-d5785c6bc-n9n8q_e0e047cd-d665-4b39-8588-b0d396e3b36d/manager/0.log" Dec 11 09:10:48 crc kubenswrapper[4860]: I1211 09:10:48.079107 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-59475655bd-8sqgf_47efc842-7de4-461c-bafc-0fbaabb6c3ad/webhook-server/0.log" Dec 11 09:10:48 crc kubenswrapper[4860]: I1211 09:10:48.199745 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-qmw55_b5eafccf-57d6-4bcb-a43f-5ef3a2f04db0/kube-rbac-proxy/0.log" Dec 11 09:10:48 crc kubenswrapper[4860]: I1211 09:10:48.752381 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c6crs_130f6554-c476-49a0-8270-593008c1ffc6/frr/0.log" Dec 11 09:10:48 crc kubenswrapper[4860]: I1211 09:10:48.787993 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-qmw55_b5eafccf-57d6-4bcb-a43f-5ef3a2f04db0/speaker/0.log" Dec 11 09:11:00 crc kubenswrapper[4860]: I1211 09:11:00.414417 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4vr6tn_da496881-c195-4cc2-9277-c8118bf651dc/util/0.log" Dec 11 09:11:00 crc kubenswrapper[4860]: I1211 09:11:00.614884 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4vr6tn_da496881-c195-4cc2-9277-c8118bf651dc/pull/0.log" Dec 11 09:11:00 crc kubenswrapper[4860]: I1211 09:11:00.617426 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4vr6tn_da496881-c195-4cc2-9277-c8118bf651dc/pull/0.log" Dec 11 09:11:00 crc kubenswrapper[4860]: I1211 09:11:00.668930 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4vr6tn_da496881-c195-4cc2-9277-c8118bf651dc/util/0.log" Dec 11 09:11:00 crc kubenswrapper[4860]: I1211 09:11:00.823723 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4vr6tn_da496881-c195-4cc2-9277-c8118bf651dc/pull/0.log" Dec 11 09:11:00 crc kubenswrapper[4860]: I1211 09:11:00.831533 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4vr6tn_da496881-c195-4cc2-9277-c8118bf651dc/util/0.log" Dec 11 09:11:00 crc kubenswrapper[4860]: I1211 09:11:00.864224 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4vr6tn_da496881-c195-4cc2-9277-c8118bf651dc/extract/0.log" Dec 11 09:11:01 crc kubenswrapper[4860]: I1211 09:11:01.154888 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8nkkjk_f80508d8-cf59-4656-af41-b0688b7b997c/util/0.log" Dec 11 09:11:01 crc kubenswrapper[4860]: I1211 09:11:01.418358 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8nkkjk_f80508d8-cf59-4656-af41-b0688b7b997c/util/0.log" Dec 11 09:11:01 crc kubenswrapper[4860]: I1211 09:11:01.427488 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8nkkjk_f80508d8-cf59-4656-af41-b0688b7b997c/pull/0.log" Dec 11 09:11:01 crc kubenswrapper[4860]: I1211 09:11:01.430918 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8nkkjk_f80508d8-cf59-4656-af41-b0688b7b997c/pull/0.log" Dec 11 09:11:01 crc kubenswrapper[4860]: I1211 09:11:01.587159 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8nkkjk_f80508d8-cf59-4656-af41-b0688b7b997c/pull/0.log" Dec 11 09:11:01 crc kubenswrapper[4860]: I1211 09:11:01.621274 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8nkkjk_f80508d8-cf59-4656-af41-b0688b7b997c/util/0.log" Dec 11 09:11:01 crc kubenswrapper[4860]: I1211 09:11:01.658825 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8nkkjk_f80508d8-cf59-4656-af41-b0688b7b997c/extract/0.log" Dec 11 09:11:01 crc kubenswrapper[4860]: I1211 09:11:01.816749 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-bg2t9_79181f73-fb9e-40c5-96b8-0ad95545f1f7/extract-utilities/0.log" Dec 11 09:11:02 crc kubenswrapper[4860]: I1211 09:11:02.009797 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-bg2t9_79181f73-fb9e-40c5-96b8-0ad95545f1f7/extract-utilities/0.log" Dec 11 09:11:02 crc kubenswrapper[4860]: I1211 09:11:02.054968 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-bg2t9_79181f73-fb9e-40c5-96b8-0ad95545f1f7/extract-content/0.log" Dec 11 09:11:02 crc kubenswrapper[4860]: I1211 09:11:02.066180 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-bg2t9_79181f73-fb9e-40c5-96b8-0ad95545f1f7/extract-content/0.log" Dec 11 09:11:02 crc kubenswrapper[4860]: I1211 09:11:02.219227 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-bg2t9_79181f73-fb9e-40c5-96b8-0ad95545f1f7/extract-utilities/0.log" Dec 11 09:11:02 crc kubenswrapper[4860]: I1211 09:11:02.379365 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-bg2t9_79181f73-fb9e-40c5-96b8-0ad95545f1f7/extract-content/0.log" Dec 11 09:11:02 crc kubenswrapper[4860]: I1211 09:11:02.494499 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-kzkbs_f1bc94aa-96ad-4e1f-a407-9968d42069aa/extract-utilities/0.log" Dec 11 09:11:02 crc kubenswrapper[4860]: I1211 09:11:02.675539 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-kzkbs_f1bc94aa-96ad-4e1f-a407-9968d42069aa/extract-content/0.log" Dec 11 09:11:02 crc kubenswrapper[4860]: I1211 09:11:02.692351 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-kzkbs_f1bc94aa-96ad-4e1f-a407-9968d42069aa/extract-utilities/0.log" Dec 11 09:11:02 crc kubenswrapper[4860]: I1211 09:11:02.739944 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-bg2t9_79181f73-fb9e-40c5-96b8-0ad95545f1f7/registry-server/0.log" Dec 11 09:11:02 crc kubenswrapper[4860]: I1211 09:11:02.753833 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-kzkbs_f1bc94aa-96ad-4e1f-a407-9968d42069aa/extract-content/0.log" Dec 11 09:11:02 crc kubenswrapper[4860]: I1211 09:11:02.972012 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-kzkbs_f1bc94aa-96ad-4e1f-a407-9968d42069aa/extract-utilities/0.log" Dec 11 09:11:02 crc kubenswrapper[4860]: I1211 09:11:02.984746 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-kzkbs_f1bc94aa-96ad-4e1f-a407-9968d42069aa/extract-content/0.log" Dec 11 09:11:03 crc kubenswrapper[4860]: I1211 09:11:03.226296 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-trp75_dd2052c5-96c6-4226-858e-61e3ddbfeef9/marketplace-operator/0.log" Dec 11 09:11:03 crc kubenswrapper[4860]: I1211 09:11:03.365177 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-nkvw8_74e80c39-0b0c-42f6-88b6-a800ec910606/extract-utilities/0.log" Dec 11 09:11:03 crc kubenswrapper[4860]: I1211 09:11:03.476390 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-kzkbs_f1bc94aa-96ad-4e1f-a407-9968d42069aa/registry-server/0.log" Dec 11 09:11:03 crc kubenswrapper[4860]: I1211 09:11:03.510496 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-nkvw8_74e80c39-0b0c-42f6-88b6-a800ec910606/extract-content/0.log" Dec 11 09:11:03 crc kubenswrapper[4860]: I1211 09:11:03.549249 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-nkvw8_74e80c39-0b0c-42f6-88b6-a800ec910606/extract-utilities/0.log" Dec 11 09:11:03 crc kubenswrapper[4860]: I1211 09:11:03.582058 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-nkvw8_74e80c39-0b0c-42f6-88b6-a800ec910606/extract-content/0.log" Dec 11 09:11:03 crc kubenswrapper[4860]: I1211 09:11:03.801555 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-nkvw8_74e80c39-0b0c-42f6-88b6-a800ec910606/extract-content/0.log" Dec 11 09:11:03 crc kubenswrapper[4860]: I1211 09:11:03.809714 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-nkvw8_74e80c39-0b0c-42f6-88b6-a800ec910606/extract-utilities/0.log" Dec 11 09:11:03 crc kubenswrapper[4860]: I1211 09:11:03.980300 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-7vj8m_7d8a48c0-183c-4699-86d1-f6d0ef767cdf/extract-utilities/0.log" Dec 11 09:11:04 crc kubenswrapper[4860]: I1211 09:11:04.033403 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-nkvw8_74e80c39-0b0c-42f6-88b6-a800ec910606/registry-server/0.log" Dec 11 09:11:04 crc kubenswrapper[4860]: I1211 09:11:04.159473 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-7vj8m_7d8a48c0-183c-4699-86d1-f6d0ef767cdf/extract-utilities/0.log" Dec 11 09:11:04 crc kubenswrapper[4860]: I1211 09:11:04.176918 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-7vj8m_7d8a48c0-183c-4699-86d1-f6d0ef767cdf/extract-content/0.log" Dec 11 09:11:04 crc kubenswrapper[4860]: I1211 09:11:04.240282 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-7vj8m_7d8a48c0-183c-4699-86d1-f6d0ef767cdf/extract-content/0.log" Dec 11 09:11:04 crc kubenswrapper[4860]: I1211 09:11:04.391428 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-7vj8m_7d8a48c0-183c-4699-86d1-f6d0ef767cdf/extract-utilities/0.log" Dec 11 09:11:04 crc kubenswrapper[4860]: I1211 09:11:04.425510 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-7vj8m_7d8a48c0-183c-4699-86d1-f6d0ef767cdf/extract-content/0.log" Dec 11 09:11:04 crc kubenswrapper[4860]: I1211 09:11:04.997462 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-7vj8m_7d8a48c0-183c-4699-86d1-f6d0ef767cdf/registry-server/0.log" Dec 11 09:11:08 crc kubenswrapper[4860]: I1211 09:11:08.795893 4860 patch_prober.go:28] interesting pod/machine-config-daemon-99qgp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 09:11:08 crc kubenswrapper[4860]: I1211 09:11:08.796260 4860 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 09:11:26 crc kubenswrapper[4860]: E1211 09:11:26.353786 4860 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.169:59256->38.102.83.169:39441: write tcp 38.102.83.169:59256->38.102.83.169:39441: write: broken pipe Dec 11 09:11:38 crc kubenswrapper[4860]: I1211 09:11:38.795115 4860 patch_prober.go:28] interesting pod/machine-config-daemon-99qgp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 09:11:38 crc kubenswrapper[4860]: I1211 09:11:38.795830 4860 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 09:11:38 crc kubenswrapper[4860]: I1211 09:11:38.795894 4860 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" Dec 11 09:11:38 crc kubenswrapper[4860]: I1211 09:11:38.797034 4860 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8eb6bd97d2b0921d3b9d3592fa7a8bac43db52ba5376a780109acc0b8111e2bf"} pod="openshift-machine-config-operator/machine-config-daemon-99qgp" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 11 09:11:38 crc kubenswrapper[4860]: I1211 09:11:38.797114 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" containerName="machine-config-daemon" containerID="cri-o://8eb6bd97d2b0921d3b9d3592fa7a8bac43db52ba5376a780109acc0b8111e2bf" gracePeriod=600 Dec 11 09:11:39 crc kubenswrapper[4860]: I1211 09:11:39.423549 4860 generic.go:334] "Generic (PLEG): container finished" podID="31c30642-6e60-41b4-a477-0d802424e0aa" containerID="8eb6bd97d2b0921d3b9d3592fa7a8bac43db52ba5376a780109acc0b8111e2bf" exitCode=0 Dec 11 09:11:39 crc kubenswrapper[4860]: I1211 09:11:39.423805 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" event={"ID":"31c30642-6e60-41b4-a477-0d802424e0aa","Type":"ContainerDied","Data":"8eb6bd97d2b0921d3b9d3592fa7a8bac43db52ba5376a780109acc0b8111e2bf"} Dec 11 09:11:39 crc kubenswrapper[4860]: I1211 09:11:39.423879 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" event={"ID":"31c30642-6e60-41b4-a477-0d802424e0aa","Type":"ContainerStarted","Data":"6704d8c98ab428aac5737fa114668f6cf3119541ee0fe55ea326d429fcbe2d82"} Dec 11 09:11:39 crc kubenswrapper[4860]: I1211 09:11:39.423904 4860 scope.go:117] "RemoveContainer" containerID="76665d219a809fd146709b7837eb798329ce7fb1c53d5b15c247e7e654985e07" Dec 11 09:12:34 crc kubenswrapper[4860]: I1211 09:12:34.381755 4860 scope.go:117] "RemoveContainer" containerID="619dd1ba11a09a1196a4e956e2fe414ed7746b8f33710bf71c0d0768c234b4ce" Dec 11 09:12:34 crc kubenswrapper[4860]: I1211 09:12:34.437527 4860 scope.go:117] "RemoveContainer" containerID="a2484ecb14b06167fc67c85af14815f6d29a85589a426f77ae1d14bc6b809412" Dec 11 09:12:34 crc kubenswrapper[4860]: I1211 09:12:34.480714 4860 scope.go:117] "RemoveContainer" containerID="81768794f2f872c0777df7a56b46523db04d2fb2246d0258f185edb17452da85" Dec 11 09:12:46 crc kubenswrapper[4860]: I1211 09:12:46.089287 4860 generic.go:334] "Generic (PLEG): container finished" podID="11670ddb-0c7a-4ac0-b039-a5254f80b5f8" containerID="4cefe6af6c5989c2d95f88a602fdcd910e7ecf41908dc35b81db7ef0bb0d72e7" exitCode=0 Dec 11 09:12:46 crc kubenswrapper[4860]: I1211 09:12:46.089374 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-n96c5/must-gather-6gnzt" event={"ID":"11670ddb-0c7a-4ac0-b039-a5254f80b5f8","Type":"ContainerDied","Data":"4cefe6af6c5989c2d95f88a602fdcd910e7ecf41908dc35b81db7ef0bb0d72e7"} Dec 11 09:12:46 crc kubenswrapper[4860]: I1211 09:12:46.090380 4860 scope.go:117] "RemoveContainer" containerID="4cefe6af6c5989c2d95f88a602fdcd910e7ecf41908dc35b81db7ef0bb0d72e7" Dec 11 09:12:46 crc kubenswrapper[4860]: I1211 09:12:46.899241 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-n96c5_must-gather-6gnzt_11670ddb-0c7a-4ac0-b039-a5254f80b5f8/gather/0.log" Dec 11 09:12:54 crc kubenswrapper[4860]: I1211 09:12:54.940350 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-n96c5/must-gather-6gnzt"] Dec 11 09:12:54 crc kubenswrapper[4860]: I1211 09:12:54.941311 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-n96c5/must-gather-6gnzt" podUID="11670ddb-0c7a-4ac0-b039-a5254f80b5f8" containerName="copy" containerID="cri-o://c5b8630a7391dc3185f8f8b32bc1a078cd3af6b7c3b10143b59bca15218a1732" gracePeriod=2 Dec 11 09:12:54 crc kubenswrapper[4860]: I1211 09:12:54.952164 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-n96c5/must-gather-6gnzt"] Dec 11 09:12:55 crc kubenswrapper[4860]: I1211 09:12:55.178304 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-n96c5_must-gather-6gnzt_11670ddb-0c7a-4ac0-b039-a5254f80b5f8/copy/0.log" Dec 11 09:12:55 crc kubenswrapper[4860]: I1211 09:12:55.179235 4860 generic.go:334] "Generic (PLEG): container finished" podID="11670ddb-0c7a-4ac0-b039-a5254f80b5f8" containerID="c5b8630a7391dc3185f8f8b32bc1a078cd3af6b7c3b10143b59bca15218a1732" exitCode=143 Dec 11 09:12:55 crc kubenswrapper[4860]: I1211 09:12:55.389439 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-n96c5_must-gather-6gnzt_11670ddb-0c7a-4ac0-b039-a5254f80b5f8/copy/0.log" Dec 11 09:12:55 crc kubenswrapper[4860]: I1211 09:12:55.389830 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-n96c5/must-gather-6gnzt" Dec 11 09:12:55 crc kubenswrapper[4860]: I1211 09:12:55.471215 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/11670ddb-0c7a-4ac0-b039-a5254f80b5f8-must-gather-output\") pod \"11670ddb-0c7a-4ac0-b039-a5254f80b5f8\" (UID: \"11670ddb-0c7a-4ac0-b039-a5254f80b5f8\") " Dec 11 09:12:55 crc kubenswrapper[4860]: I1211 09:12:55.471331 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j9jwp\" (UniqueName: \"kubernetes.io/projected/11670ddb-0c7a-4ac0-b039-a5254f80b5f8-kube-api-access-j9jwp\") pod \"11670ddb-0c7a-4ac0-b039-a5254f80b5f8\" (UID: \"11670ddb-0c7a-4ac0-b039-a5254f80b5f8\") " Dec 11 09:12:55 crc kubenswrapper[4860]: I1211 09:12:55.483102 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/11670ddb-0c7a-4ac0-b039-a5254f80b5f8-kube-api-access-j9jwp" (OuterVolumeSpecName: "kube-api-access-j9jwp") pod "11670ddb-0c7a-4ac0-b039-a5254f80b5f8" (UID: "11670ddb-0c7a-4ac0-b039-a5254f80b5f8"). InnerVolumeSpecName "kube-api-access-j9jwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 09:12:55 crc kubenswrapper[4860]: I1211 09:12:55.574031 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j9jwp\" (UniqueName: \"kubernetes.io/projected/11670ddb-0c7a-4ac0-b039-a5254f80b5f8-kube-api-access-j9jwp\") on node \"crc\" DevicePath \"\"" Dec 11 09:12:55 crc kubenswrapper[4860]: I1211 09:12:55.625214 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/11670ddb-0c7a-4ac0-b039-a5254f80b5f8-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "11670ddb-0c7a-4ac0-b039-a5254f80b5f8" (UID: "11670ddb-0c7a-4ac0-b039-a5254f80b5f8"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 09:12:55 crc kubenswrapper[4860]: I1211 09:12:55.676355 4860 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/11670ddb-0c7a-4ac0-b039-a5254f80b5f8-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 11 09:12:56 crc kubenswrapper[4860]: I1211 09:12:56.190329 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-n96c5_must-gather-6gnzt_11670ddb-0c7a-4ac0-b039-a5254f80b5f8/copy/0.log" Dec 11 09:12:56 crc kubenswrapper[4860]: I1211 09:12:56.190979 4860 scope.go:117] "RemoveContainer" containerID="c5b8630a7391dc3185f8f8b32bc1a078cd3af6b7c3b10143b59bca15218a1732" Dec 11 09:12:56 crc kubenswrapper[4860]: I1211 09:12:56.191018 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-n96c5/must-gather-6gnzt" Dec 11 09:12:56 crc kubenswrapper[4860]: I1211 09:12:56.213366 4860 scope.go:117] "RemoveContainer" containerID="4cefe6af6c5989c2d95f88a602fdcd910e7ecf41908dc35b81db7ef0bb0d72e7" Dec 11 09:12:57 crc kubenswrapper[4860]: I1211 09:12:57.597007 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="11670ddb-0c7a-4ac0-b039-a5254f80b5f8" path="/var/lib/kubelet/pods/11670ddb-0c7a-4ac0-b039-a5254f80b5f8/volumes" Dec 11 09:14:08 crc kubenswrapper[4860]: I1211 09:14:08.795210 4860 patch_prober.go:28] interesting pod/machine-config-daemon-99qgp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 09:14:08 crc kubenswrapper[4860]: I1211 09:14:08.796850 4860 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 09:14:38 crc kubenswrapper[4860]: I1211 09:14:38.795799 4860 patch_prober.go:28] interesting pod/machine-config-daemon-99qgp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 09:14:38 crc kubenswrapper[4860]: I1211 09:14:38.796340 4860 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 09:15:00 crc kubenswrapper[4860]: I1211 09:15:00.174588 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29424075-w8lqq"] Dec 11 09:15:00 crc kubenswrapper[4860]: E1211 09:15:00.175909 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d3489b0-ae39-422e-a7f0-7969aa384f6f" containerName="extract-content" Dec 11 09:15:00 crc kubenswrapper[4860]: I1211 09:15:00.175938 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d3489b0-ae39-422e-a7f0-7969aa384f6f" containerName="extract-content" Dec 11 09:15:00 crc kubenswrapper[4860]: E1211 09:15:00.175969 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d3489b0-ae39-422e-a7f0-7969aa384f6f" containerName="registry-server" Dec 11 09:15:00 crc kubenswrapper[4860]: I1211 09:15:00.175990 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d3489b0-ae39-422e-a7f0-7969aa384f6f" containerName="registry-server" Dec 11 09:15:00 crc kubenswrapper[4860]: E1211 09:15:00.176023 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11670ddb-0c7a-4ac0-b039-a5254f80b5f8" containerName="copy" Dec 11 09:15:00 crc kubenswrapper[4860]: I1211 09:15:00.176032 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="11670ddb-0c7a-4ac0-b039-a5254f80b5f8" containerName="copy" Dec 11 09:15:00 crc kubenswrapper[4860]: E1211 09:15:00.176045 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11670ddb-0c7a-4ac0-b039-a5254f80b5f8" containerName="gather" Dec 11 09:15:00 crc kubenswrapper[4860]: I1211 09:15:00.176053 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="11670ddb-0c7a-4ac0-b039-a5254f80b5f8" containerName="gather" Dec 11 09:15:00 crc kubenswrapper[4860]: E1211 09:15:00.176073 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dbe01adc-96a1-49b3-8541-7032fbc5d8ae" containerName="container-00" Dec 11 09:15:00 crc kubenswrapper[4860]: I1211 09:15:00.176080 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="dbe01adc-96a1-49b3-8541-7032fbc5d8ae" containerName="container-00" Dec 11 09:15:00 crc kubenswrapper[4860]: E1211 09:15:00.176096 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d3489b0-ae39-422e-a7f0-7969aa384f6f" containerName="extract-utilities" Dec 11 09:15:00 crc kubenswrapper[4860]: I1211 09:15:00.176106 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d3489b0-ae39-422e-a7f0-7969aa384f6f" containerName="extract-utilities" Dec 11 09:15:00 crc kubenswrapper[4860]: I1211 09:15:00.176383 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="11670ddb-0c7a-4ac0-b039-a5254f80b5f8" containerName="gather" Dec 11 09:15:00 crc kubenswrapper[4860]: I1211 09:15:00.176406 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d3489b0-ae39-422e-a7f0-7969aa384f6f" containerName="registry-server" Dec 11 09:15:00 crc kubenswrapper[4860]: I1211 09:15:00.176426 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="11670ddb-0c7a-4ac0-b039-a5254f80b5f8" containerName="copy" Dec 11 09:15:00 crc kubenswrapper[4860]: I1211 09:15:00.176440 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="dbe01adc-96a1-49b3-8541-7032fbc5d8ae" containerName="container-00" Dec 11 09:15:00 crc kubenswrapper[4860]: I1211 09:15:00.177367 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29424075-w8lqq" Dec 11 09:15:00 crc kubenswrapper[4860]: I1211 09:15:00.180999 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 11 09:15:00 crc kubenswrapper[4860]: I1211 09:15:00.183677 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 11 09:15:00 crc kubenswrapper[4860]: I1211 09:15:00.190754 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29424075-w8lqq"] Dec 11 09:15:00 crc kubenswrapper[4860]: I1211 09:15:00.276050 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c24e63af-3b46-4702-978c-ed0e7256ad61-secret-volume\") pod \"collect-profiles-29424075-w8lqq\" (UID: \"c24e63af-3b46-4702-978c-ed0e7256ad61\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424075-w8lqq" Dec 11 09:15:00 crc kubenswrapper[4860]: I1211 09:15:00.276206 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cnp5n\" (UniqueName: \"kubernetes.io/projected/c24e63af-3b46-4702-978c-ed0e7256ad61-kube-api-access-cnp5n\") pod \"collect-profiles-29424075-w8lqq\" (UID: \"c24e63af-3b46-4702-978c-ed0e7256ad61\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424075-w8lqq" Dec 11 09:15:00 crc kubenswrapper[4860]: I1211 09:15:00.276285 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c24e63af-3b46-4702-978c-ed0e7256ad61-config-volume\") pod \"collect-profiles-29424075-w8lqq\" (UID: \"c24e63af-3b46-4702-978c-ed0e7256ad61\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424075-w8lqq" Dec 11 09:15:00 crc kubenswrapper[4860]: I1211 09:15:00.378119 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c24e63af-3b46-4702-978c-ed0e7256ad61-secret-volume\") pod \"collect-profiles-29424075-w8lqq\" (UID: \"c24e63af-3b46-4702-978c-ed0e7256ad61\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424075-w8lqq" Dec 11 09:15:00 crc kubenswrapper[4860]: I1211 09:15:00.378239 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cnp5n\" (UniqueName: \"kubernetes.io/projected/c24e63af-3b46-4702-978c-ed0e7256ad61-kube-api-access-cnp5n\") pod \"collect-profiles-29424075-w8lqq\" (UID: \"c24e63af-3b46-4702-978c-ed0e7256ad61\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424075-w8lqq" Dec 11 09:15:00 crc kubenswrapper[4860]: I1211 09:15:00.378309 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c24e63af-3b46-4702-978c-ed0e7256ad61-config-volume\") pod \"collect-profiles-29424075-w8lqq\" (UID: \"c24e63af-3b46-4702-978c-ed0e7256ad61\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424075-w8lqq" Dec 11 09:15:00 crc kubenswrapper[4860]: I1211 09:15:00.379434 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c24e63af-3b46-4702-978c-ed0e7256ad61-config-volume\") pod \"collect-profiles-29424075-w8lqq\" (UID: \"c24e63af-3b46-4702-978c-ed0e7256ad61\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424075-w8lqq" Dec 11 09:15:00 crc kubenswrapper[4860]: I1211 09:15:00.390900 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c24e63af-3b46-4702-978c-ed0e7256ad61-secret-volume\") pod \"collect-profiles-29424075-w8lqq\" (UID: \"c24e63af-3b46-4702-978c-ed0e7256ad61\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424075-w8lqq" Dec 11 09:15:00 crc kubenswrapper[4860]: I1211 09:15:00.398307 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cnp5n\" (UniqueName: \"kubernetes.io/projected/c24e63af-3b46-4702-978c-ed0e7256ad61-kube-api-access-cnp5n\") pod \"collect-profiles-29424075-w8lqq\" (UID: \"c24e63af-3b46-4702-978c-ed0e7256ad61\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424075-w8lqq" Dec 11 09:15:00 crc kubenswrapper[4860]: I1211 09:15:00.513200 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29424075-w8lqq" Dec 11 09:15:00 crc kubenswrapper[4860]: I1211 09:15:00.980680 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29424075-w8lqq"] Dec 11 09:15:01 crc kubenswrapper[4860]: I1211 09:15:01.326248 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29424075-w8lqq" event={"ID":"c24e63af-3b46-4702-978c-ed0e7256ad61","Type":"ContainerStarted","Data":"c6e0c6c90976537256d01ba795b2b791001403604532182185a46a0948b99462"} Dec 11 09:15:01 crc kubenswrapper[4860]: I1211 09:15:01.326845 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29424075-w8lqq" event={"ID":"c24e63af-3b46-4702-978c-ed0e7256ad61","Type":"ContainerStarted","Data":"ea7a30c3ae610013b8a59576f0271fdd1a4543e8606aa6e9211704600bb1e55f"} Dec 11 09:15:01 crc kubenswrapper[4860]: I1211 09:15:01.350577 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29424075-w8lqq" podStartSLOduration=1.350557569 podStartE2EDuration="1.350557569s" podCreationTimestamp="2025-12-11 09:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 09:15:01.345359546 +0000 UTC m=+3834.073878611" watchObservedRunningTime="2025-12-11 09:15:01.350557569 +0000 UTC m=+3834.079076624" Dec 11 09:15:02 crc kubenswrapper[4860]: I1211 09:15:02.337205 4860 generic.go:334] "Generic (PLEG): container finished" podID="c24e63af-3b46-4702-978c-ed0e7256ad61" containerID="c6e0c6c90976537256d01ba795b2b791001403604532182185a46a0948b99462" exitCode=0 Dec 11 09:15:02 crc kubenswrapper[4860]: I1211 09:15:02.337255 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29424075-w8lqq" event={"ID":"c24e63af-3b46-4702-978c-ed0e7256ad61","Type":"ContainerDied","Data":"c6e0c6c90976537256d01ba795b2b791001403604532182185a46a0948b99462"} Dec 11 09:15:03 crc kubenswrapper[4860]: I1211 09:15:03.665082 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29424075-w8lqq" Dec 11 09:15:03 crc kubenswrapper[4860]: I1211 09:15:03.745790 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c24e63af-3b46-4702-978c-ed0e7256ad61-config-volume\") pod \"c24e63af-3b46-4702-978c-ed0e7256ad61\" (UID: \"c24e63af-3b46-4702-978c-ed0e7256ad61\") " Dec 11 09:15:03 crc kubenswrapper[4860]: I1211 09:15:03.746011 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cnp5n\" (UniqueName: \"kubernetes.io/projected/c24e63af-3b46-4702-978c-ed0e7256ad61-kube-api-access-cnp5n\") pod \"c24e63af-3b46-4702-978c-ed0e7256ad61\" (UID: \"c24e63af-3b46-4702-978c-ed0e7256ad61\") " Dec 11 09:15:03 crc kubenswrapper[4860]: I1211 09:15:03.746073 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c24e63af-3b46-4702-978c-ed0e7256ad61-secret-volume\") pod \"c24e63af-3b46-4702-978c-ed0e7256ad61\" (UID: \"c24e63af-3b46-4702-978c-ed0e7256ad61\") " Dec 11 09:15:03 crc kubenswrapper[4860]: I1211 09:15:03.746720 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c24e63af-3b46-4702-978c-ed0e7256ad61-config-volume" (OuterVolumeSpecName: "config-volume") pod "c24e63af-3b46-4702-978c-ed0e7256ad61" (UID: "c24e63af-3b46-4702-978c-ed0e7256ad61"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 09:15:03 crc kubenswrapper[4860]: I1211 09:15:03.752196 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c24e63af-3b46-4702-978c-ed0e7256ad61-kube-api-access-cnp5n" (OuterVolumeSpecName: "kube-api-access-cnp5n") pod "c24e63af-3b46-4702-978c-ed0e7256ad61" (UID: "c24e63af-3b46-4702-978c-ed0e7256ad61"). InnerVolumeSpecName "kube-api-access-cnp5n". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 09:15:03 crc kubenswrapper[4860]: I1211 09:15:03.752603 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c24e63af-3b46-4702-978c-ed0e7256ad61-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "c24e63af-3b46-4702-978c-ed0e7256ad61" (UID: "c24e63af-3b46-4702-978c-ed0e7256ad61"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 09:15:03 crc kubenswrapper[4860]: I1211 09:15:03.848774 4860 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c24e63af-3b46-4702-978c-ed0e7256ad61-config-volume\") on node \"crc\" DevicePath \"\"" Dec 11 09:15:03 crc kubenswrapper[4860]: I1211 09:15:03.848823 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cnp5n\" (UniqueName: \"kubernetes.io/projected/c24e63af-3b46-4702-978c-ed0e7256ad61-kube-api-access-cnp5n\") on node \"crc\" DevicePath \"\"" Dec 11 09:15:03 crc kubenswrapper[4860]: I1211 09:15:03.848841 4860 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c24e63af-3b46-4702-978c-ed0e7256ad61-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 11 09:15:04 crc kubenswrapper[4860]: I1211 09:15:04.356977 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29424075-w8lqq" event={"ID":"c24e63af-3b46-4702-978c-ed0e7256ad61","Type":"ContainerDied","Data":"ea7a30c3ae610013b8a59576f0271fdd1a4543e8606aa6e9211704600bb1e55f"} Dec 11 09:15:04 crc kubenswrapper[4860]: I1211 09:15:04.357018 4860 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ea7a30c3ae610013b8a59576f0271fdd1a4543e8606aa6e9211704600bb1e55f" Dec 11 09:15:04 crc kubenswrapper[4860]: I1211 09:15:04.357050 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29424075-w8lqq" Dec 11 09:15:04 crc kubenswrapper[4860]: I1211 09:15:04.421993 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29424030-m6tp2"] Dec 11 09:15:04 crc kubenswrapper[4860]: I1211 09:15:04.430769 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29424030-m6tp2"] Dec 11 09:15:05 crc kubenswrapper[4860]: I1211 09:15:05.593627 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5330e7d4-9a3d-41cc-9f94-e17ad44f2080" path="/var/lib/kubelet/pods/5330e7d4-9a3d-41cc-9f94-e17ad44f2080/volumes" Dec 11 09:15:08 crc kubenswrapper[4860]: I1211 09:15:08.795132 4860 patch_prober.go:28] interesting pod/machine-config-daemon-99qgp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 09:15:08 crc kubenswrapper[4860]: I1211 09:15:08.795751 4860 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 09:15:08 crc kubenswrapper[4860]: I1211 09:15:08.795803 4860 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" Dec 11 09:15:08 crc kubenswrapper[4860]: I1211 09:15:08.796664 4860 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6704d8c98ab428aac5737fa114668f6cf3119541ee0fe55ea326d429fcbe2d82"} pod="openshift-machine-config-operator/machine-config-daemon-99qgp" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 11 09:15:08 crc kubenswrapper[4860]: I1211 09:15:08.796744 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" containerName="machine-config-daemon" containerID="cri-o://6704d8c98ab428aac5737fa114668f6cf3119541ee0fe55ea326d429fcbe2d82" gracePeriod=600 Dec 11 09:15:08 crc kubenswrapper[4860]: E1211 09:15:08.921575 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-99qgp_openshift-machine-config-operator(31c30642-6e60-41b4-a477-0d802424e0aa)\"" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" Dec 11 09:15:09 crc kubenswrapper[4860]: I1211 09:15:09.402044 4860 generic.go:334] "Generic (PLEG): container finished" podID="31c30642-6e60-41b4-a477-0d802424e0aa" containerID="6704d8c98ab428aac5737fa114668f6cf3119541ee0fe55ea326d429fcbe2d82" exitCode=0 Dec 11 09:15:09 crc kubenswrapper[4860]: I1211 09:15:09.402103 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" event={"ID":"31c30642-6e60-41b4-a477-0d802424e0aa","Type":"ContainerDied","Data":"6704d8c98ab428aac5737fa114668f6cf3119541ee0fe55ea326d429fcbe2d82"} Dec 11 09:15:09 crc kubenswrapper[4860]: I1211 09:15:09.402717 4860 scope.go:117] "RemoveContainer" containerID="8eb6bd97d2b0921d3b9d3592fa7a8bac43db52ba5376a780109acc0b8111e2bf" Dec 11 09:15:09 crc kubenswrapper[4860]: I1211 09:15:09.403552 4860 scope.go:117] "RemoveContainer" containerID="6704d8c98ab428aac5737fa114668f6cf3119541ee0fe55ea326d429fcbe2d82" Dec 11 09:15:09 crc kubenswrapper[4860]: E1211 09:15:09.404466 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-99qgp_openshift-machine-config-operator(31c30642-6e60-41b4-a477-0d802424e0aa)\"" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" Dec 11 09:15:22 crc kubenswrapper[4860]: I1211 09:15:22.578521 4860 scope.go:117] "RemoveContainer" containerID="6704d8c98ab428aac5737fa114668f6cf3119541ee0fe55ea326d429fcbe2d82" Dec 11 09:15:22 crc kubenswrapper[4860]: E1211 09:15:22.579257 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-99qgp_openshift-machine-config-operator(31c30642-6e60-41b4-a477-0d802424e0aa)\"" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" Dec 11 09:15:34 crc kubenswrapper[4860]: I1211 09:15:34.592702 4860 scope.go:117] "RemoveContainer" containerID="25be5020a4cf38879be0a9e4ba9401a639ce81df3089ca082995ed05ab6b1bd7" Dec 11 09:15:37 crc kubenswrapper[4860]: I1211 09:15:37.586607 4860 scope.go:117] "RemoveContainer" containerID="6704d8c98ab428aac5737fa114668f6cf3119541ee0fe55ea326d429fcbe2d82" Dec 11 09:15:37 crc kubenswrapper[4860]: E1211 09:15:37.587419 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-99qgp_openshift-machine-config-operator(31c30642-6e60-41b4-a477-0d802424e0aa)\"" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" Dec 11 09:15:51 crc kubenswrapper[4860]: I1211 09:15:51.580534 4860 scope.go:117] "RemoveContainer" containerID="6704d8c98ab428aac5737fa114668f6cf3119541ee0fe55ea326d429fcbe2d82" Dec 11 09:15:51 crc kubenswrapper[4860]: E1211 09:15:51.583727 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-99qgp_openshift-machine-config-operator(31c30642-6e60-41b4-a477-0d802424e0aa)\"" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" Dec 11 09:15:54 crc kubenswrapper[4860]: I1211 09:15:54.469286 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-fjd59/must-gather-phzqz"] Dec 11 09:15:54 crc kubenswrapper[4860]: E1211 09:15:54.470450 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c24e63af-3b46-4702-978c-ed0e7256ad61" containerName="collect-profiles" Dec 11 09:15:54 crc kubenswrapper[4860]: I1211 09:15:54.470469 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="c24e63af-3b46-4702-978c-ed0e7256ad61" containerName="collect-profiles" Dec 11 09:15:54 crc kubenswrapper[4860]: I1211 09:15:54.470706 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="c24e63af-3b46-4702-978c-ed0e7256ad61" containerName="collect-profiles" Dec 11 09:15:54 crc kubenswrapper[4860]: I1211 09:15:54.472007 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-fjd59/must-gather-phzqz" Dec 11 09:15:54 crc kubenswrapper[4860]: I1211 09:15:54.474673 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-fjd59"/"kube-root-ca.crt" Dec 11 09:15:54 crc kubenswrapper[4860]: I1211 09:15:54.475238 4860 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-fjd59"/"default-dockercfg-rh9sk" Dec 11 09:15:54 crc kubenswrapper[4860]: I1211 09:15:54.477722 4860 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-fjd59"/"openshift-service-ca.crt" Dec 11 09:15:54 crc kubenswrapper[4860]: I1211 09:15:54.496919 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-fjd59/must-gather-phzqz"] Dec 11 09:15:54 crc kubenswrapper[4860]: I1211 09:15:54.581781 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qqwtc\" (UniqueName: \"kubernetes.io/projected/38bf9964-7add-4228-9abc-c1a5bd18d568-kube-api-access-qqwtc\") pod \"must-gather-phzqz\" (UID: \"38bf9964-7add-4228-9abc-c1a5bd18d568\") " pod="openshift-must-gather-fjd59/must-gather-phzqz" Dec 11 09:15:54 crc kubenswrapper[4860]: I1211 09:15:54.581875 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/38bf9964-7add-4228-9abc-c1a5bd18d568-must-gather-output\") pod \"must-gather-phzqz\" (UID: \"38bf9964-7add-4228-9abc-c1a5bd18d568\") " pod="openshift-must-gather-fjd59/must-gather-phzqz" Dec 11 09:15:54 crc kubenswrapper[4860]: I1211 09:15:54.708986 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qqwtc\" (UniqueName: \"kubernetes.io/projected/38bf9964-7add-4228-9abc-c1a5bd18d568-kube-api-access-qqwtc\") pod \"must-gather-phzqz\" (UID: \"38bf9964-7add-4228-9abc-c1a5bd18d568\") " pod="openshift-must-gather-fjd59/must-gather-phzqz" Dec 11 09:15:54 crc kubenswrapper[4860]: I1211 09:15:54.709142 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/38bf9964-7add-4228-9abc-c1a5bd18d568-must-gather-output\") pod \"must-gather-phzqz\" (UID: \"38bf9964-7add-4228-9abc-c1a5bd18d568\") " pod="openshift-must-gather-fjd59/must-gather-phzqz" Dec 11 09:15:54 crc kubenswrapper[4860]: I1211 09:15:54.711284 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/38bf9964-7add-4228-9abc-c1a5bd18d568-must-gather-output\") pod \"must-gather-phzqz\" (UID: \"38bf9964-7add-4228-9abc-c1a5bd18d568\") " pod="openshift-must-gather-fjd59/must-gather-phzqz" Dec 11 09:15:54 crc kubenswrapper[4860]: I1211 09:15:54.731057 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qqwtc\" (UniqueName: \"kubernetes.io/projected/38bf9964-7add-4228-9abc-c1a5bd18d568-kube-api-access-qqwtc\") pod \"must-gather-phzqz\" (UID: \"38bf9964-7add-4228-9abc-c1a5bd18d568\") " pod="openshift-must-gather-fjd59/must-gather-phzqz" Dec 11 09:15:54 crc kubenswrapper[4860]: I1211 09:15:54.812115 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-fjd59/must-gather-phzqz" Dec 11 09:15:55 crc kubenswrapper[4860]: I1211 09:15:55.282736 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-fjd59/must-gather-phzqz"] Dec 11 09:15:56 crc kubenswrapper[4860]: I1211 09:15:56.097755 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-fjd59/must-gather-phzqz" event={"ID":"38bf9964-7add-4228-9abc-c1a5bd18d568","Type":"ContainerStarted","Data":"9aad5d3923ab3c049951961f15f9728b5ebf276e736f9ca6270c47c2dcd36786"} Dec 11 09:15:56 crc kubenswrapper[4860]: I1211 09:15:56.098067 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-fjd59/must-gather-phzqz" event={"ID":"38bf9964-7add-4228-9abc-c1a5bd18d568","Type":"ContainerStarted","Data":"d2806d73fffa4fc0a3571875988c8f69f884ca94d9b9fca50b635eadd9b5a188"} Dec 11 09:15:56 crc kubenswrapper[4860]: I1211 09:15:56.098079 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-fjd59/must-gather-phzqz" event={"ID":"38bf9964-7add-4228-9abc-c1a5bd18d568","Type":"ContainerStarted","Data":"ce94c624a5c8753264ca401a75989ffef1bf7f25b7f21fd813e2ad767382dd08"} Dec 11 09:15:56 crc kubenswrapper[4860]: I1211 09:15:56.133463 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-fjd59/must-gather-phzqz" podStartSLOduration=2.133434233 podStartE2EDuration="2.133434233s" podCreationTimestamp="2025-12-11 09:15:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 09:15:56.119149083 +0000 UTC m=+3888.847668138" watchObservedRunningTime="2025-12-11 09:15:56.133434233 +0000 UTC m=+3888.861953288" Dec 11 09:15:59 crc kubenswrapper[4860]: I1211 09:15:59.436431 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-fjd59/crc-debug-mqm7t"] Dec 11 09:15:59 crc kubenswrapper[4860]: I1211 09:15:59.438252 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-fjd59/crc-debug-mqm7t" Dec 11 09:15:59 crc kubenswrapper[4860]: I1211 09:15:59.509982 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zsqlh\" (UniqueName: \"kubernetes.io/projected/c208d478-bfb6-4483-b01b-2ef7fbe27c39-kube-api-access-zsqlh\") pod \"crc-debug-mqm7t\" (UID: \"c208d478-bfb6-4483-b01b-2ef7fbe27c39\") " pod="openshift-must-gather-fjd59/crc-debug-mqm7t" Dec 11 09:15:59 crc kubenswrapper[4860]: I1211 09:15:59.510378 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c208d478-bfb6-4483-b01b-2ef7fbe27c39-host\") pod \"crc-debug-mqm7t\" (UID: \"c208d478-bfb6-4483-b01b-2ef7fbe27c39\") " pod="openshift-must-gather-fjd59/crc-debug-mqm7t" Dec 11 09:15:59 crc kubenswrapper[4860]: I1211 09:15:59.612018 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c208d478-bfb6-4483-b01b-2ef7fbe27c39-host\") pod \"crc-debug-mqm7t\" (UID: \"c208d478-bfb6-4483-b01b-2ef7fbe27c39\") " pod="openshift-must-gather-fjd59/crc-debug-mqm7t" Dec 11 09:15:59 crc kubenswrapper[4860]: I1211 09:15:59.612129 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zsqlh\" (UniqueName: \"kubernetes.io/projected/c208d478-bfb6-4483-b01b-2ef7fbe27c39-kube-api-access-zsqlh\") pod \"crc-debug-mqm7t\" (UID: \"c208d478-bfb6-4483-b01b-2ef7fbe27c39\") " pod="openshift-must-gather-fjd59/crc-debug-mqm7t" Dec 11 09:15:59 crc kubenswrapper[4860]: I1211 09:15:59.612227 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c208d478-bfb6-4483-b01b-2ef7fbe27c39-host\") pod \"crc-debug-mqm7t\" (UID: \"c208d478-bfb6-4483-b01b-2ef7fbe27c39\") " pod="openshift-must-gather-fjd59/crc-debug-mqm7t" Dec 11 09:15:59 crc kubenswrapper[4860]: I1211 09:15:59.635173 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zsqlh\" (UniqueName: \"kubernetes.io/projected/c208d478-bfb6-4483-b01b-2ef7fbe27c39-kube-api-access-zsqlh\") pod \"crc-debug-mqm7t\" (UID: \"c208d478-bfb6-4483-b01b-2ef7fbe27c39\") " pod="openshift-must-gather-fjd59/crc-debug-mqm7t" Dec 11 09:15:59 crc kubenswrapper[4860]: I1211 09:15:59.780283 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-fjd59/crc-debug-mqm7t" Dec 11 09:16:00 crc kubenswrapper[4860]: I1211 09:16:00.137866 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-fjd59/crc-debug-mqm7t" event={"ID":"c208d478-bfb6-4483-b01b-2ef7fbe27c39","Type":"ContainerStarted","Data":"03f9edfc7cfb73ed955c48cb1497915d80834e9d9c1e0e7d2a559fab81efe2e3"} Dec 11 09:16:00 crc kubenswrapper[4860]: I1211 09:16:00.138247 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-fjd59/crc-debug-mqm7t" event={"ID":"c208d478-bfb6-4483-b01b-2ef7fbe27c39","Type":"ContainerStarted","Data":"0e4bbfaa22a6ad5dc830a97e9347788cc893113304833ffe01ffc2e6c8783d59"} Dec 11 09:16:00 crc kubenswrapper[4860]: I1211 09:16:00.189901 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-fjd59/crc-debug-mqm7t" podStartSLOduration=1.189875757 podStartE2EDuration="1.189875757s" podCreationTimestamp="2025-12-11 09:15:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 09:16:00.188150423 +0000 UTC m=+3892.916669478" watchObservedRunningTime="2025-12-11 09:16:00.189875757 +0000 UTC m=+3892.918394822" Dec 11 09:16:04 crc kubenswrapper[4860]: I1211 09:16:04.578981 4860 scope.go:117] "RemoveContainer" containerID="6704d8c98ab428aac5737fa114668f6cf3119541ee0fe55ea326d429fcbe2d82" Dec 11 09:16:04 crc kubenswrapper[4860]: E1211 09:16:04.579747 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-99qgp_openshift-machine-config-operator(31c30642-6e60-41b4-a477-0d802424e0aa)\"" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" Dec 11 09:16:18 crc kubenswrapper[4860]: I1211 09:16:18.579010 4860 scope.go:117] "RemoveContainer" containerID="6704d8c98ab428aac5737fa114668f6cf3119541ee0fe55ea326d429fcbe2d82" Dec 11 09:16:18 crc kubenswrapper[4860]: E1211 09:16:18.581075 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-99qgp_openshift-machine-config-operator(31c30642-6e60-41b4-a477-0d802424e0aa)\"" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" Dec 11 09:16:27 crc kubenswrapper[4860]: I1211 09:16:27.082325 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-26cn4"] Dec 11 09:16:27 crc kubenswrapper[4860]: I1211 09:16:27.087563 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-26cn4" Dec 11 09:16:27 crc kubenswrapper[4860]: I1211 09:16:27.109587 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-26cn4"] Dec 11 09:16:27 crc kubenswrapper[4860]: I1211 09:16:27.242069 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0fae3628-0190-474c-82f2-f53bed062cdd-utilities\") pod \"certified-operators-26cn4\" (UID: \"0fae3628-0190-474c-82f2-f53bed062cdd\") " pod="openshift-marketplace/certified-operators-26cn4" Dec 11 09:16:27 crc kubenswrapper[4860]: I1211 09:16:27.242186 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0fae3628-0190-474c-82f2-f53bed062cdd-catalog-content\") pod \"certified-operators-26cn4\" (UID: \"0fae3628-0190-474c-82f2-f53bed062cdd\") " pod="openshift-marketplace/certified-operators-26cn4" Dec 11 09:16:27 crc kubenswrapper[4860]: I1211 09:16:27.242212 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s4jtm\" (UniqueName: \"kubernetes.io/projected/0fae3628-0190-474c-82f2-f53bed062cdd-kube-api-access-s4jtm\") pod \"certified-operators-26cn4\" (UID: \"0fae3628-0190-474c-82f2-f53bed062cdd\") " pod="openshift-marketplace/certified-operators-26cn4" Dec 11 09:16:27 crc kubenswrapper[4860]: I1211 09:16:27.344409 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0fae3628-0190-474c-82f2-f53bed062cdd-utilities\") pod \"certified-operators-26cn4\" (UID: \"0fae3628-0190-474c-82f2-f53bed062cdd\") " pod="openshift-marketplace/certified-operators-26cn4" Dec 11 09:16:27 crc kubenswrapper[4860]: I1211 09:16:27.344775 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0fae3628-0190-474c-82f2-f53bed062cdd-catalog-content\") pod \"certified-operators-26cn4\" (UID: \"0fae3628-0190-474c-82f2-f53bed062cdd\") " pod="openshift-marketplace/certified-operators-26cn4" Dec 11 09:16:27 crc kubenswrapper[4860]: I1211 09:16:27.344799 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s4jtm\" (UniqueName: \"kubernetes.io/projected/0fae3628-0190-474c-82f2-f53bed062cdd-kube-api-access-s4jtm\") pod \"certified-operators-26cn4\" (UID: \"0fae3628-0190-474c-82f2-f53bed062cdd\") " pod="openshift-marketplace/certified-operators-26cn4" Dec 11 09:16:27 crc kubenswrapper[4860]: I1211 09:16:27.345623 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0fae3628-0190-474c-82f2-f53bed062cdd-utilities\") pod \"certified-operators-26cn4\" (UID: \"0fae3628-0190-474c-82f2-f53bed062cdd\") " pod="openshift-marketplace/certified-operators-26cn4" Dec 11 09:16:27 crc kubenswrapper[4860]: I1211 09:16:27.345927 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0fae3628-0190-474c-82f2-f53bed062cdd-catalog-content\") pod \"certified-operators-26cn4\" (UID: \"0fae3628-0190-474c-82f2-f53bed062cdd\") " pod="openshift-marketplace/certified-operators-26cn4" Dec 11 09:16:27 crc kubenswrapper[4860]: I1211 09:16:27.376489 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s4jtm\" (UniqueName: \"kubernetes.io/projected/0fae3628-0190-474c-82f2-f53bed062cdd-kube-api-access-s4jtm\") pod \"certified-operators-26cn4\" (UID: \"0fae3628-0190-474c-82f2-f53bed062cdd\") " pod="openshift-marketplace/certified-operators-26cn4" Dec 11 09:16:27 crc kubenswrapper[4860]: I1211 09:16:27.437958 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-26cn4" Dec 11 09:16:28 crc kubenswrapper[4860]: I1211 09:16:28.264811 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-26cn4"] Dec 11 09:16:28 crc kubenswrapper[4860]: I1211 09:16:28.695374 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-26cn4" event={"ID":"0fae3628-0190-474c-82f2-f53bed062cdd","Type":"ContainerStarted","Data":"bd43c580023d92d9f1211f07e003fa49e4f05a6af0719b3fcafcf661ca8505e2"} Dec 11 09:16:28 crc kubenswrapper[4860]: I1211 09:16:28.695823 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-26cn4" event={"ID":"0fae3628-0190-474c-82f2-f53bed062cdd","Type":"ContainerStarted","Data":"54cfd064066937dbec21d44f8d1776910fecf4c227757297239f7b1925c2f316"} Dec 11 09:16:29 crc kubenswrapper[4860]: I1211 09:16:29.706800 4860 generic.go:334] "Generic (PLEG): container finished" podID="0fae3628-0190-474c-82f2-f53bed062cdd" containerID="bd43c580023d92d9f1211f07e003fa49e4f05a6af0719b3fcafcf661ca8505e2" exitCode=0 Dec 11 09:16:29 crc kubenswrapper[4860]: I1211 09:16:29.706876 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-26cn4" event={"ID":"0fae3628-0190-474c-82f2-f53bed062cdd","Type":"ContainerDied","Data":"bd43c580023d92d9f1211f07e003fa49e4f05a6af0719b3fcafcf661ca8505e2"} Dec 11 09:16:29 crc kubenswrapper[4860]: I1211 09:16:29.709145 4860 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 11 09:16:30 crc kubenswrapper[4860]: I1211 09:16:30.578846 4860 scope.go:117] "RemoveContainer" containerID="6704d8c98ab428aac5737fa114668f6cf3119541ee0fe55ea326d429fcbe2d82" Dec 11 09:16:30 crc kubenswrapper[4860]: E1211 09:16:30.579562 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-99qgp_openshift-machine-config-operator(31c30642-6e60-41b4-a477-0d802424e0aa)\"" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" Dec 11 09:16:31 crc kubenswrapper[4860]: I1211 09:16:31.735488 4860 generic.go:334] "Generic (PLEG): container finished" podID="0fae3628-0190-474c-82f2-f53bed062cdd" containerID="96c2d86c53e4843b60f065108e65fee14c55e63a7ee02cd1ccc6a27eeb25c2a5" exitCode=0 Dec 11 09:16:31 crc kubenswrapper[4860]: I1211 09:16:31.735669 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-26cn4" event={"ID":"0fae3628-0190-474c-82f2-f53bed062cdd","Type":"ContainerDied","Data":"96c2d86c53e4843b60f065108e65fee14c55e63a7ee02cd1ccc6a27eeb25c2a5"} Dec 11 09:16:32 crc kubenswrapper[4860]: I1211 09:16:32.750555 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-26cn4" event={"ID":"0fae3628-0190-474c-82f2-f53bed062cdd","Type":"ContainerStarted","Data":"59fefac103c580d9390af8933ac69331ab667365dfd196fe1ac5792eba01e711"} Dec 11 09:16:32 crc kubenswrapper[4860]: I1211 09:16:32.776511 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-26cn4" podStartSLOduration=3.277330048 podStartE2EDuration="5.776491927s" podCreationTimestamp="2025-12-11 09:16:27 +0000 UTC" firstStartedPulling="2025-12-11 09:16:29.708941099 +0000 UTC m=+3922.437460154" lastFinishedPulling="2025-12-11 09:16:32.208102978 +0000 UTC m=+3924.936622033" observedRunningTime="2025-12-11 09:16:32.771512161 +0000 UTC m=+3925.500031216" watchObservedRunningTime="2025-12-11 09:16:32.776491927 +0000 UTC m=+3925.505010982" Dec 11 09:16:37 crc kubenswrapper[4860]: I1211 09:16:37.438823 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-26cn4" Dec 11 09:16:37 crc kubenswrapper[4860]: I1211 09:16:37.439437 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-26cn4" Dec 11 09:16:37 crc kubenswrapper[4860]: I1211 09:16:37.492914 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-26cn4" Dec 11 09:16:37 crc kubenswrapper[4860]: I1211 09:16:37.846568 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-26cn4" Dec 11 09:16:37 crc kubenswrapper[4860]: I1211 09:16:37.923994 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-26cn4"] Dec 11 09:16:39 crc kubenswrapper[4860]: I1211 09:16:39.811357 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-26cn4" podUID="0fae3628-0190-474c-82f2-f53bed062cdd" containerName="registry-server" containerID="cri-o://59fefac103c580d9390af8933ac69331ab667365dfd196fe1ac5792eba01e711" gracePeriod=2 Dec 11 09:16:40 crc kubenswrapper[4860]: I1211 09:16:40.823395 4860 generic.go:334] "Generic (PLEG): container finished" podID="0fae3628-0190-474c-82f2-f53bed062cdd" containerID="59fefac103c580d9390af8933ac69331ab667365dfd196fe1ac5792eba01e711" exitCode=0 Dec 11 09:16:40 crc kubenswrapper[4860]: I1211 09:16:40.823466 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-26cn4" event={"ID":"0fae3628-0190-474c-82f2-f53bed062cdd","Type":"ContainerDied","Data":"59fefac103c580d9390af8933ac69331ab667365dfd196fe1ac5792eba01e711"} Dec 11 09:16:41 crc kubenswrapper[4860]: I1211 09:16:41.272235 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-26cn4" Dec 11 09:16:41 crc kubenswrapper[4860]: I1211 09:16:41.366906 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0fae3628-0190-474c-82f2-f53bed062cdd-catalog-content\") pod \"0fae3628-0190-474c-82f2-f53bed062cdd\" (UID: \"0fae3628-0190-474c-82f2-f53bed062cdd\") " Dec 11 09:16:41 crc kubenswrapper[4860]: I1211 09:16:41.367314 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4jtm\" (UniqueName: \"kubernetes.io/projected/0fae3628-0190-474c-82f2-f53bed062cdd-kube-api-access-s4jtm\") pod \"0fae3628-0190-474c-82f2-f53bed062cdd\" (UID: \"0fae3628-0190-474c-82f2-f53bed062cdd\") " Dec 11 09:16:41 crc kubenswrapper[4860]: I1211 09:16:41.367449 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0fae3628-0190-474c-82f2-f53bed062cdd-utilities\") pod \"0fae3628-0190-474c-82f2-f53bed062cdd\" (UID: \"0fae3628-0190-474c-82f2-f53bed062cdd\") " Dec 11 09:16:41 crc kubenswrapper[4860]: I1211 09:16:41.368423 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0fae3628-0190-474c-82f2-f53bed062cdd-utilities" (OuterVolumeSpecName: "utilities") pod "0fae3628-0190-474c-82f2-f53bed062cdd" (UID: "0fae3628-0190-474c-82f2-f53bed062cdd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 09:16:41 crc kubenswrapper[4860]: I1211 09:16:41.373024 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0fae3628-0190-474c-82f2-f53bed062cdd-kube-api-access-s4jtm" (OuterVolumeSpecName: "kube-api-access-s4jtm") pod "0fae3628-0190-474c-82f2-f53bed062cdd" (UID: "0fae3628-0190-474c-82f2-f53bed062cdd"). InnerVolumeSpecName "kube-api-access-s4jtm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 09:16:41 crc kubenswrapper[4860]: I1211 09:16:41.429346 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0fae3628-0190-474c-82f2-f53bed062cdd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0fae3628-0190-474c-82f2-f53bed062cdd" (UID: "0fae3628-0190-474c-82f2-f53bed062cdd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 09:16:41 crc kubenswrapper[4860]: I1211 09:16:41.469455 4860 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0fae3628-0190-474c-82f2-f53bed062cdd-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 09:16:41 crc kubenswrapper[4860]: I1211 09:16:41.469773 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4jtm\" (UniqueName: \"kubernetes.io/projected/0fae3628-0190-474c-82f2-f53bed062cdd-kube-api-access-s4jtm\") on node \"crc\" DevicePath \"\"" Dec 11 09:16:41 crc kubenswrapper[4860]: I1211 09:16:41.469873 4860 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0fae3628-0190-474c-82f2-f53bed062cdd-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 09:16:41 crc kubenswrapper[4860]: I1211 09:16:41.851076 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-26cn4" event={"ID":"0fae3628-0190-474c-82f2-f53bed062cdd","Type":"ContainerDied","Data":"54cfd064066937dbec21d44f8d1776910fecf4c227757297239f7b1925c2f316"} Dec 11 09:16:41 crc kubenswrapper[4860]: I1211 09:16:41.851144 4860 scope.go:117] "RemoveContainer" containerID="59fefac103c580d9390af8933ac69331ab667365dfd196fe1ac5792eba01e711" Dec 11 09:16:41 crc kubenswrapper[4860]: I1211 09:16:41.851202 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-26cn4" Dec 11 09:16:41 crc kubenswrapper[4860]: I1211 09:16:41.881104 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-26cn4"] Dec 11 09:16:41 crc kubenswrapper[4860]: I1211 09:16:41.889926 4860 scope.go:117] "RemoveContainer" containerID="96c2d86c53e4843b60f065108e65fee14c55e63a7ee02cd1ccc6a27eeb25c2a5" Dec 11 09:16:41 crc kubenswrapper[4860]: I1211 09:16:41.892606 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-26cn4"] Dec 11 09:16:41 crc kubenswrapper[4860]: I1211 09:16:41.931216 4860 scope.go:117] "RemoveContainer" containerID="bd43c580023d92d9f1211f07e003fa49e4f05a6af0719b3fcafcf661ca8505e2" Dec 11 09:16:43 crc kubenswrapper[4860]: I1211 09:16:43.579777 4860 scope.go:117] "RemoveContainer" containerID="6704d8c98ab428aac5737fa114668f6cf3119541ee0fe55ea326d429fcbe2d82" Dec 11 09:16:43 crc kubenswrapper[4860]: E1211 09:16:43.580354 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-99qgp_openshift-machine-config-operator(31c30642-6e60-41b4-a477-0d802424e0aa)\"" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" Dec 11 09:16:43 crc kubenswrapper[4860]: I1211 09:16:43.592977 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0fae3628-0190-474c-82f2-f53bed062cdd" path="/var/lib/kubelet/pods/0fae3628-0190-474c-82f2-f53bed062cdd/volumes" Dec 11 09:16:43 crc kubenswrapper[4860]: I1211 09:16:43.873080 4860 generic.go:334] "Generic (PLEG): container finished" podID="c208d478-bfb6-4483-b01b-2ef7fbe27c39" containerID="03f9edfc7cfb73ed955c48cb1497915d80834e9d9c1e0e7d2a559fab81efe2e3" exitCode=0 Dec 11 09:16:43 crc kubenswrapper[4860]: I1211 09:16:43.873126 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-fjd59/crc-debug-mqm7t" event={"ID":"c208d478-bfb6-4483-b01b-2ef7fbe27c39","Type":"ContainerDied","Data":"03f9edfc7cfb73ed955c48cb1497915d80834e9d9c1e0e7d2a559fab81efe2e3"} Dec 11 09:16:44 crc kubenswrapper[4860]: I1211 09:16:44.978131 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-fjd59/crc-debug-mqm7t" Dec 11 09:16:45 crc kubenswrapper[4860]: I1211 09:16:45.023151 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-fjd59/crc-debug-mqm7t"] Dec 11 09:16:45 crc kubenswrapper[4860]: I1211 09:16:45.033111 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-fjd59/crc-debug-mqm7t"] Dec 11 09:16:45 crc kubenswrapper[4860]: I1211 09:16:45.050241 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zsqlh\" (UniqueName: \"kubernetes.io/projected/c208d478-bfb6-4483-b01b-2ef7fbe27c39-kube-api-access-zsqlh\") pod \"c208d478-bfb6-4483-b01b-2ef7fbe27c39\" (UID: \"c208d478-bfb6-4483-b01b-2ef7fbe27c39\") " Dec 11 09:16:45 crc kubenswrapper[4860]: I1211 09:16:45.050534 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c208d478-bfb6-4483-b01b-2ef7fbe27c39-host\") pod \"c208d478-bfb6-4483-b01b-2ef7fbe27c39\" (UID: \"c208d478-bfb6-4483-b01b-2ef7fbe27c39\") " Dec 11 09:16:45 crc kubenswrapper[4860]: I1211 09:16:45.050728 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c208d478-bfb6-4483-b01b-2ef7fbe27c39-host" (OuterVolumeSpecName: "host") pod "c208d478-bfb6-4483-b01b-2ef7fbe27c39" (UID: "c208d478-bfb6-4483-b01b-2ef7fbe27c39"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 09:16:45 crc kubenswrapper[4860]: I1211 09:16:45.051421 4860 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c208d478-bfb6-4483-b01b-2ef7fbe27c39-host\") on node \"crc\" DevicePath \"\"" Dec 11 09:16:45 crc kubenswrapper[4860]: I1211 09:16:45.056465 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c208d478-bfb6-4483-b01b-2ef7fbe27c39-kube-api-access-zsqlh" (OuterVolumeSpecName: "kube-api-access-zsqlh") pod "c208d478-bfb6-4483-b01b-2ef7fbe27c39" (UID: "c208d478-bfb6-4483-b01b-2ef7fbe27c39"). InnerVolumeSpecName "kube-api-access-zsqlh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 09:16:45 crc kubenswrapper[4860]: I1211 09:16:45.153231 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zsqlh\" (UniqueName: \"kubernetes.io/projected/c208d478-bfb6-4483-b01b-2ef7fbe27c39-kube-api-access-zsqlh\") on node \"crc\" DevicePath \"\"" Dec 11 09:16:45 crc kubenswrapper[4860]: I1211 09:16:45.591225 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c208d478-bfb6-4483-b01b-2ef7fbe27c39" path="/var/lib/kubelet/pods/c208d478-bfb6-4483-b01b-2ef7fbe27c39/volumes" Dec 11 09:16:45 crc kubenswrapper[4860]: I1211 09:16:45.895260 4860 scope.go:117] "RemoveContainer" containerID="03f9edfc7cfb73ed955c48cb1497915d80834e9d9c1e0e7d2a559fab81efe2e3" Dec 11 09:16:45 crc kubenswrapper[4860]: I1211 09:16:45.895305 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-fjd59/crc-debug-mqm7t" Dec 11 09:16:46 crc kubenswrapper[4860]: I1211 09:16:46.215663 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-fjd59/crc-debug-xx2nv"] Dec 11 09:16:46 crc kubenswrapper[4860]: E1211 09:16:46.216058 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0fae3628-0190-474c-82f2-f53bed062cdd" containerName="extract-utilities" Dec 11 09:16:46 crc kubenswrapper[4860]: I1211 09:16:46.216072 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="0fae3628-0190-474c-82f2-f53bed062cdd" containerName="extract-utilities" Dec 11 09:16:46 crc kubenswrapper[4860]: E1211 09:16:46.216092 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0fae3628-0190-474c-82f2-f53bed062cdd" containerName="registry-server" Dec 11 09:16:46 crc kubenswrapper[4860]: I1211 09:16:46.216098 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="0fae3628-0190-474c-82f2-f53bed062cdd" containerName="registry-server" Dec 11 09:16:46 crc kubenswrapper[4860]: E1211 09:16:46.216114 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c208d478-bfb6-4483-b01b-2ef7fbe27c39" containerName="container-00" Dec 11 09:16:46 crc kubenswrapper[4860]: I1211 09:16:46.216119 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="c208d478-bfb6-4483-b01b-2ef7fbe27c39" containerName="container-00" Dec 11 09:16:46 crc kubenswrapper[4860]: E1211 09:16:46.216129 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0fae3628-0190-474c-82f2-f53bed062cdd" containerName="extract-content" Dec 11 09:16:46 crc kubenswrapper[4860]: I1211 09:16:46.216135 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="0fae3628-0190-474c-82f2-f53bed062cdd" containerName="extract-content" Dec 11 09:16:46 crc kubenswrapper[4860]: I1211 09:16:46.216312 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="0fae3628-0190-474c-82f2-f53bed062cdd" containerName="registry-server" Dec 11 09:16:46 crc kubenswrapper[4860]: I1211 09:16:46.216332 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="c208d478-bfb6-4483-b01b-2ef7fbe27c39" containerName="container-00" Dec 11 09:16:46 crc kubenswrapper[4860]: I1211 09:16:46.217061 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-fjd59/crc-debug-xx2nv" Dec 11 09:16:46 crc kubenswrapper[4860]: I1211 09:16:46.273968 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g8z6h\" (UniqueName: \"kubernetes.io/projected/4830b0c0-fa1d-4091-ae44-9c3ea901590b-kube-api-access-g8z6h\") pod \"crc-debug-xx2nv\" (UID: \"4830b0c0-fa1d-4091-ae44-9c3ea901590b\") " pod="openshift-must-gather-fjd59/crc-debug-xx2nv" Dec 11 09:16:46 crc kubenswrapper[4860]: I1211 09:16:46.274107 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4830b0c0-fa1d-4091-ae44-9c3ea901590b-host\") pod \"crc-debug-xx2nv\" (UID: \"4830b0c0-fa1d-4091-ae44-9c3ea901590b\") " pod="openshift-must-gather-fjd59/crc-debug-xx2nv" Dec 11 09:16:46 crc kubenswrapper[4860]: I1211 09:16:46.375844 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4830b0c0-fa1d-4091-ae44-9c3ea901590b-host\") pod \"crc-debug-xx2nv\" (UID: \"4830b0c0-fa1d-4091-ae44-9c3ea901590b\") " pod="openshift-must-gather-fjd59/crc-debug-xx2nv" Dec 11 09:16:46 crc kubenswrapper[4860]: I1211 09:16:46.375996 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g8z6h\" (UniqueName: \"kubernetes.io/projected/4830b0c0-fa1d-4091-ae44-9c3ea901590b-kube-api-access-g8z6h\") pod \"crc-debug-xx2nv\" (UID: \"4830b0c0-fa1d-4091-ae44-9c3ea901590b\") " pod="openshift-must-gather-fjd59/crc-debug-xx2nv" Dec 11 09:16:46 crc kubenswrapper[4860]: I1211 09:16:46.376108 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4830b0c0-fa1d-4091-ae44-9c3ea901590b-host\") pod \"crc-debug-xx2nv\" (UID: \"4830b0c0-fa1d-4091-ae44-9c3ea901590b\") " pod="openshift-must-gather-fjd59/crc-debug-xx2nv" Dec 11 09:16:46 crc kubenswrapper[4860]: I1211 09:16:46.402060 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g8z6h\" (UniqueName: \"kubernetes.io/projected/4830b0c0-fa1d-4091-ae44-9c3ea901590b-kube-api-access-g8z6h\") pod \"crc-debug-xx2nv\" (UID: \"4830b0c0-fa1d-4091-ae44-9c3ea901590b\") " pod="openshift-must-gather-fjd59/crc-debug-xx2nv" Dec 11 09:16:46 crc kubenswrapper[4860]: I1211 09:16:46.538974 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-fjd59/crc-debug-xx2nv" Dec 11 09:16:46 crc kubenswrapper[4860]: I1211 09:16:46.907177 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-fjd59/crc-debug-xx2nv" event={"ID":"4830b0c0-fa1d-4091-ae44-9c3ea901590b","Type":"ContainerStarted","Data":"d2e89457c427f86ded2d8e106401b16edb8234b5a02e346ea774d6b94fc639fe"} Dec 11 09:16:46 crc kubenswrapper[4860]: I1211 09:16:46.907243 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-fjd59/crc-debug-xx2nv" event={"ID":"4830b0c0-fa1d-4091-ae44-9c3ea901590b","Type":"ContainerStarted","Data":"13de61ec1cda484d8ea5e02b76fbdacc8b3f3211aefff145e3091bee1854bf63"} Dec 11 09:16:46 crc kubenswrapper[4860]: I1211 09:16:46.927372 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-fjd59/crc-debug-xx2nv" podStartSLOduration=0.927344686 podStartE2EDuration="927.344686ms" podCreationTimestamp="2025-12-11 09:16:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 09:16:46.917841817 +0000 UTC m=+3939.646360872" watchObservedRunningTime="2025-12-11 09:16:46.927344686 +0000 UTC m=+3939.655863741" Dec 11 09:16:47 crc kubenswrapper[4860]: I1211 09:16:47.920657 4860 generic.go:334] "Generic (PLEG): container finished" podID="4830b0c0-fa1d-4091-ae44-9c3ea901590b" containerID="d2e89457c427f86ded2d8e106401b16edb8234b5a02e346ea774d6b94fc639fe" exitCode=0 Dec 11 09:16:47 crc kubenswrapper[4860]: I1211 09:16:47.920759 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-fjd59/crc-debug-xx2nv" event={"ID":"4830b0c0-fa1d-4091-ae44-9c3ea901590b","Type":"ContainerDied","Data":"d2e89457c427f86ded2d8e106401b16edb8234b5a02e346ea774d6b94fc639fe"} Dec 11 09:16:49 crc kubenswrapper[4860]: I1211 09:16:49.052497 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-fjd59/crc-debug-xx2nv" Dec 11 09:16:49 crc kubenswrapper[4860]: I1211 09:16:49.096533 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-fjd59/crc-debug-xx2nv"] Dec 11 09:16:49 crc kubenswrapper[4860]: I1211 09:16:49.105097 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-fjd59/crc-debug-xx2nv"] Dec 11 09:16:49 crc kubenswrapper[4860]: I1211 09:16:49.144987 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g8z6h\" (UniqueName: \"kubernetes.io/projected/4830b0c0-fa1d-4091-ae44-9c3ea901590b-kube-api-access-g8z6h\") pod \"4830b0c0-fa1d-4091-ae44-9c3ea901590b\" (UID: \"4830b0c0-fa1d-4091-ae44-9c3ea901590b\") " Dec 11 09:16:49 crc kubenswrapper[4860]: I1211 09:16:49.145056 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4830b0c0-fa1d-4091-ae44-9c3ea901590b-host\") pod \"4830b0c0-fa1d-4091-ae44-9c3ea901590b\" (UID: \"4830b0c0-fa1d-4091-ae44-9c3ea901590b\") " Dec 11 09:16:49 crc kubenswrapper[4860]: I1211 09:16:49.145215 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4830b0c0-fa1d-4091-ae44-9c3ea901590b-host" (OuterVolumeSpecName: "host") pod "4830b0c0-fa1d-4091-ae44-9c3ea901590b" (UID: "4830b0c0-fa1d-4091-ae44-9c3ea901590b"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 09:16:49 crc kubenswrapper[4860]: I1211 09:16:49.145443 4860 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4830b0c0-fa1d-4091-ae44-9c3ea901590b-host\") on node \"crc\" DevicePath \"\"" Dec 11 09:16:49 crc kubenswrapper[4860]: I1211 09:16:49.151786 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4830b0c0-fa1d-4091-ae44-9c3ea901590b-kube-api-access-g8z6h" (OuterVolumeSpecName: "kube-api-access-g8z6h") pod "4830b0c0-fa1d-4091-ae44-9c3ea901590b" (UID: "4830b0c0-fa1d-4091-ae44-9c3ea901590b"). InnerVolumeSpecName "kube-api-access-g8z6h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 09:16:49 crc kubenswrapper[4860]: I1211 09:16:49.247514 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g8z6h\" (UniqueName: \"kubernetes.io/projected/4830b0c0-fa1d-4091-ae44-9c3ea901590b-kube-api-access-g8z6h\") on node \"crc\" DevicePath \"\"" Dec 11 09:16:49 crc kubenswrapper[4860]: I1211 09:16:49.596171 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4830b0c0-fa1d-4091-ae44-9c3ea901590b" path="/var/lib/kubelet/pods/4830b0c0-fa1d-4091-ae44-9c3ea901590b/volumes" Dec 11 09:16:49 crc kubenswrapper[4860]: I1211 09:16:49.942937 4860 scope.go:117] "RemoveContainer" containerID="d2e89457c427f86ded2d8e106401b16edb8234b5a02e346ea774d6b94fc639fe" Dec 11 09:16:49 crc kubenswrapper[4860]: I1211 09:16:49.942992 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-fjd59/crc-debug-xx2nv" Dec 11 09:16:50 crc kubenswrapper[4860]: I1211 09:16:50.263841 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-fjd59/crc-debug-s69hh"] Dec 11 09:16:50 crc kubenswrapper[4860]: E1211 09:16:50.264680 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4830b0c0-fa1d-4091-ae44-9c3ea901590b" containerName="container-00" Dec 11 09:16:50 crc kubenswrapper[4860]: I1211 09:16:50.264698 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="4830b0c0-fa1d-4091-ae44-9c3ea901590b" containerName="container-00" Dec 11 09:16:50 crc kubenswrapper[4860]: I1211 09:16:50.264941 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="4830b0c0-fa1d-4091-ae44-9c3ea901590b" containerName="container-00" Dec 11 09:16:50 crc kubenswrapper[4860]: I1211 09:16:50.265924 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-fjd59/crc-debug-s69hh" Dec 11 09:16:50 crc kubenswrapper[4860]: I1211 09:16:50.375798 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-db7zx\" (UniqueName: \"kubernetes.io/projected/ae00af42-ebd6-4098-83f7-c89dcd32d6a5-kube-api-access-db7zx\") pod \"crc-debug-s69hh\" (UID: \"ae00af42-ebd6-4098-83f7-c89dcd32d6a5\") " pod="openshift-must-gather-fjd59/crc-debug-s69hh" Dec 11 09:16:50 crc kubenswrapper[4860]: I1211 09:16:50.376461 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ae00af42-ebd6-4098-83f7-c89dcd32d6a5-host\") pod \"crc-debug-s69hh\" (UID: \"ae00af42-ebd6-4098-83f7-c89dcd32d6a5\") " pod="openshift-must-gather-fjd59/crc-debug-s69hh" Dec 11 09:16:50 crc kubenswrapper[4860]: I1211 09:16:50.478868 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ae00af42-ebd6-4098-83f7-c89dcd32d6a5-host\") pod \"crc-debug-s69hh\" (UID: \"ae00af42-ebd6-4098-83f7-c89dcd32d6a5\") " pod="openshift-must-gather-fjd59/crc-debug-s69hh" Dec 11 09:16:50 crc kubenswrapper[4860]: I1211 09:16:50.478977 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-db7zx\" (UniqueName: \"kubernetes.io/projected/ae00af42-ebd6-4098-83f7-c89dcd32d6a5-kube-api-access-db7zx\") pod \"crc-debug-s69hh\" (UID: \"ae00af42-ebd6-4098-83f7-c89dcd32d6a5\") " pod="openshift-must-gather-fjd59/crc-debug-s69hh" Dec 11 09:16:50 crc kubenswrapper[4860]: I1211 09:16:50.479061 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ae00af42-ebd6-4098-83f7-c89dcd32d6a5-host\") pod \"crc-debug-s69hh\" (UID: \"ae00af42-ebd6-4098-83f7-c89dcd32d6a5\") " pod="openshift-must-gather-fjd59/crc-debug-s69hh" Dec 11 09:16:50 crc kubenswrapper[4860]: I1211 09:16:50.498466 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-db7zx\" (UniqueName: \"kubernetes.io/projected/ae00af42-ebd6-4098-83f7-c89dcd32d6a5-kube-api-access-db7zx\") pod \"crc-debug-s69hh\" (UID: \"ae00af42-ebd6-4098-83f7-c89dcd32d6a5\") " pod="openshift-must-gather-fjd59/crc-debug-s69hh" Dec 11 09:16:50 crc kubenswrapper[4860]: I1211 09:16:50.585533 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-fjd59/crc-debug-s69hh" Dec 11 09:16:50 crc kubenswrapper[4860]: W1211 09:16:50.618221 4860 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podae00af42_ebd6_4098_83f7_c89dcd32d6a5.slice/crio-cd9911e5e687658cf3d6615a88506e9bc29a1cc695493bb384e24140f324ed36 WatchSource:0}: Error finding container cd9911e5e687658cf3d6615a88506e9bc29a1cc695493bb384e24140f324ed36: Status 404 returned error can't find the container with id cd9911e5e687658cf3d6615a88506e9bc29a1cc695493bb384e24140f324ed36 Dec 11 09:16:50 crc kubenswrapper[4860]: I1211 09:16:50.951104 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-fjd59/crc-debug-s69hh" event={"ID":"ae00af42-ebd6-4098-83f7-c89dcd32d6a5","Type":"ContainerStarted","Data":"cd9911e5e687658cf3d6615a88506e9bc29a1cc695493bb384e24140f324ed36"} Dec 11 09:16:51 crc kubenswrapper[4860]: I1211 09:16:51.964592 4860 generic.go:334] "Generic (PLEG): container finished" podID="ae00af42-ebd6-4098-83f7-c89dcd32d6a5" containerID="c3913d353b7af7b67d65954ca3ace1235533f604450e744d5da288f6b59f06ca" exitCode=0 Dec 11 09:16:51 crc kubenswrapper[4860]: I1211 09:16:51.964698 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-fjd59/crc-debug-s69hh" event={"ID":"ae00af42-ebd6-4098-83f7-c89dcd32d6a5","Type":"ContainerDied","Data":"c3913d353b7af7b67d65954ca3ace1235533f604450e744d5da288f6b59f06ca"} Dec 11 09:16:52 crc kubenswrapper[4860]: I1211 09:16:52.002658 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-fjd59/crc-debug-s69hh"] Dec 11 09:16:52 crc kubenswrapper[4860]: I1211 09:16:52.013614 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-fjd59/crc-debug-s69hh"] Dec 11 09:16:53 crc kubenswrapper[4860]: I1211 09:16:53.099090 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-fjd59/crc-debug-s69hh" Dec 11 09:16:53 crc kubenswrapper[4860]: I1211 09:16:53.233211 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-db7zx\" (UniqueName: \"kubernetes.io/projected/ae00af42-ebd6-4098-83f7-c89dcd32d6a5-kube-api-access-db7zx\") pod \"ae00af42-ebd6-4098-83f7-c89dcd32d6a5\" (UID: \"ae00af42-ebd6-4098-83f7-c89dcd32d6a5\") " Dec 11 09:16:53 crc kubenswrapper[4860]: I1211 09:16:53.233338 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ae00af42-ebd6-4098-83f7-c89dcd32d6a5-host\") pod \"ae00af42-ebd6-4098-83f7-c89dcd32d6a5\" (UID: \"ae00af42-ebd6-4098-83f7-c89dcd32d6a5\") " Dec 11 09:16:53 crc kubenswrapper[4860]: I1211 09:16:53.233964 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ae00af42-ebd6-4098-83f7-c89dcd32d6a5-host" (OuterVolumeSpecName: "host") pod "ae00af42-ebd6-4098-83f7-c89dcd32d6a5" (UID: "ae00af42-ebd6-4098-83f7-c89dcd32d6a5"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 11 09:16:53 crc kubenswrapper[4860]: I1211 09:16:53.239438 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae00af42-ebd6-4098-83f7-c89dcd32d6a5-kube-api-access-db7zx" (OuterVolumeSpecName: "kube-api-access-db7zx") pod "ae00af42-ebd6-4098-83f7-c89dcd32d6a5" (UID: "ae00af42-ebd6-4098-83f7-c89dcd32d6a5"). InnerVolumeSpecName "kube-api-access-db7zx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 09:16:53 crc kubenswrapper[4860]: I1211 09:16:53.336059 4860 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ae00af42-ebd6-4098-83f7-c89dcd32d6a5-host\") on node \"crc\" DevicePath \"\"" Dec 11 09:16:53 crc kubenswrapper[4860]: I1211 09:16:53.336096 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-db7zx\" (UniqueName: \"kubernetes.io/projected/ae00af42-ebd6-4098-83f7-c89dcd32d6a5-kube-api-access-db7zx\") on node \"crc\" DevicePath \"\"" Dec 11 09:16:53 crc kubenswrapper[4860]: I1211 09:16:53.590166 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ae00af42-ebd6-4098-83f7-c89dcd32d6a5" path="/var/lib/kubelet/pods/ae00af42-ebd6-4098-83f7-c89dcd32d6a5/volumes" Dec 11 09:16:53 crc kubenswrapper[4860]: I1211 09:16:53.984068 4860 scope.go:117] "RemoveContainer" containerID="c3913d353b7af7b67d65954ca3ace1235533f604450e744d5da288f6b59f06ca" Dec 11 09:16:53 crc kubenswrapper[4860]: I1211 09:16:53.984116 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-fjd59/crc-debug-s69hh" Dec 11 09:16:54 crc kubenswrapper[4860]: I1211 09:16:54.578860 4860 scope.go:117] "RemoveContainer" containerID="6704d8c98ab428aac5737fa114668f6cf3119541ee0fe55ea326d429fcbe2d82" Dec 11 09:16:54 crc kubenswrapper[4860]: E1211 09:16:54.580876 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-99qgp_openshift-machine-config-operator(31c30642-6e60-41b4-a477-0d802424e0aa)\"" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" Dec 11 09:17:09 crc kubenswrapper[4860]: I1211 09:17:09.579481 4860 scope.go:117] "RemoveContainer" containerID="6704d8c98ab428aac5737fa114668f6cf3119541ee0fe55ea326d429fcbe2d82" Dec 11 09:17:09 crc kubenswrapper[4860]: E1211 09:17:09.580259 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-99qgp_openshift-machine-config-operator(31c30642-6e60-41b4-a477-0d802424e0aa)\"" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" Dec 11 09:17:11 crc kubenswrapper[4860]: I1211 09:17:11.629024 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-6bbf49f946-sw5n4_4a941157-652e-4efa-9165-ba55161bfb51/barbican-api/0.log" Dec 11 09:17:11 crc kubenswrapper[4860]: I1211 09:17:11.729011 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-6bbf49f946-sw5n4_4a941157-652e-4efa-9165-ba55161bfb51/barbican-api-log/0.log" Dec 11 09:17:11 crc kubenswrapper[4860]: I1211 09:17:11.815404 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-684c9ddd98-t29bd_d8954df2-1f06-4bbe-bb33-a5ac8fdf9675/barbican-keystone-listener/0.log" Dec 11 09:17:11 crc kubenswrapper[4860]: I1211 09:17:11.864712 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-684c9ddd98-t29bd_d8954df2-1f06-4bbe-bb33-a5ac8fdf9675/barbican-keystone-listener-log/0.log" Dec 11 09:17:12 crc kubenswrapper[4860]: I1211 09:17:12.022525 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-568d665c75-4vptt_a7e1def2-f417-40f5-a7dc-c63148858073/barbican-worker/0.log" Dec 11 09:17:12 crc kubenswrapper[4860]: I1211 09:17:12.074334 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-568d665c75-4vptt_a7e1def2-f417-40f5-a7dc-c63148858073/barbican-worker-log/0.log" Dec 11 09:17:12 crc kubenswrapper[4860]: I1211 09:17:12.176790 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-tm5lx_aa177563-2177-4361-b45a-7ffe6f88da96/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Dec 11 09:17:12 crc kubenswrapper[4860]: I1211 09:17:12.337372 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_f0d70280-9b14-4d9d-9b7a-7818f388c4fc/ceilometer-central-agent/0.log" Dec 11 09:17:12 crc kubenswrapper[4860]: I1211 09:17:12.398477 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_f0d70280-9b14-4d9d-9b7a-7818f388c4fc/proxy-httpd/0.log" Dec 11 09:17:12 crc kubenswrapper[4860]: I1211 09:17:12.411845 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_f0d70280-9b14-4d9d-9b7a-7818f388c4fc/ceilometer-notification-agent/0.log" Dec 11 09:17:12 crc kubenswrapper[4860]: I1211 09:17:12.493343 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_f0d70280-9b14-4d9d-9b7a-7818f388c4fc/sg-core/0.log" Dec 11 09:17:12 crc kubenswrapper[4860]: I1211 09:17:12.604262 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_8a6d65b6-5393-4edc-a9c3-3660c8860fd1/cinder-api-log/0.log" Dec 11 09:17:12 crc kubenswrapper[4860]: I1211 09:17:12.612432 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_8a6d65b6-5393-4edc-a9c3-3660c8860fd1/cinder-api/0.log" Dec 11 09:17:12 crc kubenswrapper[4860]: I1211 09:17:12.801391 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_5dc4bc3d-458e-4251-81f7-68b5b47a7275/probe/0.log" Dec 11 09:17:12 crc kubenswrapper[4860]: I1211 09:17:12.811545 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_5dc4bc3d-458e-4251-81f7-68b5b47a7275/cinder-scheduler/0.log" Dec 11 09:17:12 crc kubenswrapper[4860]: I1211 09:17:12.958975 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-7zg5t_155181af-f5ab-41fb-9c0f-820c03b91b23/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 11 09:17:13 crc kubenswrapper[4860]: I1211 09:17:13.014976 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-bv6xw_2651e591-78fd-410d-b8a0-bcb0da2066bb/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 11 09:17:13 crc kubenswrapper[4860]: I1211 09:17:13.178338 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-8c6f6df99-zx6zh_2e567b4f-94b9-4662-825a-045e124c7948/init/0.log" Dec 11 09:17:13 crc kubenswrapper[4860]: I1211 09:17:13.359156 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-8c6f6df99-zx6zh_2e567b4f-94b9-4662-825a-045e124c7948/init/0.log" Dec 11 09:17:13 crc kubenswrapper[4860]: I1211 09:17:13.392127 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-zvzpf_1fb47321-3b60-4961-abf2-87ee1d57a58a/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Dec 11 09:17:13 crc kubenswrapper[4860]: I1211 09:17:13.393054 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-8c6f6df99-zx6zh_2e567b4f-94b9-4662-825a-045e124c7948/dnsmasq-dns/0.log" Dec 11 09:17:13 crc kubenswrapper[4860]: I1211 09:17:13.595578 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_f51c05dd-1881-4644-9a40-7e6e2174c591/glance-log/0.log" Dec 11 09:17:13 crc kubenswrapper[4860]: I1211 09:17:13.617426 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_f51c05dd-1881-4644-9a40-7e6e2174c591/glance-httpd/0.log" Dec 11 09:17:13 crc kubenswrapper[4860]: I1211 09:17:13.806328 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_b7bf9af0-6051-4d90-8918-9fd00cb869cd/glance-httpd/0.log" Dec 11 09:17:13 crc kubenswrapper[4860]: I1211 09:17:13.840112 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_b7bf9af0-6051-4d90-8918-9fd00cb869cd/glance-log/0.log" Dec 11 09:17:13 crc kubenswrapper[4860]: I1211 09:17:13.942784 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-6b4697fc68-vnd26_78b0f75e-d447-4541-be11-9afe4549d2e1/horizon/0.log" Dec 11 09:17:14 crc kubenswrapper[4860]: I1211 09:17:14.150889 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-lwc56_ddaa83e7-d6db-4169-abdb-226ec9f50204/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Dec 11 09:17:14 crc kubenswrapper[4860]: I1211 09:17:14.387787 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-j8pcd_f9f3c692-bca8-49fb-a602-cc747b566849/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 11 09:17:14 crc kubenswrapper[4860]: I1211 09:17:14.407948 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-6b4697fc68-vnd26_78b0f75e-d447-4541-be11-9afe4549d2e1/horizon-log/0.log" Dec 11 09:17:14 crc kubenswrapper[4860]: I1211 09:17:14.629659 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-748d5bd978-4sz6v_e208d74e-f50f-443c-aa2d-eccdb584413d/keystone-api/0.log" Dec 11 09:17:14 crc kubenswrapper[4860]: I1211 09:17:14.652556 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29424061-gq8tl_8fb3c7c5-8c06-4e05-ad79-11f0671e16da/keystone-cron/0.log" Dec 11 09:17:14 crc kubenswrapper[4860]: I1211 09:17:14.850554 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_eceff086-19de-4aed-8ac3-5242d9b949ab/kube-state-metrics/0.log" Dec 11 09:17:14 crc kubenswrapper[4860]: I1211 09:17:14.911562 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-5z2qw_b0eed689-a827-47b6-81ba-91e76db6f35b/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Dec 11 09:17:15 crc kubenswrapper[4860]: I1211 09:17:15.237018 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-5fccb6555c-xwc9b_7739fee0-ccf5-49e9-bc21-4acd841daf55/neutron-api/0.log" Dec 11 09:17:15 crc kubenswrapper[4860]: I1211 09:17:15.304160 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-5fccb6555c-xwc9b_7739fee0-ccf5-49e9-bc21-4acd841daf55/neutron-httpd/0.log" Dec 11 09:17:15 crc kubenswrapper[4860]: I1211 09:17:15.471250 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-gfrj6_70691461-ffb3-4a01-aa11-bbe665e77b35/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Dec 11 09:17:16 crc kubenswrapper[4860]: I1211 09:17:16.118138 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_581da06a-7d8d-48dd-a5d9-d95ec1539a9f/nova-cell0-conductor-conductor/0.log" Dec 11 09:17:16 crc kubenswrapper[4860]: I1211 09:17:16.146703 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_fcbd0284-56fd-4cb0-895e-78e6a571868a/nova-api-log/0.log" Dec 11 09:17:16 crc kubenswrapper[4860]: I1211 09:17:16.516115 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_773668a5-7449-4ced-b8d0-509faab0eabb/nova-cell1-conductor-conductor/0.log" Dec 11 09:17:16 crc kubenswrapper[4860]: I1211 09:17:16.554475 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_f4725b54-31da-4b85-a4f7-200b1d6a3069/nova-cell1-novncproxy-novncproxy/0.log" Dec 11 09:17:16 crc kubenswrapper[4860]: I1211 09:17:16.584638 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_fcbd0284-56fd-4cb0-895e-78e6a571868a/nova-api-api/0.log" Dec 11 09:17:16 crc kubenswrapper[4860]: I1211 09:17:16.785708 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-x42ls_fac07a15-8332-4374-aaca-6577650e3183/nova-edpm-deployment-openstack-edpm-ipam/0.log" Dec 11 09:17:16 crc kubenswrapper[4860]: I1211 09:17:16.909916 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_4b75f4be-3810-4273-a18e-381ab6c6ef2b/nova-metadata-log/0.log" Dec 11 09:17:17 crc kubenswrapper[4860]: I1211 09:17:17.256622 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_08a3b710-84c1-4bbc-9547-a350c0d76318/nova-scheduler-scheduler/0.log" Dec 11 09:17:17 crc kubenswrapper[4860]: I1211 09:17:17.335209 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_a0838f24-14b5-4018-ad26-8cd0394b4488/mysql-bootstrap/0.log" Dec 11 09:17:18 crc kubenswrapper[4860]: I1211 09:17:18.199270 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_a0838f24-14b5-4018-ad26-8cd0394b4488/mysql-bootstrap/0.log" Dec 11 09:17:18 crc kubenswrapper[4860]: I1211 09:17:18.244800 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_a0838f24-14b5-4018-ad26-8cd0394b4488/galera/0.log" Dec 11 09:17:18 crc kubenswrapper[4860]: I1211 09:17:18.256248 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_4b75f4be-3810-4273-a18e-381ab6c6ef2b/nova-metadata-metadata/0.log" Dec 11 09:17:18 crc kubenswrapper[4860]: I1211 09:17:18.273697 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-x2slc"] Dec 11 09:17:18 crc kubenswrapper[4860]: E1211 09:17:18.277715 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae00af42-ebd6-4098-83f7-c89dcd32d6a5" containerName="container-00" Dec 11 09:17:18 crc kubenswrapper[4860]: I1211 09:17:18.279697 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae00af42-ebd6-4098-83f7-c89dcd32d6a5" containerName="container-00" Dec 11 09:17:18 crc kubenswrapper[4860]: I1211 09:17:18.280054 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae00af42-ebd6-4098-83f7-c89dcd32d6a5" containerName="container-00" Dec 11 09:17:18 crc kubenswrapper[4860]: I1211 09:17:18.281938 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-x2slc" Dec 11 09:17:18 crc kubenswrapper[4860]: I1211 09:17:18.292340 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-x2slc"] Dec 11 09:17:18 crc kubenswrapper[4860]: I1211 09:17:18.417601 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tp85h\" (UniqueName: \"kubernetes.io/projected/c8e0c201-97a3-494e-af29-c3530510b379-kube-api-access-tp85h\") pod \"community-operators-x2slc\" (UID: \"c8e0c201-97a3-494e-af29-c3530510b379\") " pod="openshift-marketplace/community-operators-x2slc" Dec 11 09:17:18 crc kubenswrapper[4860]: I1211 09:17:18.417756 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8e0c201-97a3-494e-af29-c3530510b379-utilities\") pod \"community-operators-x2slc\" (UID: \"c8e0c201-97a3-494e-af29-c3530510b379\") " pod="openshift-marketplace/community-operators-x2slc" Dec 11 09:17:18 crc kubenswrapper[4860]: I1211 09:17:18.417848 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8e0c201-97a3-494e-af29-c3530510b379-catalog-content\") pod \"community-operators-x2slc\" (UID: \"c8e0c201-97a3-494e-af29-c3530510b379\") " pod="openshift-marketplace/community-operators-x2slc" Dec 11 09:17:18 crc kubenswrapper[4860]: I1211 09:17:18.444228 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_c0d5e719-9d28-4e31-9349-f882c4b72567/mysql-bootstrap/0.log" Dec 11 09:17:18 crc kubenswrapper[4860]: I1211 09:17:18.520021 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tp85h\" (UniqueName: \"kubernetes.io/projected/c8e0c201-97a3-494e-af29-c3530510b379-kube-api-access-tp85h\") pod \"community-operators-x2slc\" (UID: \"c8e0c201-97a3-494e-af29-c3530510b379\") " pod="openshift-marketplace/community-operators-x2slc" Dec 11 09:17:18 crc kubenswrapper[4860]: I1211 09:17:18.520092 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8e0c201-97a3-494e-af29-c3530510b379-utilities\") pod \"community-operators-x2slc\" (UID: \"c8e0c201-97a3-494e-af29-c3530510b379\") " pod="openshift-marketplace/community-operators-x2slc" Dec 11 09:17:18 crc kubenswrapper[4860]: I1211 09:17:18.520156 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8e0c201-97a3-494e-af29-c3530510b379-catalog-content\") pod \"community-operators-x2slc\" (UID: \"c8e0c201-97a3-494e-af29-c3530510b379\") " pod="openshift-marketplace/community-operators-x2slc" Dec 11 09:17:18 crc kubenswrapper[4860]: I1211 09:17:18.520813 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8e0c201-97a3-494e-af29-c3530510b379-catalog-content\") pod \"community-operators-x2slc\" (UID: \"c8e0c201-97a3-494e-af29-c3530510b379\") " pod="openshift-marketplace/community-operators-x2slc" Dec 11 09:17:18 crc kubenswrapper[4860]: I1211 09:17:18.521064 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8e0c201-97a3-494e-af29-c3530510b379-utilities\") pod \"community-operators-x2slc\" (UID: \"c8e0c201-97a3-494e-af29-c3530510b379\") " pod="openshift-marketplace/community-operators-x2slc" Dec 11 09:17:18 crc kubenswrapper[4860]: I1211 09:17:18.559015 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tp85h\" (UniqueName: \"kubernetes.io/projected/c8e0c201-97a3-494e-af29-c3530510b379-kube-api-access-tp85h\") pod \"community-operators-x2slc\" (UID: \"c8e0c201-97a3-494e-af29-c3530510b379\") " pod="openshift-marketplace/community-operators-x2slc" Dec 11 09:17:18 crc kubenswrapper[4860]: I1211 09:17:18.623882 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-x2slc" Dec 11 09:17:18 crc kubenswrapper[4860]: I1211 09:17:18.853782 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_6f6e1905-c5e6-495b-9076-fa5edd9a01ac/openstackclient/0.log" Dec 11 09:17:18 crc kubenswrapper[4860]: I1211 09:17:18.866143 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_c0d5e719-9d28-4e31-9349-f882c4b72567/mysql-bootstrap/0.log" Dec 11 09:17:19 crc kubenswrapper[4860]: I1211 09:17:19.037847 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_c0d5e719-9d28-4e31-9349-f882c4b72567/galera/0.log" Dec 11 09:17:19 crc kubenswrapper[4860]: I1211 09:17:19.171555 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-dbdtz_e834572f-9650-45d3-b978-c39141b44b74/ovn-controller/0.log" Dec 11 09:17:19 crc kubenswrapper[4860]: I1211 09:17:19.302314 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-x2slc"] Dec 11 09:17:19 crc kubenswrapper[4860]: I1211 09:17:19.350526 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-lgzmd_06395cef-744f-402e-844e-275a5025b81f/openstack-network-exporter/0.log" Dec 11 09:17:19 crc kubenswrapper[4860]: I1211 09:17:19.999628 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-q86ch_e73beb7d-c40c-491b-99f8-fabe037df9ba/ovsdb-server-init/0.log" Dec 11 09:17:20 crc kubenswrapper[4860]: I1211 09:17:20.182432 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-q86ch_e73beb7d-c40c-491b-99f8-fabe037df9ba/ovs-vswitchd/0.log" Dec 11 09:17:20 crc kubenswrapper[4860]: I1211 09:17:20.240404 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-q86ch_e73beb7d-c40c-491b-99f8-fabe037df9ba/ovsdb-server/0.log" Dec 11 09:17:20 crc kubenswrapper[4860]: I1211 09:17:20.249523 4860 generic.go:334] "Generic (PLEG): container finished" podID="c8e0c201-97a3-494e-af29-c3530510b379" containerID="9e13597c5fff95d04c27e72c24228b7d4bfb1df540658a2b4a3a657207d858ce" exitCode=0 Dec 11 09:17:20 crc kubenswrapper[4860]: I1211 09:17:20.249593 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x2slc" event={"ID":"c8e0c201-97a3-494e-af29-c3530510b379","Type":"ContainerDied","Data":"9e13597c5fff95d04c27e72c24228b7d4bfb1df540658a2b4a3a657207d858ce"} Dec 11 09:17:20 crc kubenswrapper[4860]: I1211 09:17:20.249682 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x2slc" event={"ID":"c8e0c201-97a3-494e-af29-c3530510b379","Type":"ContainerStarted","Data":"ba2cd81bd2daa945dde73cb6939a2519e60b3462499225f413c9ad53e9c298c8"} Dec 11 09:17:20 crc kubenswrapper[4860]: I1211 09:17:20.264265 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-q86ch_e73beb7d-c40c-491b-99f8-fabe037df9ba/ovsdb-server-init/0.log" Dec 11 09:17:20 crc kubenswrapper[4860]: I1211 09:17:20.466659 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-8ptks_c2dda62c-b6df-481f-a388-a6bfc3b02c9e/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Dec 11 09:17:20 crc kubenswrapper[4860]: I1211 09:17:20.492561 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_ad53aa7d-2d02-4201-a90b-05bc8cca2f68/openstack-network-exporter/0.log" Dec 11 09:17:20 crc kubenswrapper[4860]: I1211 09:17:20.499413 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_ad53aa7d-2d02-4201-a90b-05bc8cca2f68/ovn-northd/0.log" Dec 11 09:17:20 crc kubenswrapper[4860]: I1211 09:17:20.729599 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_c017f36f-48a3-4919-ac22-6a85a5c73530/ovsdbserver-nb/0.log" Dec 11 09:17:20 crc kubenswrapper[4860]: I1211 09:17:20.751727 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_c017f36f-48a3-4919-ac22-6a85a5c73530/openstack-network-exporter/0.log" Dec 11 09:17:21 crc kubenswrapper[4860]: I1211 09:17:21.060278 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_9657082d-ff9c-4a54-8b4d-54e6ea411590/ovsdbserver-sb/0.log" Dec 11 09:17:21 crc kubenswrapper[4860]: I1211 09:17:21.148583 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_9657082d-ff9c-4a54-8b4d-54e6ea411590/openstack-network-exporter/0.log" Dec 11 09:17:21 crc kubenswrapper[4860]: I1211 09:17:21.267354 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-55ccdb6bbd-grdt6_78683923-6a60-4a9a-bb81-51ae61f55376/placement-api/0.log" Dec 11 09:17:21 crc kubenswrapper[4860]: I1211 09:17:21.384936 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-55ccdb6bbd-grdt6_78683923-6a60-4a9a-bb81-51ae61f55376/placement-log/0.log" Dec 11 09:17:21 crc kubenswrapper[4860]: I1211 09:17:21.483293 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_9cfe9bd2-a1fe-4cae-8f48-c007bfd95961/setup-container/0.log" Dec 11 09:17:21 crc kubenswrapper[4860]: I1211 09:17:21.580479 4860 scope.go:117] "RemoveContainer" containerID="6704d8c98ab428aac5737fa114668f6cf3119541ee0fe55ea326d429fcbe2d82" Dec 11 09:17:21 crc kubenswrapper[4860]: E1211 09:17:21.580781 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-99qgp_openshift-machine-config-operator(31c30642-6e60-41b4-a477-0d802424e0aa)\"" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" Dec 11 09:17:21 crc kubenswrapper[4860]: I1211 09:17:21.713471 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_a1d839a1-4174-4abc-9e3d-5579603687e6/setup-container/0.log" Dec 11 09:17:21 crc kubenswrapper[4860]: I1211 09:17:21.714015 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_9cfe9bd2-a1fe-4cae-8f48-c007bfd95961/rabbitmq/0.log" Dec 11 09:17:21 crc kubenswrapper[4860]: I1211 09:17:21.750293 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_9cfe9bd2-a1fe-4cae-8f48-c007bfd95961/setup-container/0.log" Dec 11 09:17:21 crc kubenswrapper[4860]: I1211 09:17:21.981713 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_a1d839a1-4174-4abc-9e3d-5579603687e6/setup-container/0.log" Dec 11 09:17:22 crc kubenswrapper[4860]: I1211 09:17:22.029730 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_a1d839a1-4174-4abc-9e3d-5579603687e6/rabbitmq/0.log" Dec 11 09:17:22 crc kubenswrapper[4860]: I1211 09:17:22.067496 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-g9pmk_07a4b9f9-a1a7-416d-89a5-e8a57c2484cc/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 11 09:17:22 crc kubenswrapper[4860]: I1211 09:17:22.277049 4860 generic.go:334] "Generic (PLEG): container finished" podID="c8e0c201-97a3-494e-af29-c3530510b379" containerID="47a3cbe6fa9dfae93a88adc7632ad869679dea26db9f238c000a85a430c53495" exitCode=0 Dec 11 09:17:22 crc kubenswrapper[4860]: I1211 09:17:22.277091 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x2slc" event={"ID":"c8e0c201-97a3-494e-af29-c3530510b379","Type":"ContainerDied","Data":"47a3cbe6fa9dfae93a88adc7632ad869679dea26db9f238c000a85a430c53495"} Dec 11 09:17:22 crc kubenswrapper[4860]: I1211 09:17:22.305673 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-zb8rn_a84f5e60-25bb-4412-a38c-fa4e7a32c177/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Dec 11 09:17:22 crc kubenswrapper[4860]: I1211 09:17:22.376420 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-2vqvr_b0ccaf6b-d322-440b-87c4-560b61547f52/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Dec 11 09:17:22 crc kubenswrapper[4860]: I1211 09:17:22.634923 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-c6n8q_bc86a63a-f36d-4893-a55e-df029aa586ea/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 11 09:17:22 crc kubenswrapper[4860]: I1211 09:17:22.723076 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-rl5hl_6a0f9071-e3f7-43f2-a0ec-8ab22fd9c785/ssh-known-hosts-edpm-deployment/0.log" Dec 11 09:17:22 crc kubenswrapper[4860]: I1211 09:17:22.929549 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-865bbb47ff-66xpd_ca8cd60f-8a69-44f4-9512-387dd82af334/proxy-server/0.log" Dec 11 09:17:23 crc kubenswrapper[4860]: I1211 09:17:23.070953 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-865bbb47ff-66xpd_ca8cd60f-8a69-44f4-9512-387dd82af334/proxy-httpd/0.log" Dec 11 09:17:23 crc kubenswrapper[4860]: I1211 09:17:23.119041 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-bphmx_79bcab38-b87b-44ff-ba2b-6027fd20719d/swift-ring-rebalance/0.log" Dec 11 09:17:23 crc kubenswrapper[4860]: I1211 09:17:23.288228 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x2slc" event={"ID":"c8e0c201-97a3-494e-af29-c3530510b379","Type":"ContainerStarted","Data":"6d1e26e88f7d765b81df17c4c8051b4137910930e4e905220e286ae937c793c2"} Dec 11 09:17:23 crc kubenswrapper[4860]: I1211 09:17:23.308755 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-x2slc" podStartSLOduration=2.796403572 podStartE2EDuration="5.308731815s" podCreationTimestamp="2025-12-11 09:17:18 +0000 UTC" firstStartedPulling="2025-12-11 09:17:20.261058762 +0000 UTC m=+3972.989577817" lastFinishedPulling="2025-12-11 09:17:22.773387005 +0000 UTC m=+3975.501906060" observedRunningTime="2025-12-11 09:17:23.306361981 +0000 UTC m=+3976.034881056" watchObservedRunningTime="2025-12-11 09:17:23.308731815 +0000 UTC m=+3976.037250890" Dec 11 09:17:23 crc kubenswrapper[4860]: I1211 09:17:23.359663 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a672e32c-a924-4080-a7d1-bf7012b6725b/account-reaper/0.log" Dec 11 09:17:23 crc kubenswrapper[4860]: I1211 09:17:23.362183 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a672e32c-a924-4080-a7d1-bf7012b6725b/account-auditor/0.log" Dec 11 09:17:23 crc kubenswrapper[4860]: I1211 09:17:23.397946 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a672e32c-a924-4080-a7d1-bf7012b6725b/account-replicator/0.log" Dec 11 09:17:23 crc kubenswrapper[4860]: I1211 09:17:23.465740 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a672e32c-a924-4080-a7d1-bf7012b6725b/account-server/0.log" Dec 11 09:17:23 crc kubenswrapper[4860]: I1211 09:17:23.626536 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a672e32c-a924-4080-a7d1-bf7012b6725b/container-replicator/0.log" Dec 11 09:17:23 crc kubenswrapper[4860]: I1211 09:17:23.633673 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a672e32c-a924-4080-a7d1-bf7012b6725b/container-auditor/0.log" Dec 11 09:17:23 crc kubenswrapper[4860]: I1211 09:17:23.642421 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a672e32c-a924-4080-a7d1-bf7012b6725b/container-server/0.log" Dec 11 09:17:23 crc kubenswrapper[4860]: I1211 09:17:23.772712 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a672e32c-a924-4080-a7d1-bf7012b6725b/container-updater/0.log" Dec 11 09:17:24 crc kubenswrapper[4860]: I1211 09:17:24.055216 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a672e32c-a924-4080-a7d1-bf7012b6725b/object-expirer/0.log" Dec 11 09:17:24 crc kubenswrapper[4860]: I1211 09:17:24.058119 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a672e32c-a924-4080-a7d1-bf7012b6725b/object-replicator/0.log" Dec 11 09:17:24 crc kubenswrapper[4860]: I1211 09:17:24.096261 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a672e32c-a924-4080-a7d1-bf7012b6725b/object-auditor/0.log" Dec 11 09:17:24 crc kubenswrapper[4860]: I1211 09:17:24.227917 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a672e32c-a924-4080-a7d1-bf7012b6725b/object-server/0.log" Dec 11 09:17:24 crc kubenswrapper[4860]: I1211 09:17:24.302392 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a672e32c-a924-4080-a7d1-bf7012b6725b/object-updater/0.log" Dec 11 09:17:24 crc kubenswrapper[4860]: I1211 09:17:24.343392 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a672e32c-a924-4080-a7d1-bf7012b6725b/rsync/0.log" Dec 11 09:17:24 crc kubenswrapper[4860]: I1211 09:17:24.410343 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a672e32c-a924-4080-a7d1-bf7012b6725b/swift-recon-cron/0.log" Dec 11 09:17:24 crc kubenswrapper[4860]: I1211 09:17:24.688529 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-wmwhw_beb319f7-9b48-413d-9561-e6c0dc62238f/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Dec 11 09:17:24 crc kubenswrapper[4860]: I1211 09:17:24.728440 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_3b659d05-76c9-4b04-a05e-a7dd488e9852/tempest-tests-tempest-tests-runner/0.log" Dec 11 09:17:24 crc kubenswrapper[4860]: I1211 09:17:24.993926 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-ljr6l_27ee41d4-1afd-4700-828c-7388caf498c5/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 11 09:17:25 crc kubenswrapper[4860]: I1211 09:17:25.005034 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_b658df4a-886e-49de-b953-87539b28329b/test-operator-logs-container/0.log" Dec 11 09:17:25 crc kubenswrapper[4860]: I1211 09:17:25.525819 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-xhr7k"] Dec 11 09:17:25 crc kubenswrapper[4860]: I1211 09:17:25.527981 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xhr7k" Dec 11 09:17:25 crc kubenswrapper[4860]: I1211 09:17:25.556713 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xhr7k"] Dec 11 09:17:25 crc kubenswrapper[4860]: I1211 09:17:25.566862 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c880ffe1-896a-41c5-8cea-4a6ca96ca971-catalog-content\") pod \"redhat-operators-xhr7k\" (UID: \"c880ffe1-896a-41c5-8cea-4a6ca96ca971\") " pod="openshift-marketplace/redhat-operators-xhr7k" Dec 11 09:17:25 crc kubenswrapper[4860]: I1211 09:17:25.566958 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v5j85\" (UniqueName: \"kubernetes.io/projected/c880ffe1-896a-41c5-8cea-4a6ca96ca971-kube-api-access-v5j85\") pod \"redhat-operators-xhr7k\" (UID: \"c880ffe1-896a-41c5-8cea-4a6ca96ca971\") " pod="openshift-marketplace/redhat-operators-xhr7k" Dec 11 09:17:25 crc kubenswrapper[4860]: I1211 09:17:25.567011 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c880ffe1-896a-41c5-8cea-4a6ca96ca971-utilities\") pod \"redhat-operators-xhr7k\" (UID: \"c880ffe1-896a-41c5-8cea-4a6ca96ca971\") " pod="openshift-marketplace/redhat-operators-xhr7k" Dec 11 09:17:25 crc kubenswrapper[4860]: I1211 09:17:25.670201 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c880ffe1-896a-41c5-8cea-4a6ca96ca971-catalog-content\") pod \"redhat-operators-xhr7k\" (UID: \"c880ffe1-896a-41c5-8cea-4a6ca96ca971\") " pod="openshift-marketplace/redhat-operators-xhr7k" Dec 11 09:17:25 crc kubenswrapper[4860]: I1211 09:17:25.670299 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v5j85\" (UniqueName: \"kubernetes.io/projected/c880ffe1-896a-41c5-8cea-4a6ca96ca971-kube-api-access-v5j85\") pod \"redhat-operators-xhr7k\" (UID: \"c880ffe1-896a-41c5-8cea-4a6ca96ca971\") " pod="openshift-marketplace/redhat-operators-xhr7k" Dec 11 09:17:25 crc kubenswrapper[4860]: I1211 09:17:25.670374 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c880ffe1-896a-41c5-8cea-4a6ca96ca971-utilities\") pod \"redhat-operators-xhr7k\" (UID: \"c880ffe1-896a-41c5-8cea-4a6ca96ca971\") " pod="openshift-marketplace/redhat-operators-xhr7k" Dec 11 09:17:25 crc kubenswrapper[4860]: I1211 09:17:25.671686 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c880ffe1-896a-41c5-8cea-4a6ca96ca971-catalog-content\") pod \"redhat-operators-xhr7k\" (UID: \"c880ffe1-896a-41c5-8cea-4a6ca96ca971\") " pod="openshift-marketplace/redhat-operators-xhr7k" Dec 11 09:17:25 crc kubenswrapper[4860]: I1211 09:17:25.672136 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c880ffe1-896a-41c5-8cea-4a6ca96ca971-utilities\") pod \"redhat-operators-xhr7k\" (UID: \"c880ffe1-896a-41c5-8cea-4a6ca96ca971\") " pod="openshift-marketplace/redhat-operators-xhr7k" Dec 11 09:17:25 crc kubenswrapper[4860]: I1211 09:17:25.692402 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v5j85\" (UniqueName: \"kubernetes.io/projected/c880ffe1-896a-41c5-8cea-4a6ca96ca971-kube-api-access-v5j85\") pod \"redhat-operators-xhr7k\" (UID: \"c880ffe1-896a-41c5-8cea-4a6ca96ca971\") " pod="openshift-marketplace/redhat-operators-xhr7k" Dec 11 09:17:25 crc kubenswrapper[4860]: I1211 09:17:25.860788 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xhr7k" Dec 11 09:17:26 crc kubenswrapper[4860]: I1211 09:17:26.365042 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xhr7k"] Dec 11 09:17:27 crc kubenswrapper[4860]: I1211 09:17:27.341012 4860 generic.go:334] "Generic (PLEG): container finished" podID="c880ffe1-896a-41c5-8cea-4a6ca96ca971" containerID="906d59b91fd7d000faff22bd48e1b733b7a36c777a8bc783b32b1e03024f71ac" exitCode=0 Dec 11 09:17:27 crc kubenswrapper[4860]: I1211 09:17:27.341283 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xhr7k" event={"ID":"c880ffe1-896a-41c5-8cea-4a6ca96ca971","Type":"ContainerDied","Data":"906d59b91fd7d000faff22bd48e1b733b7a36c777a8bc783b32b1e03024f71ac"} Dec 11 09:17:27 crc kubenswrapper[4860]: I1211 09:17:27.341309 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xhr7k" event={"ID":"c880ffe1-896a-41c5-8cea-4a6ca96ca971","Type":"ContainerStarted","Data":"232c9369a35039184728d1c227862f4ab075939b6559169a60b25921f2301bc5"} Dec 11 09:17:28 crc kubenswrapper[4860]: I1211 09:17:28.624508 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-x2slc" Dec 11 09:17:28 crc kubenswrapper[4860]: I1211 09:17:28.625045 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-x2slc" Dec 11 09:17:28 crc kubenswrapper[4860]: I1211 09:17:28.672891 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-x2slc" Dec 11 09:17:29 crc kubenswrapper[4860]: I1211 09:17:29.367237 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xhr7k" event={"ID":"c880ffe1-896a-41c5-8cea-4a6ca96ca971","Type":"ContainerStarted","Data":"acc9f0b777498fea85afdccd27a92519387204dda1675d4462967d8728e15280"} Dec 11 09:17:29 crc kubenswrapper[4860]: I1211 09:17:29.438314 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-x2slc" Dec 11 09:17:30 crc kubenswrapper[4860]: I1211 09:17:30.300442 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-x2slc"] Dec 11 09:17:31 crc kubenswrapper[4860]: I1211 09:17:31.385090 4860 generic.go:334] "Generic (PLEG): container finished" podID="c880ffe1-896a-41c5-8cea-4a6ca96ca971" containerID="acc9f0b777498fea85afdccd27a92519387204dda1675d4462967d8728e15280" exitCode=0 Dec 11 09:17:31 crc kubenswrapper[4860]: I1211 09:17:31.385146 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xhr7k" event={"ID":"c880ffe1-896a-41c5-8cea-4a6ca96ca971","Type":"ContainerDied","Data":"acc9f0b777498fea85afdccd27a92519387204dda1675d4462967d8728e15280"} Dec 11 09:17:31 crc kubenswrapper[4860]: I1211 09:17:31.386038 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-x2slc" podUID="c8e0c201-97a3-494e-af29-c3530510b379" containerName="registry-server" containerID="cri-o://6d1e26e88f7d765b81df17c4c8051b4137910930e4e905220e286ae937c793c2" gracePeriod=2 Dec 11 09:17:32 crc kubenswrapper[4860]: I1211 09:17:32.284978 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_1752934c-842c-45b6-88ec-0fc804374532/memcached/0.log" Dec 11 09:17:34 crc kubenswrapper[4860]: I1211 09:17:34.082997 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-x2slc" Dec 11 09:17:34 crc kubenswrapper[4860]: I1211 09:17:34.135688 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tp85h\" (UniqueName: \"kubernetes.io/projected/c8e0c201-97a3-494e-af29-c3530510b379-kube-api-access-tp85h\") pod \"c8e0c201-97a3-494e-af29-c3530510b379\" (UID: \"c8e0c201-97a3-494e-af29-c3530510b379\") " Dec 11 09:17:34 crc kubenswrapper[4860]: I1211 09:17:34.135791 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8e0c201-97a3-494e-af29-c3530510b379-catalog-content\") pod \"c8e0c201-97a3-494e-af29-c3530510b379\" (UID: \"c8e0c201-97a3-494e-af29-c3530510b379\") " Dec 11 09:17:34 crc kubenswrapper[4860]: I1211 09:17:34.135982 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8e0c201-97a3-494e-af29-c3530510b379-utilities\") pod \"c8e0c201-97a3-494e-af29-c3530510b379\" (UID: \"c8e0c201-97a3-494e-af29-c3530510b379\") " Dec 11 09:17:34 crc kubenswrapper[4860]: I1211 09:17:34.137452 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c8e0c201-97a3-494e-af29-c3530510b379-utilities" (OuterVolumeSpecName: "utilities") pod "c8e0c201-97a3-494e-af29-c3530510b379" (UID: "c8e0c201-97a3-494e-af29-c3530510b379"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 09:17:34 crc kubenswrapper[4860]: I1211 09:17:34.159159 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8e0c201-97a3-494e-af29-c3530510b379-kube-api-access-tp85h" (OuterVolumeSpecName: "kube-api-access-tp85h") pod "c8e0c201-97a3-494e-af29-c3530510b379" (UID: "c8e0c201-97a3-494e-af29-c3530510b379"). InnerVolumeSpecName "kube-api-access-tp85h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 09:17:34 crc kubenswrapper[4860]: I1211 09:17:34.199062 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c8e0c201-97a3-494e-af29-c3530510b379-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c8e0c201-97a3-494e-af29-c3530510b379" (UID: "c8e0c201-97a3-494e-af29-c3530510b379"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 09:17:34 crc kubenswrapper[4860]: I1211 09:17:34.240702 4860 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8e0c201-97a3-494e-af29-c3530510b379-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 09:17:34 crc kubenswrapper[4860]: I1211 09:17:34.240775 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tp85h\" (UniqueName: \"kubernetes.io/projected/c8e0c201-97a3-494e-af29-c3530510b379-kube-api-access-tp85h\") on node \"crc\" DevicePath \"\"" Dec 11 09:17:34 crc kubenswrapper[4860]: I1211 09:17:34.240791 4860 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8e0c201-97a3-494e-af29-c3530510b379-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 09:17:34 crc kubenswrapper[4860]: I1211 09:17:34.415161 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xhr7k" event={"ID":"c880ffe1-896a-41c5-8cea-4a6ca96ca971","Type":"ContainerStarted","Data":"4145a557ba02326aac48e6d24fc6a4abc9c4e016cc8fa534787b3cbc52f7b340"} Dec 11 09:17:34 crc kubenswrapper[4860]: I1211 09:17:34.422230 4860 generic.go:334] "Generic (PLEG): container finished" podID="c8e0c201-97a3-494e-af29-c3530510b379" containerID="6d1e26e88f7d765b81df17c4c8051b4137910930e4e905220e286ae937c793c2" exitCode=0 Dec 11 09:17:34 crc kubenswrapper[4860]: I1211 09:17:34.422289 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x2slc" event={"ID":"c8e0c201-97a3-494e-af29-c3530510b379","Type":"ContainerDied","Data":"6d1e26e88f7d765b81df17c4c8051b4137910930e4e905220e286ae937c793c2"} Dec 11 09:17:34 crc kubenswrapper[4860]: I1211 09:17:34.422326 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x2slc" event={"ID":"c8e0c201-97a3-494e-af29-c3530510b379","Type":"ContainerDied","Data":"ba2cd81bd2daa945dde73cb6939a2519e60b3462499225f413c9ad53e9c298c8"} Dec 11 09:17:34 crc kubenswrapper[4860]: I1211 09:17:34.422347 4860 scope.go:117] "RemoveContainer" containerID="6d1e26e88f7d765b81df17c4c8051b4137910930e4e905220e286ae937c793c2" Dec 11 09:17:34 crc kubenswrapper[4860]: I1211 09:17:34.422522 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-x2slc" Dec 11 09:17:34 crc kubenswrapper[4860]: I1211 09:17:34.446176 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-xhr7k" podStartSLOduration=2.933030251 podStartE2EDuration="9.446150808s" podCreationTimestamp="2025-12-11 09:17:25 +0000 UTC" firstStartedPulling="2025-12-11 09:17:27.350389893 +0000 UTC m=+3980.078908948" lastFinishedPulling="2025-12-11 09:17:33.86351045 +0000 UTC m=+3986.592029505" observedRunningTime="2025-12-11 09:17:34.438863628 +0000 UTC m=+3987.167382683" watchObservedRunningTime="2025-12-11 09:17:34.446150808 +0000 UTC m=+3987.174669863" Dec 11 09:17:34 crc kubenswrapper[4860]: I1211 09:17:34.473016 4860 scope.go:117] "RemoveContainer" containerID="47a3cbe6fa9dfae93a88adc7632ad869679dea26db9f238c000a85a430c53495" Dec 11 09:17:34 crc kubenswrapper[4860]: I1211 09:17:34.481426 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-x2slc"] Dec 11 09:17:34 crc kubenswrapper[4860]: I1211 09:17:34.515319 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-x2slc"] Dec 11 09:17:34 crc kubenswrapper[4860]: I1211 09:17:34.542383 4860 scope.go:117] "RemoveContainer" containerID="9e13597c5fff95d04c27e72c24228b7d4bfb1df540658a2b4a3a657207d858ce" Dec 11 09:17:34 crc kubenswrapper[4860]: I1211 09:17:34.575472 4860 scope.go:117] "RemoveContainer" containerID="6d1e26e88f7d765b81df17c4c8051b4137910930e4e905220e286ae937c793c2" Dec 11 09:17:34 crc kubenswrapper[4860]: E1211 09:17:34.576882 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6d1e26e88f7d765b81df17c4c8051b4137910930e4e905220e286ae937c793c2\": container with ID starting with 6d1e26e88f7d765b81df17c4c8051b4137910930e4e905220e286ae937c793c2 not found: ID does not exist" containerID="6d1e26e88f7d765b81df17c4c8051b4137910930e4e905220e286ae937c793c2" Dec 11 09:17:34 crc kubenswrapper[4860]: I1211 09:17:34.576929 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d1e26e88f7d765b81df17c4c8051b4137910930e4e905220e286ae937c793c2"} err="failed to get container status \"6d1e26e88f7d765b81df17c4c8051b4137910930e4e905220e286ae937c793c2\": rpc error: code = NotFound desc = could not find container \"6d1e26e88f7d765b81df17c4c8051b4137910930e4e905220e286ae937c793c2\": container with ID starting with 6d1e26e88f7d765b81df17c4c8051b4137910930e4e905220e286ae937c793c2 not found: ID does not exist" Dec 11 09:17:34 crc kubenswrapper[4860]: I1211 09:17:34.576959 4860 scope.go:117] "RemoveContainer" containerID="47a3cbe6fa9dfae93a88adc7632ad869679dea26db9f238c000a85a430c53495" Dec 11 09:17:34 crc kubenswrapper[4860]: E1211 09:17:34.578517 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"47a3cbe6fa9dfae93a88adc7632ad869679dea26db9f238c000a85a430c53495\": container with ID starting with 47a3cbe6fa9dfae93a88adc7632ad869679dea26db9f238c000a85a430c53495 not found: ID does not exist" containerID="47a3cbe6fa9dfae93a88adc7632ad869679dea26db9f238c000a85a430c53495" Dec 11 09:17:34 crc kubenswrapper[4860]: I1211 09:17:34.578561 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"47a3cbe6fa9dfae93a88adc7632ad869679dea26db9f238c000a85a430c53495"} err="failed to get container status \"47a3cbe6fa9dfae93a88adc7632ad869679dea26db9f238c000a85a430c53495\": rpc error: code = NotFound desc = could not find container \"47a3cbe6fa9dfae93a88adc7632ad869679dea26db9f238c000a85a430c53495\": container with ID starting with 47a3cbe6fa9dfae93a88adc7632ad869679dea26db9f238c000a85a430c53495 not found: ID does not exist" Dec 11 09:17:34 crc kubenswrapper[4860]: I1211 09:17:34.578591 4860 scope.go:117] "RemoveContainer" containerID="9e13597c5fff95d04c27e72c24228b7d4bfb1df540658a2b4a3a657207d858ce" Dec 11 09:17:34 crc kubenswrapper[4860]: E1211 09:17:34.580548 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9e13597c5fff95d04c27e72c24228b7d4bfb1df540658a2b4a3a657207d858ce\": container with ID starting with 9e13597c5fff95d04c27e72c24228b7d4bfb1df540658a2b4a3a657207d858ce not found: ID does not exist" containerID="9e13597c5fff95d04c27e72c24228b7d4bfb1df540658a2b4a3a657207d858ce" Dec 11 09:17:34 crc kubenswrapper[4860]: I1211 09:17:34.580587 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e13597c5fff95d04c27e72c24228b7d4bfb1df540658a2b4a3a657207d858ce"} err="failed to get container status \"9e13597c5fff95d04c27e72c24228b7d4bfb1df540658a2b4a3a657207d858ce\": rpc error: code = NotFound desc = could not find container \"9e13597c5fff95d04c27e72c24228b7d4bfb1df540658a2b4a3a657207d858ce\": container with ID starting with 9e13597c5fff95d04c27e72c24228b7d4bfb1df540658a2b4a3a657207d858ce not found: ID does not exist" Dec 11 09:17:35 crc kubenswrapper[4860]: I1211 09:17:35.579596 4860 scope.go:117] "RemoveContainer" containerID="6704d8c98ab428aac5737fa114668f6cf3119541ee0fe55ea326d429fcbe2d82" Dec 11 09:17:35 crc kubenswrapper[4860]: E1211 09:17:35.579926 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-99qgp_openshift-machine-config-operator(31c30642-6e60-41b4-a477-0d802424e0aa)\"" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" Dec 11 09:17:35 crc kubenswrapper[4860]: I1211 09:17:35.591035 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c8e0c201-97a3-494e-af29-c3530510b379" path="/var/lib/kubelet/pods/c8e0c201-97a3-494e-af29-c3530510b379/volumes" Dec 11 09:17:35 crc kubenswrapper[4860]: I1211 09:17:35.861488 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-xhr7k" Dec 11 09:17:35 crc kubenswrapper[4860]: I1211 09:17:35.861896 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-xhr7k" Dec 11 09:17:36 crc kubenswrapper[4860]: I1211 09:17:36.915793 4860 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-xhr7k" podUID="c880ffe1-896a-41c5-8cea-4a6ca96ca971" containerName="registry-server" probeResult="failure" output=< Dec 11 09:17:36 crc kubenswrapper[4860]: timeout: failed to connect service ":50051" within 1s Dec 11 09:17:36 crc kubenswrapper[4860]: > Dec 11 09:17:45 crc kubenswrapper[4860]: I1211 09:17:45.918470 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-xhr7k" Dec 11 09:17:45 crc kubenswrapper[4860]: I1211 09:17:45.973334 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-xhr7k" Dec 11 09:17:46 crc kubenswrapper[4860]: I1211 09:17:46.165188 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xhr7k"] Dec 11 09:17:47 crc kubenswrapper[4860]: I1211 09:17:47.564486 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-xhr7k" podUID="c880ffe1-896a-41c5-8cea-4a6ca96ca971" containerName="registry-server" containerID="cri-o://4145a557ba02326aac48e6d24fc6a4abc9c4e016cc8fa534787b3cbc52f7b340" gracePeriod=2 Dec 11 09:17:48 crc kubenswrapper[4860]: I1211 09:17:48.102717 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xhr7k" Dec 11 09:17:48 crc kubenswrapper[4860]: I1211 09:17:48.218203 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c880ffe1-896a-41c5-8cea-4a6ca96ca971-utilities\") pod \"c880ffe1-896a-41c5-8cea-4a6ca96ca971\" (UID: \"c880ffe1-896a-41c5-8cea-4a6ca96ca971\") " Dec 11 09:17:48 crc kubenswrapper[4860]: I1211 09:17:48.218364 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v5j85\" (UniqueName: \"kubernetes.io/projected/c880ffe1-896a-41c5-8cea-4a6ca96ca971-kube-api-access-v5j85\") pod \"c880ffe1-896a-41c5-8cea-4a6ca96ca971\" (UID: \"c880ffe1-896a-41c5-8cea-4a6ca96ca971\") " Dec 11 09:17:48 crc kubenswrapper[4860]: I1211 09:17:48.218423 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c880ffe1-896a-41c5-8cea-4a6ca96ca971-catalog-content\") pod \"c880ffe1-896a-41c5-8cea-4a6ca96ca971\" (UID: \"c880ffe1-896a-41c5-8cea-4a6ca96ca971\") " Dec 11 09:17:48 crc kubenswrapper[4860]: I1211 09:17:48.219431 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c880ffe1-896a-41c5-8cea-4a6ca96ca971-utilities" (OuterVolumeSpecName: "utilities") pod "c880ffe1-896a-41c5-8cea-4a6ca96ca971" (UID: "c880ffe1-896a-41c5-8cea-4a6ca96ca971"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 09:17:48 crc kubenswrapper[4860]: I1211 09:17:48.224187 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c880ffe1-896a-41c5-8cea-4a6ca96ca971-kube-api-access-v5j85" (OuterVolumeSpecName: "kube-api-access-v5j85") pod "c880ffe1-896a-41c5-8cea-4a6ca96ca971" (UID: "c880ffe1-896a-41c5-8cea-4a6ca96ca971"). InnerVolumeSpecName "kube-api-access-v5j85". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 09:17:48 crc kubenswrapper[4860]: I1211 09:17:48.320792 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v5j85\" (UniqueName: \"kubernetes.io/projected/c880ffe1-896a-41c5-8cea-4a6ca96ca971-kube-api-access-v5j85\") on node \"crc\" DevicePath \"\"" Dec 11 09:17:48 crc kubenswrapper[4860]: I1211 09:17:48.321126 4860 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c880ffe1-896a-41c5-8cea-4a6ca96ca971-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 09:17:48 crc kubenswrapper[4860]: I1211 09:17:48.341811 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c880ffe1-896a-41c5-8cea-4a6ca96ca971-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c880ffe1-896a-41c5-8cea-4a6ca96ca971" (UID: "c880ffe1-896a-41c5-8cea-4a6ca96ca971"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 09:17:48 crc kubenswrapper[4860]: I1211 09:17:48.423034 4860 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c880ffe1-896a-41c5-8cea-4a6ca96ca971-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 09:17:48 crc kubenswrapper[4860]: I1211 09:17:48.576411 4860 generic.go:334] "Generic (PLEG): container finished" podID="c880ffe1-896a-41c5-8cea-4a6ca96ca971" containerID="4145a557ba02326aac48e6d24fc6a4abc9c4e016cc8fa534787b3cbc52f7b340" exitCode=0 Dec 11 09:17:48 crc kubenswrapper[4860]: I1211 09:17:48.576451 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xhr7k" event={"ID":"c880ffe1-896a-41c5-8cea-4a6ca96ca971","Type":"ContainerDied","Data":"4145a557ba02326aac48e6d24fc6a4abc9c4e016cc8fa534787b3cbc52f7b340"} Dec 11 09:17:48 crc kubenswrapper[4860]: I1211 09:17:48.576469 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xhr7k" Dec 11 09:17:48 crc kubenswrapper[4860]: I1211 09:17:48.576486 4860 scope.go:117] "RemoveContainer" containerID="4145a557ba02326aac48e6d24fc6a4abc9c4e016cc8fa534787b3cbc52f7b340" Dec 11 09:17:48 crc kubenswrapper[4860]: I1211 09:17:48.576476 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xhr7k" event={"ID":"c880ffe1-896a-41c5-8cea-4a6ca96ca971","Type":"ContainerDied","Data":"232c9369a35039184728d1c227862f4ab075939b6559169a60b25921f2301bc5"} Dec 11 09:17:48 crc kubenswrapper[4860]: I1211 09:17:48.578278 4860 scope.go:117] "RemoveContainer" containerID="6704d8c98ab428aac5737fa114668f6cf3119541ee0fe55ea326d429fcbe2d82" Dec 11 09:17:48 crc kubenswrapper[4860]: E1211 09:17:48.578555 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-99qgp_openshift-machine-config-operator(31c30642-6e60-41b4-a477-0d802424e0aa)\"" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" Dec 11 09:17:48 crc kubenswrapper[4860]: I1211 09:17:48.598713 4860 scope.go:117] "RemoveContainer" containerID="acc9f0b777498fea85afdccd27a92519387204dda1675d4462967d8728e15280" Dec 11 09:17:48 crc kubenswrapper[4860]: I1211 09:17:48.613248 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xhr7k"] Dec 11 09:17:48 crc kubenswrapper[4860]: I1211 09:17:48.621599 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-xhr7k"] Dec 11 09:17:48 crc kubenswrapper[4860]: I1211 09:17:48.794689 4860 scope.go:117] "RemoveContainer" containerID="906d59b91fd7d000faff22bd48e1b733b7a36c777a8bc783b32b1e03024f71ac" Dec 11 09:17:48 crc kubenswrapper[4860]: I1211 09:17:48.859764 4860 scope.go:117] "RemoveContainer" containerID="4145a557ba02326aac48e6d24fc6a4abc9c4e016cc8fa534787b3cbc52f7b340" Dec 11 09:17:48 crc kubenswrapper[4860]: E1211 09:17:48.860247 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4145a557ba02326aac48e6d24fc6a4abc9c4e016cc8fa534787b3cbc52f7b340\": container with ID starting with 4145a557ba02326aac48e6d24fc6a4abc9c4e016cc8fa534787b3cbc52f7b340 not found: ID does not exist" containerID="4145a557ba02326aac48e6d24fc6a4abc9c4e016cc8fa534787b3cbc52f7b340" Dec 11 09:17:48 crc kubenswrapper[4860]: I1211 09:17:48.860284 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4145a557ba02326aac48e6d24fc6a4abc9c4e016cc8fa534787b3cbc52f7b340"} err="failed to get container status \"4145a557ba02326aac48e6d24fc6a4abc9c4e016cc8fa534787b3cbc52f7b340\": rpc error: code = NotFound desc = could not find container \"4145a557ba02326aac48e6d24fc6a4abc9c4e016cc8fa534787b3cbc52f7b340\": container with ID starting with 4145a557ba02326aac48e6d24fc6a4abc9c4e016cc8fa534787b3cbc52f7b340 not found: ID does not exist" Dec 11 09:17:48 crc kubenswrapper[4860]: I1211 09:17:48.860308 4860 scope.go:117] "RemoveContainer" containerID="acc9f0b777498fea85afdccd27a92519387204dda1675d4462967d8728e15280" Dec 11 09:17:48 crc kubenswrapper[4860]: E1211 09:17:48.860593 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"acc9f0b777498fea85afdccd27a92519387204dda1675d4462967d8728e15280\": container with ID starting with acc9f0b777498fea85afdccd27a92519387204dda1675d4462967d8728e15280 not found: ID does not exist" containerID="acc9f0b777498fea85afdccd27a92519387204dda1675d4462967d8728e15280" Dec 11 09:17:48 crc kubenswrapper[4860]: I1211 09:17:48.860616 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"acc9f0b777498fea85afdccd27a92519387204dda1675d4462967d8728e15280"} err="failed to get container status \"acc9f0b777498fea85afdccd27a92519387204dda1675d4462967d8728e15280\": rpc error: code = NotFound desc = could not find container \"acc9f0b777498fea85afdccd27a92519387204dda1675d4462967d8728e15280\": container with ID starting with acc9f0b777498fea85afdccd27a92519387204dda1675d4462967d8728e15280 not found: ID does not exist" Dec 11 09:17:48 crc kubenswrapper[4860]: I1211 09:17:48.860630 4860 scope.go:117] "RemoveContainer" containerID="906d59b91fd7d000faff22bd48e1b733b7a36c777a8bc783b32b1e03024f71ac" Dec 11 09:17:48 crc kubenswrapper[4860]: E1211 09:17:48.860893 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"906d59b91fd7d000faff22bd48e1b733b7a36c777a8bc783b32b1e03024f71ac\": container with ID starting with 906d59b91fd7d000faff22bd48e1b733b7a36c777a8bc783b32b1e03024f71ac not found: ID does not exist" containerID="906d59b91fd7d000faff22bd48e1b733b7a36c777a8bc783b32b1e03024f71ac" Dec 11 09:17:48 crc kubenswrapper[4860]: I1211 09:17:48.860913 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"906d59b91fd7d000faff22bd48e1b733b7a36c777a8bc783b32b1e03024f71ac"} err="failed to get container status \"906d59b91fd7d000faff22bd48e1b733b7a36c777a8bc783b32b1e03024f71ac\": rpc error: code = NotFound desc = could not find container \"906d59b91fd7d000faff22bd48e1b733b7a36c777a8bc783b32b1e03024f71ac\": container with ID starting with 906d59b91fd7d000faff22bd48e1b733b7a36c777a8bc783b32b1e03024f71ac not found: ID does not exist" Dec 11 09:17:49 crc kubenswrapper[4860]: I1211 09:17:49.599500 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c880ffe1-896a-41c5-8cea-4a6ca96ca971" path="/var/lib/kubelet/pods/c880ffe1-896a-41c5-8cea-4a6ca96ca971/volumes" Dec 11 09:17:53 crc kubenswrapper[4860]: I1211 09:17:53.980391 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_941ec9b01b7a72127ba64306553d6766ab71ba59f32f9ca01c76932a62n7v8f_bf285af3-cbc9-4e8f-bb3a-bc1d16a3ea54/util/0.log" Dec 11 09:17:54 crc kubenswrapper[4860]: I1211 09:17:54.169536 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_941ec9b01b7a72127ba64306553d6766ab71ba59f32f9ca01c76932a62n7v8f_bf285af3-cbc9-4e8f-bb3a-bc1d16a3ea54/util/0.log" Dec 11 09:17:54 crc kubenswrapper[4860]: I1211 09:17:54.197836 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_941ec9b01b7a72127ba64306553d6766ab71ba59f32f9ca01c76932a62n7v8f_bf285af3-cbc9-4e8f-bb3a-bc1d16a3ea54/pull/0.log" Dec 11 09:17:54 crc kubenswrapper[4860]: I1211 09:17:54.213442 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_941ec9b01b7a72127ba64306553d6766ab71ba59f32f9ca01c76932a62n7v8f_bf285af3-cbc9-4e8f-bb3a-bc1d16a3ea54/pull/0.log" Dec 11 09:17:54 crc kubenswrapper[4860]: I1211 09:17:54.375149 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_941ec9b01b7a72127ba64306553d6766ab71ba59f32f9ca01c76932a62n7v8f_bf285af3-cbc9-4e8f-bb3a-bc1d16a3ea54/util/0.log" Dec 11 09:17:54 crc kubenswrapper[4860]: I1211 09:17:54.379150 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_941ec9b01b7a72127ba64306553d6766ab71ba59f32f9ca01c76932a62n7v8f_bf285af3-cbc9-4e8f-bb3a-bc1d16a3ea54/pull/0.log" Dec 11 09:17:54 crc kubenswrapper[4860]: I1211 09:17:54.394029 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_941ec9b01b7a72127ba64306553d6766ab71ba59f32f9ca01c76932a62n7v8f_bf285af3-cbc9-4e8f-bb3a-bc1d16a3ea54/extract/0.log" Dec 11 09:17:54 crc kubenswrapper[4860]: I1211 09:17:54.579571 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-bshq2_64f09777-fca3-412f-98f8-5cd189cb9fbd/kube-rbac-proxy/0.log" Dec 11 09:17:54 crc kubenswrapper[4860]: I1211 09:17:54.689558 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-bshq2_64f09777-fca3-412f-98f8-5cd189cb9fbd/manager/0.log" Dec 11 09:17:54 crc kubenswrapper[4860]: I1211 09:17:54.702177 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6c677c69b-zrrf6_323ad77f-a0d6-461d-b483-2ffd74ebb9b7/kube-rbac-proxy/0.log" Dec 11 09:17:54 crc kubenswrapper[4860]: I1211 09:17:54.961807 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6c677c69b-zrrf6_323ad77f-a0d6-461d-b483-2ffd74ebb9b7/manager/0.log" Dec 11 09:17:55 crc kubenswrapper[4860]: I1211 09:17:55.059313 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-697fb699cf-6t74n_a3bf483c-6438-4b7e-abb8-447b9b146644/kube-rbac-proxy/0.log" Dec 11 09:17:55 crc kubenswrapper[4860]: I1211 09:17:55.072832 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-697fb699cf-6t74n_a3bf483c-6438-4b7e-abb8-447b9b146644/manager/0.log" Dec 11 09:17:55 crc kubenswrapper[4860]: I1211 09:17:55.223187 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-5697bb5779-kgzwt_e00c360d-20a4-4e2e-81ea-e4ffa1bc62fc/kube-rbac-proxy/0.log" Dec 11 09:17:55 crc kubenswrapper[4860]: I1211 09:17:55.308533 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-5697bb5779-kgzwt_e00c360d-20a4-4e2e-81ea-e4ffa1bc62fc/manager/0.log" Dec 11 09:17:55 crc kubenswrapper[4860]: I1211 09:17:55.493507 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-lrhhk_505d980f-cfcb-42b2-876e-a4730abe7ea6/manager/0.log" Dec 11 09:17:55 crc kubenswrapper[4860]: I1211 09:17:55.493968 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-lrhhk_505d980f-cfcb-42b2-876e-a4730abe7ea6/kube-rbac-proxy/0.log" Dec 11 09:17:55 crc kubenswrapper[4860]: I1211 09:17:55.540034 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-65fv4_bb231e77-e19b-41ed-bf7a-ae354e1089ec/kube-rbac-proxy/0.log" Dec 11 09:17:55 crc kubenswrapper[4860]: I1211 09:17:55.727344 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-65fv4_bb231e77-e19b-41ed-bf7a-ae354e1089ec/manager/0.log" Dec 11 09:17:55 crc kubenswrapper[4860]: I1211 09:17:55.764449 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-78d48bff9d-sq5gd_7f1d9cd2-1e0c-4c4d-86a6-b91a38eadba2/kube-rbac-proxy/0.log" Dec 11 09:17:55 crc kubenswrapper[4860]: I1211 09:17:55.996142 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-78d48bff9d-sq5gd_7f1d9cd2-1e0c-4c4d-86a6-b91a38eadba2/manager/0.log" Dec 11 09:17:56 crc kubenswrapper[4860]: I1211 09:17:56.027614 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-967d97867-5wvld_91bb3ff9-fd19-4c79-bac4-621cb114c783/manager/0.log" Dec 11 09:17:56 crc kubenswrapper[4860]: I1211 09:17:56.033024 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-967d97867-5wvld_91bb3ff9-fd19-4c79-bac4-621cb114c783/kube-rbac-proxy/0.log" Dec 11 09:17:56 crc kubenswrapper[4860]: I1211 09:17:56.223797 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-6bpb9_d5e0079f-f500-407f-b06f-a610a19c741b/kube-rbac-proxy/0.log" Dec 11 09:17:56 crc kubenswrapper[4860]: I1211 09:17:56.316303 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-6bpb9_d5e0079f-f500-407f-b06f-a610a19c741b/manager/0.log" Dec 11 09:17:56 crc kubenswrapper[4860]: I1211 09:17:56.436682 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5b5fd79c9c-lpm8h_aae650b9-dcbc-482b-b49f-0a45651bb4b4/kube-rbac-proxy/0.log" Dec 11 09:17:56 crc kubenswrapper[4860]: I1211 09:17:56.462509 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5b5fd79c9c-lpm8h_aae650b9-dcbc-482b-b49f-0a45651bb4b4/manager/0.log" Dec 11 09:17:56 crc kubenswrapper[4860]: I1211 09:17:56.549063 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-79c8c4686c-4tw58_5a589b13-7f29-40ae-9804-7c5c2bc3d745/kube-rbac-proxy/0.log" Dec 11 09:17:56 crc kubenswrapper[4860]: I1211 09:17:56.643815 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-79c8c4686c-4tw58_5a589b13-7f29-40ae-9804-7c5c2bc3d745/manager/0.log" Dec 11 09:17:56 crc kubenswrapper[4860]: I1211 09:17:56.707181 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-ht7db_dc79abc2-7b6c-46c4-b9b4-a9c76696a8b2/kube-rbac-proxy/0.log" Dec 11 09:17:56 crc kubenswrapper[4860]: I1211 09:17:56.816979 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-ht7db_dc79abc2-7b6c-46c4-b9b4-a9c76696a8b2/manager/0.log" Dec 11 09:17:56 crc kubenswrapper[4860]: I1211 09:17:56.927160 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-zrhjm_5aff136d-8f48-4cb7-8ddc-72ce966607b7/kube-rbac-proxy/0.log" Dec 11 09:17:56 crc kubenswrapper[4860]: I1211 09:17:56.992319 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-zrhjm_5aff136d-8f48-4cb7-8ddc-72ce966607b7/manager/0.log" Dec 11 09:17:57 crc kubenswrapper[4860]: I1211 09:17:57.141860 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-bvs87_689923af-60de-49e9-bbed-bebaa63adc4b/kube-rbac-proxy/0.log" Dec 11 09:17:57 crc kubenswrapper[4860]: I1211 09:17:57.211666 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-bvs87_689923af-60de-49e9-bbed-bebaa63adc4b/manager/0.log" Dec 11 09:17:57 crc kubenswrapper[4860]: I1211 09:17:57.265047 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-84b575879fgp9c8_04044c38-9aa4-4439-bad4-cf687e00ed5a/kube-rbac-proxy/0.log" Dec 11 09:17:57 crc kubenswrapper[4860]: I1211 09:17:57.340426 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-84b575879fgp9c8_04044c38-9aa4-4439-bad4-cf687e00ed5a/manager/0.log" Dec 11 09:17:57 crc kubenswrapper[4860]: I1211 09:17:57.775130 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-7d57466ffb-mg466_f57d6313-d688-4a8d-a9cb-face8d0e0ef6/operator/0.log" Dec 11 09:17:57 crc kubenswrapper[4860]: I1211 09:17:57.873913 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-hm8hr_aec175c6-0b2d-40d8-855a-6132fb5e66a0/registry-server/0.log" Dec 11 09:17:58 crc kubenswrapper[4860]: I1211 09:17:58.032127 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-nvt56_99f7e97f-7cea-4785-b95e-133adf238ac3/kube-rbac-proxy/0.log" Dec 11 09:17:58 crc kubenswrapper[4860]: I1211 09:17:58.140323 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-nvt56_99f7e97f-7cea-4785-b95e-133adf238ac3/manager/0.log" Dec 11 09:17:58 crc kubenswrapper[4860]: I1211 09:17:58.277243 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-2qmxt_a52f8df6-7c28-4776-aca5-bd8b47e82fe8/kube-rbac-proxy/0.log" Dec 11 09:17:58 crc kubenswrapper[4860]: I1211 09:17:58.296020 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-2qmxt_a52f8df6-7c28-4776-aca5-bd8b47e82fe8/manager/0.log" Dec 11 09:17:58 crc kubenswrapper[4860]: I1211 09:17:58.525783 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-jsftr_22f3c8f5-1781-4ba2-8610-c9ada5539d92/operator/0.log" Dec 11 09:17:58 crc kubenswrapper[4860]: I1211 09:17:58.631860 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-6d784644c6-68pll_4bbd2122-dfb6-486b-8196-375ffeb93527/manager/0.log" Dec 11 09:17:58 crc kubenswrapper[4860]: I1211 09:17:58.643404 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-9d58d64bc-p7dkp_211d9c9c-f584-43ca-8db1-7b81f6307c21/kube-rbac-proxy/0.log" Dec 11 09:17:58 crc kubenswrapper[4860]: I1211 09:17:58.650076 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-9d58d64bc-p7dkp_211d9c9c-f584-43ca-8db1-7b81f6307c21/manager/0.log" Dec 11 09:17:59 crc kubenswrapper[4860]: I1211 09:17:59.022593 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-58d5ff84df-hrpn8_e36550af-98a6-49b3-9769-bd52d0da2838/kube-rbac-proxy/0.log" Dec 11 09:17:59 crc kubenswrapper[4860]: I1211 09:17:59.088911 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-fh5vc_b6c38bbc-80d0-449a-aab0-291078361ebd/kube-rbac-proxy/0.log" Dec 11 09:17:59 crc kubenswrapper[4860]: I1211 09:17:59.141073 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-58d5ff84df-hrpn8_e36550af-98a6-49b3-9769-bd52d0da2838/manager/0.log" Dec 11 09:17:59 crc kubenswrapper[4860]: I1211 09:17:59.199838 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-fh5vc_b6c38bbc-80d0-449a-aab0-291078361ebd/manager/0.log" Dec 11 09:17:59 crc kubenswrapper[4860]: I1211 09:17:59.266922 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-75944c9b7-ck9nc_8462b48d-cc3d-4714-9558-22818db99c08/kube-rbac-proxy/0.log" Dec 11 09:17:59 crc kubenswrapper[4860]: I1211 09:17:59.344429 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-75944c9b7-ck9nc_8462b48d-cc3d-4714-9558-22818db99c08/manager/0.log" Dec 11 09:17:59 crc kubenswrapper[4860]: I1211 09:17:59.578778 4860 scope.go:117] "RemoveContainer" containerID="6704d8c98ab428aac5737fa114668f6cf3119541ee0fe55ea326d429fcbe2d82" Dec 11 09:17:59 crc kubenswrapper[4860]: E1211 09:17:59.579096 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-99qgp_openshift-machine-config-operator(31c30642-6e60-41b4-a477-0d802424e0aa)\"" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" Dec 11 09:18:13 crc kubenswrapper[4860]: I1211 09:18:13.168749 4860 scope.go:117] "RemoveContainer" containerID="6704d8c98ab428aac5737fa114668f6cf3119541ee0fe55ea326d429fcbe2d82" Dec 11 09:18:13 crc kubenswrapper[4860]: E1211 09:18:13.169401 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-99qgp_openshift-machine-config-operator(31c30642-6e60-41b4-a477-0d802424e0aa)\"" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" Dec 11 09:18:18 crc kubenswrapper[4860]: I1211 09:18:18.592592 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-76lnq_7a764867-5216-47fb-8211-cea78b79f707/control-plane-machine-set-operator/0.log" Dec 11 09:18:18 crc kubenswrapper[4860]: I1211 09:18:18.802493 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-nc65d_cb55e751-9913-4a3c-a0b4-3e681129d052/kube-rbac-proxy/0.log" Dec 11 09:18:18 crc kubenswrapper[4860]: I1211 09:18:18.855116 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-nc65d_cb55e751-9913-4a3c-a0b4-3e681129d052/machine-api-operator/0.log" Dec 11 09:18:27 crc kubenswrapper[4860]: I1211 09:18:27.586206 4860 scope.go:117] "RemoveContainer" containerID="6704d8c98ab428aac5737fa114668f6cf3119541ee0fe55ea326d429fcbe2d82" Dec 11 09:18:27 crc kubenswrapper[4860]: E1211 09:18:27.587125 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-99qgp_openshift-machine-config-operator(31c30642-6e60-41b4-a477-0d802424e0aa)\"" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" Dec 11 09:18:32 crc kubenswrapper[4860]: I1211 09:18:32.678141 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-7bszh_5198d79a-f8ab-4a3a-9859-0925545321a4/cert-manager-controller/0.log" Dec 11 09:18:32 crc kubenswrapper[4860]: I1211 09:18:32.860122 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-qhfrs_29ede248-cc17-41b1-a1c4-4feaa6010cf2/cert-manager-cainjector/0.log" Dec 11 09:18:33 crc kubenswrapper[4860]: I1211 09:18:33.500131 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-rvtht_7ea21c5c-320d-4bb0-9bea-186a528a61f4/cert-manager-webhook/0.log" Dec 11 09:18:40 crc kubenswrapper[4860]: I1211 09:18:40.578672 4860 scope.go:117] "RemoveContainer" containerID="6704d8c98ab428aac5737fa114668f6cf3119541ee0fe55ea326d429fcbe2d82" Dec 11 09:18:40 crc kubenswrapper[4860]: E1211 09:18:40.579425 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-99qgp_openshift-machine-config-operator(31c30642-6e60-41b4-a477-0d802424e0aa)\"" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" Dec 11 09:18:45 crc kubenswrapper[4860]: I1211 09:18:45.392500 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-6ff7998486-2q4xw_4cc66f42-2ab8-4a37-be52-5eb9be9fdd4b/nmstate-console-plugin/0.log" Dec 11 09:18:45 crc kubenswrapper[4860]: I1211 09:18:45.573519 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-kv4zf_f17c9e82-c0f3-408d-a871-d24992f28860/nmstate-handler/0.log" Dec 11 09:18:45 crc kubenswrapper[4860]: I1211 09:18:45.666765 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f7f7578db-qv9ws_e18c8adc-cdba-4f7f-8b5d-f15a73397c2b/kube-rbac-proxy/0.log" Dec 11 09:18:45 crc kubenswrapper[4860]: I1211 09:18:45.688714 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f7f7578db-qv9ws_e18c8adc-cdba-4f7f-8b5d-f15a73397c2b/nmstate-metrics/0.log" Dec 11 09:18:45 crc kubenswrapper[4860]: I1211 09:18:45.852850 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-6769fb99d-49jqb_4d7eadcc-525c-4503-a8c1-7c0b0c04a9e8/nmstate-operator/0.log" Dec 11 09:18:45 crc kubenswrapper[4860]: I1211 09:18:45.893850 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-f8fb84555-nnjsk_295f8d4b-cf0e-41e9-a9a6-8ff38c589ecc/nmstate-webhook/0.log" Dec 11 09:18:52 crc kubenswrapper[4860]: I1211 09:18:52.579399 4860 scope.go:117] "RemoveContainer" containerID="6704d8c98ab428aac5737fa114668f6cf3119541ee0fe55ea326d429fcbe2d82" Dec 11 09:18:52 crc kubenswrapper[4860]: E1211 09:18:52.580325 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-99qgp_openshift-machine-config-operator(31c30642-6e60-41b4-a477-0d802424e0aa)\"" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" Dec 11 09:18:59 crc kubenswrapper[4860]: I1211 09:18:59.580170 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-5bddd4b946-96p7v_31da3aab-3e08-4b87-a866-d49a4224cf29/kube-rbac-proxy/0.log" Dec 11 09:18:59 crc kubenswrapper[4860]: I1211 09:18:59.689570 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-5bddd4b946-96p7v_31da3aab-3e08-4b87-a866-d49a4224cf29/controller/0.log" Dec 11 09:18:59 crc kubenswrapper[4860]: I1211 09:18:59.775472 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c6crs_130f6554-c476-49a0-8270-593008c1ffc6/cp-frr-files/0.log" Dec 11 09:18:59 crc kubenswrapper[4860]: I1211 09:18:59.960767 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c6crs_130f6554-c476-49a0-8270-593008c1ffc6/cp-frr-files/0.log" Dec 11 09:18:59 crc kubenswrapper[4860]: I1211 09:18:59.976934 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c6crs_130f6554-c476-49a0-8270-593008c1ffc6/cp-reloader/0.log" Dec 11 09:19:00 crc kubenswrapper[4860]: I1211 09:19:00.023392 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c6crs_130f6554-c476-49a0-8270-593008c1ffc6/cp-reloader/0.log" Dec 11 09:19:00 crc kubenswrapper[4860]: I1211 09:19:00.035425 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c6crs_130f6554-c476-49a0-8270-593008c1ffc6/cp-metrics/0.log" Dec 11 09:19:00 crc kubenswrapper[4860]: I1211 09:19:00.215424 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c6crs_130f6554-c476-49a0-8270-593008c1ffc6/cp-metrics/0.log" Dec 11 09:19:00 crc kubenswrapper[4860]: I1211 09:19:00.217584 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c6crs_130f6554-c476-49a0-8270-593008c1ffc6/cp-metrics/0.log" Dec 11 09:19:00 crc kubenswrapper[4860]: I1211 09:19:00.250060 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c6crs_130f6554-c476-49a0-8270-593008c1ffc6/cp-reloader/0.log" Dec 11 09:19:00 crc kubenswrapper[4860]: I1211 09:19:00.269621 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c6crs_130f6554-c476-49a0-8270-593008c1ffc6/cp-frr-files/0.log" Dec 11 09:19:00 crc kubenswrapper[4860]: I1211 09:19:00.400593 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c6crs_130f6554-c476-49a0-8270-593008c1ffc6/cp-frr-files/0.log" Dec 11 09:19:00 crc kubenswrapper[4860]: I1211 09:19:00.430534 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c6crs_130f6554-c476-49a0-8270-593008c1ffc6/cp-reloader/0.log" Dec 11 09:19:00 crc kubenswrapper[4860]: I1211 09:19:00.437386 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c6crs_130f6554-c476-49a0-8270-593008c1ffc6/cp-metrics/0.log" Dec 11 09:19:00 crc kubenswrapper[4860]: I1211 09:19:00.446219 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c6crs_130f6554-c476-49a0-8270-593008c1ffc6/controller/0.log" Dec 11 09:19:00 crc kubenswrapper[4860]: I1211 09:19:00.622475 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c6crs_130f6554-c476-49a0-8270-593008c1ffc6/kube-rbac-proxy/0.log" Dec 11 09:19:00 crc kubenswrapper[4860]: I1211 09:19:00.645159 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c6crs_130f6554-c476-49a0-8270-593008c1ffc6/kube-rbac-proxy-frr/0.log" Dec 11 09:19:00 crc kubenswrapper[4860]: I1211 09:19:00.645478 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c6crs_130f6554-c476-49a0-8270-593008c1ffc6/frr-metrics/0.log" Dec 11 09:19:00 crc kubenswrapper[4860]: I1211 09:19:00.805946 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c6crs_130f6554-c476-49a0-8270-593008c1ffc6/reloader/0.log" Dec 11 09:19:00 crc kubenswrapper[4860]: I1211 09:19:00.887222 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7784b6fcf-2548h_98afb3e0-d363-472c-886e-2b2981cb13a4/frr-k8s-webhook-server/0.log" Dec 11 09:19:01 crc kubenswrapper[4860]: I1211 09:19:01.085072 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-d5785c6bc-n9n8q_e0e047cd-d665-4b39-8588-b0d396e3b36d/manager/0.log" Dec 11 09:19:01 crc kubenswrapper[4860]: I1211 09:19:01.282078 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-59475655bd-8sqgf_47efc842-7de4-461c-bafc-0fbaabb6c3ad/webhook-server/0.log" Dec 11 09:19:01 crc kubenswrapper[4860]: I1211 09:19:01.370422 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-qmw55_b5eafccf-57d6-4bcb-a43f-5ef3a2f04db0/kube-rbac-proxy/0.log" Dec 11 09:19:01 crc kubenswrapper[4860]: I1211 09:19:01.975496 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-qmw55_b5eafccf-57d6-4bcb-a43f-5ef3a2f04db0/speaker/0.log" Dec 11 09:19:02 crc kubenswrapper[4860]: I1211 09:19:02.020093 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-c6crs_130f6554-c476-49a0-8270-593008c1ffc6/frr/0.log" Dec 11 09:19:07 crc kubenswrapper[4860]: I1211 09:19:07.587954 4860 scope.go:117] "RemoveContainer" containerID="6704d8c98ab428aac5737fa114668f6cf3119541ee0fe55ea326d429fcbe2d82" Dec 11 09:19:07 crc kubenswrapper[4860]: E1211 09:19:07.588974 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-99qgp_openshift-machine-config-operator(31c30642-6e60-41b4-a477-0d802424e0aa)\"" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" Dec 11 09:19:10 crc kubenswrapper[4860]: I1211 09:19:10.810882 4860 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-4gz5s"] Dec 11 09:19:10 crc kubenswrapper[4860]: E1211 09:19:10.811724 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c880ffe1-896a-41c5-8cea-4a6ca96ca971" containerName="extract-content" Dec 11 09:19:10 crc kubenswrapper[4860]: I1211 09:19:10.811741 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="c880ffe1-896a-41c5-8cea-4a6ca96ca971" containerName="extract-content" Dec 11 09:19:10 crc kubenswrapper[4860]: E1211 09:19:10.811768 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8e0c201-97a3-494e-af29-c3530510b379" containerName="extract-content" Dec 11 09:19:10 crc kubenswrapper[4860]: I1211 09:19:10.811776 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8e0c201-97a3-494e-af29-c3530510b379" containerName="extract-content" Dec 11 09:19:10 crc kubenswrapper[4860]: E1211 09:19:10.811799 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8e0c201-97a3-494e-af29-c3530510b379" containerName="registry-server" Dec 11 09:19:10 crc kubenswrapper[4860]: I1211 09:19:10.811808 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8e0c201-97a3-494e-af29-c3530510b379" containerName="registry-server" Dec 11 09:19:10 crc kubenswrapper[4860]: E1211 09:19:10.811824 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8e0c201-97a3-494e-af29-c3530510b379" containerName="extract-utilities" Dec 11 09:19:10 crc kubenswrapper[4860]: I1211 09:19:10.811831 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8e0c201-97a3-494e-af29-c3530510b379" containerName="extract-utilities" Dec 11 09:19:10 crc kubenswrapper[4860]: E1211 09:19:10.811858 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c880ffe1-896a-41c5-8cea-4a6ca96ca971" containerName="extract-utilities" Dec 11 09:19:10 crc kubenswrapper[4860]: I1211 09:19:10.811866 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="c880ffe1-896a-41c5-8cea-4a6ca96ca971" containerName="extract-utilities" Dec 11 09:19:10 crc kubenswrapper[4860]: E1211 09:19:10.811891 4860 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c880ffe1-896a-41c5-8cea-4a6ca96ca971" containerName="registry-server" Dec 11 09:19:10 crc kubenswrapper[4860]: I1211 09:19:10.811899 4860 state_mem.go:107] "Deleted CPUSet assignment" podUID="c880ffe1-896a-41c5-8cea-4a6ca96ca971" containerName="registry-server" Dec 11 09:19:10 crc kubenswrapper[4860]: I1211 09:19:10.812189 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="c880ffe1-896a-41c5-8cea-4a6ca96ca971" containerName="registry-server" Dec 11 09:19:10 crc kubenswrapper[4860]: I1211 09:19:10.812213 4860 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8e0c201-97a3-494e-af29-c3530510b379" containerName="registry-server" Dec 11 09:19:10 crc kubenswrapper[4860]: I1211 09:19:10.814587 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4gz5s" Dec 11 09:19:10 crc kubenswrapper[4860]: I1211 09:19:10.825547 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4gz5s"] Dec 11 09:19:10 crc kubenswrapper[4860]: I1211 09:19:10.873180 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2dbsh\" (UniqueName: \"kubernetes.io/projected/ba902791-09ec-4471-98b8-e79cb9f40013-kube-api-access-2dbsh\") pod \"redhat-marketplace-4gz5s\" (UID: \"ba902791-09ec-4471-98b8-e79cb9f40013\") " pod="openshift-marketplace/redhat-marketplace-4gz5s" Dec 11 09:19:10 crc kubenswrapper[4860]: I1211 09:19:10.873301 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba902791-09ec-4471-98b8-e79cb9f40013-catalog-content\") pod \"redhat-marketplace-4gz5s\" (UID: \"ba902791-09ec-4471-98b8-e79cb9f40013\") " pod="openshift-marketplace/redhat-marketplace-4gz5s" Dec 11 09:19:10 crc kubenswrapper[4860]: I1211 09:19:10.873331 4860 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba902791-09ec-4471-98b8-e79cb9f40013-utilities\") pod \"redhat-marketplace-4gz5s\" (UID: \"ba902791-09ec-4471-98b8-e79cb9f40013\") " pod="openshift-marketplace/redhat-marketplace-4gz5s" Dec 11 09:19:10 crc kubenswrapper[4860]: I1211 09:19:10.975071 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2dbsh\" (UniqueName: \"kubernetes.io/projected/ba902791-09ec-4471-98b8-e79cb9f40013-kube-api-access-2dbsh\") pod \"redhat-marketplace-4gz5s\" (UID: \"ba902791-09ec-4471-98b8-e79cb9f40013\") " pod="openshift-marketplace/redhat-marketplace-4gz5s" Dec 11 09:19:10 crc kubenswrapper[4860]: I1211 09:19:10.975160 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba902791-09ec-4471-98b8-e79cb9f40013-catalog-content\") pod \"redhat-marketplace-4gz5s\" (UID: \"ba902791-09ec-4471-98b8-e79cb9f40013\") " pod="openshift-marketplace/redhat-marketplace-4gz5s" Dec 11 09:19:10 crc kubenswrapper[4860]: I1211 09:19:10.975182 4860 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba902791-09ec-4471-98b8-e79cb9f40013-utilities\") pod \"redhat-marketplace-4gz5s\" (UID: \"ba902791-09ec-4471-98b8-e79cb9f40013\") " pod="openshift-marketplace/redhat-marketplace-4gz5s" Dec 11 09:19:10 crc kubenswrapper[4860]: I1211 09:19:10.975670 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba902791-09ec-4471-98b8-e79cb9f40013-catalog-content\") pod \"redhat-marketplace-4gz5s\" (UID: \"ba902791-09ec-4471-98b8-e79cb9f40013\") " pod="openshift-marketplace/redhat-marketplace-4gz5s" Dec 11 09:19:10 crc kubenswrapper[4860]: I1211 09:19:10.975801 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba902791-09ec-4471-98b8-e79cb9f40013-utilities\") pod \"redhat-marketplace-4gz5s\" (UID: \"ba902791-09ec-4471-98b8-e79cb9f40013\") " pod="openshift-marketplace/redhat-marketplace-4gz5s" Dec 11 09:19:10 crc kubenswrapper[4860]: I1211 09:19:10.996251 4860 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2dbsh\" (UniqueName: \"kubernetes.io/projected/ba902791-09ec-4471-98b8-e79cb9f40013-kube-api-access-2dbsh\") pod \"redhat-marketplace-4gz5s\" (UID: \"ba902791-09ec-4471-98b8-e79cb9f40013\") " pod="openshift-marketplace/redhat-marketplace-4gz5s" Dec 11 09:19:11 crc kubenswrapper[4860]: I1211 09:19:11.132743 4860 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4gz5s" Dec 11 09:19:11 crc kubenswrapper[4860]: I1211 09:19:11.624125 4860 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4gz5s"] Dec 11 09:19:11 crc kubenswrapper[4860]: I1211 09:19:11.724767 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4gz5s" event={"ID":"ba902791-09ec-4471-98b8-e79cb9f40013","Type":"ContainerStarted","Data":"f6edc62a2fbdabcb58e5af9b9f1f636f2dd69a2945c6ac1ff84972bf91fbdab9"} Dec 11 09:19:12 crc kubenswrapper[4860]: I1211 09:19:12.734257 4860 generic.go:334] "Generic (PLEG): container finished" podID="ba902791-09ec-4471-98b8-e79cb9f40013" containerID="4218b1c0ad730d76e3d50dc38793e1ec1b6aac5c50842277332c0064258cbb61" exitCode=0 Dec 11 09:19:12 crc kubenswrapper[4860]: I1211 09:19:12.734818 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4gz5s" event={"ID":"ba902791-09ec-4471-98b8-e79cb9f40013","Type":"ContainerDied","Data":"4218b1c0ad730d76e3d50dc38793e1ec1b6aac5c50842277332c0064258cbb61"} Dec 11 09:19:14 crc kubenswrapper[4860]: I1211 09:19:14.876897 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4vr6tn_da496881-c195-4cc2-9277-c8118bf651dc/util/0.log" Dec 11 09:19:15 crc kubenswrapper[4860]: I1211 09:19:15.025898 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4vr6tn_da496881-c195-4cc2-9277-c8118bf651dc/util/0.log" Dec 11 09:19:15 crc kubenswrapper[4860]: I1211 09:19:15.061626 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4vr6tn_da496881-c195-4cc2-9277-c8118bf651dc/pull/0.log" Dec 11 09:19:15 crc kubenswrapper[4860]: I1211 09:19:15.095558 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4vr6tn_da496881-c195-4cc2-9277-c8118bf651dc/pull/0.log" Dec 11 09:19:15 crc kubenswrapper[4860]: I1211 09:19:15.256703 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4vr6tn_da496881-c195-4cc2-9277-c8118bf651dc/pull/0.log" Dec 11 09:19:15 crc kubenswrapper[4860]: I1211 09:19:15.308172 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4vr6tn_da496881-c195-4cc2-9277-c8118bf651dc/extract/0.log" Dec 11 09:19:15 crc kubenswrapper[4860]: I1211 09:19:15.310885 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4vr6tn_da496881-c195-4cc2-9277-c8118bf651dc/util/0.log" Dec 11 09:19:15 crc kubenswrapper[4860]: I1211 09:19:15.422345 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8nkkjk_f80508d8-cf59-4656-af41-b0688b7b997c/util/0.log" Dec 11 09:19:15 crc kubenswrapper[4860]: I1211 09:19:15.599393 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8nkkjk_f80508d8-cf59-4656-af41-b0688b7b997c/util/0.log" Dec 11 09:19:15 crc kubenswrapper[4860]: I1211 09:19:15.606063 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8nkkjk_f80508d8-cf59-4656-af41-b0688b7b997c/pull/0.log" Dec 11 09:19:15 crc kubenswrapper[4860]: I1211 09:19:15.638295 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8nkkjk_f80508d8-cf59-4656-af41-b0688b7b997c/pull/0.log" Dec 11 09:19:15 crc kubenswrapper[4860]: I1211 09:19:15.761273 4860 generic.go:334] "Generic (PLEG): container finished" podID="ba902791-09ec-4471-98b8-e79cb9f40013" containerID="0d916c0c2d6fae93a30d00536cd82ff80165a66e7c7261dd36f1874145bb1f47" exitCode=0 Dec 11 09:19:15 crc kubenswrapper[4860]: I1211 09:19:15.761380 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4gz5s" event={"ID":"ba902791-09ec-4471-98b8-e79cb9f40013","Type":"ContainerDied","Data":"0d916c0c2d6fae93a30d00536cd82ff80165a66e7c7261dd36f1874145bb1f47"} Dec 11 09:19:15 crc kubenswrapper[4860]: I1211 09:19:15.839181 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8nkkjk_f80508d8-cf59-4656-af41-b0688b7b997c/util/0.log" Dec 11 09:19:15 crc kubenswrapper[4860]: I1211 09:19:15.877048 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8nkkjk_f80508d8-cf59-4656-af41-b0688b7b997c/pull/0.log" Dec 11 09:19:15 crc kubenswrapper[4860]: I1211 09:19:15.893791 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8nkkjk_f80508d8-cf59-4656-af41-b0688b7b997c/extract/0.log" Dec 11 09:19:16 crc kubenswrapper[4860]: I1211 09:19:16.027265 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-bg2t9_79181f73-fb9e-40c5-96b8-0ad95545f1f7/extract-utilities/0.log" Dec 11 09:19:16 crc kubenswrapper[4860]: I1211 09:19:16.721112 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-bg2t9_79181f73-fb9e-40c5-96b8-0ad95545f1f7/extract-utilities/0.log" Dec 11 09:19:16 crc kubenswrapper[4860]: I1211 09:19:16.943338 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-bg2t9_79181f73-fb9e-40c5-96b8-0ad95545f1f7/extract-content/0.log" Dec 11 09:19:16 crc kubenswrapper[4860]: I1211 09:19:16.981736 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-bg2t9_79181f73-fb9e-40c5-96b8-0ad95545f1f7/extract-content/0.log" Dec 11 09:19:17 crc kubenswrapper[4860]: I1211 09:19:17.162901 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-bg2t9_79181f73-fb9e-40c5-96b8-0ad95545f1f7/extract-utilities/0.log" Dec 11 09:19:17 crc kubenswrapper[4860]: I1211 09:19:17.204321 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-bg2t9_79181f73-fb9e-40c5-96b8-0ad95545f1f7/extract-content/0.log" Dec 11 09:19:17 crc kubenswrapper[4860]: I1211 09:19:17.406176 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-kzkbs_f1bc94aa-96ad-4e1f-a407-9968d42069aa/extract-utilities/0.log" Dec 11 09:19:17 crc kubenswrapper[4860]: I1211 09:19:17.767872 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-bg2t9_79181f73-fb9e-40c5-96b8-0ad95545f1f7/registry-server/0.log" Dec 11 09:19:17 crc kubenswrapper[4860]: I1211 09:19:17.791735 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4gz5s" event={"ID":"ba902791-09ec-4471-98b8-e79cb9f40013","Type":"ContainerStarted","Data":"de21d0ef0bfe8e75da5d6898af8402709c4651605fcb667a43e0b9f3fee7541d"} Dec 11 09:19:17 crc kubenswrapper[4860]: I1211 09:19:17.804476 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-kzkbs_f1bc94aa-96ad-4e1f-a407-9968d42069aa/extract-content/0.log" Dec 11 09:19:17 crc kubenswrapper[4860]: I1211 09:19:17.819303 4860 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-4gz5s" podStartSLOduration=3.639191685 podStartE2EDuration="7.819280475s" podCreationTimestamp="2025-12-11 09:19:10 +0000 UTC" firstStartedPulling="2025-12-11 09:19:12.73631581 +0000 UTC m=+4085.464834865" lastFinishedPulling="2025-12-11 09:19:16.9164046 +0000 UTC m=+4089.644923655" observedRunningTime="2025-12-11 09:19:17.818770009 +0000 UTC m=+4090.547289064" watchObservedRunningTime="2025-12-11 09:19:17.819280475 +0000 UTC m=+4090.547799530" Dec 11 09:19:17 crc kubenswrapper[4860]: I1211 09:19:17.854332 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-kzkbs_f1bc94aa-96ad-4e1f-a407-9968d42069aa/extract-content/0.log" Dec 11 09:19:17 crc kubenswrapper[4860]: I1211 09:19:17.864976 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-kzkbs_f1bc94aa-96ad-4e1f-a407-9968d42069aa/extract-utilities/0.log" Dec 11 09:19:18 crc kubenswrapper[4860]: I1211 09:19:18.061069 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-kzkbs_f1bc94aa-96ad-4e1f-a407-9968d42069aa/extract-content/0.log" Dec 11 09:19:18 crc kubenswrapper[4860]: I1211 09:19:18.061243 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-kzkbs_f1bc94aa-96ad-4e1f-a407-9968d42069aa/extract-utilities/0.log" Dec 11 09:19:18 crc kubenswrapper[4860]: I1211 09:19:18.186679 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-trp75_dd2052c5-96c6-4226-858e-61e3ddbfeef9/marketplace-operator/0.log" Dec 11 09:19:18 crc kubenswrapper[4860]: I1211 09:19:18.301265 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-4gz5s_ba902791-09ec-4471-98b8-e79cb9f40013/extract-utilities/0.log" Dec 11 09:19:18 crc kubenswrapper[4860]: I1211 09:19:18.521150 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-4gz5s_ba902791-09ec-4471-98b8-e79cb9f40013/extract-content/0.log" Dec 11 09:19:18 crc kubenswrapper[4860]: I1211 09:19:18.579682 4860 scope.go:117] "RemoveContainer" containerID="6704d8c98ab428aac5737fa114668f6cf3119541ee0fe55ea326d429fcbe2d82" Dec 11 09:19:18 crc kubenswrapper[4860]: E1211 09:19:18.580024 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-99qgp_openshift-machine-config-operator(31c30642-6e60-41b4-a477-0d802424e0aa)\"" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" Dec 11 09:19:18 crc kubenswrapper[4860]: I1211 09:19:18.642233 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-4gz5s_ba902791-09ec-4471-98b8-e79cb9f40013/extract-utilities/0.log" Dec 11 09:19:18 crc kubenswrapper[4860]: I1211 09:19:18.656376 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-4gz5s_ba902791-09ec-4471-98b8-e79cb9f40013/extract-content/0.log" Dec 11 09:19:18 crc kubenswrapper[4860]: I1211 09:19:18.759483 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-kzkbs_f1bc94aa-96ad-4e1f-a407-9968d42069aa/registry-server/0.log" Dec 11 09:19:18 crc kubenswrapper[4860]: I1211 09:19:18.815110 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-4gz5s_ba902791-09ec-4471-98b8-e79cb9f40013/extract-utilities/0.log" Dec 11 09:19:18 crc kubenswrapper[4860]: I1211 09:19:18.822013 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-4gz5s_ba902791-09ec-4471-98b8-e79cb9f40013/extract-content/0.log" Dec 11 09:19:18 crc kubenswrapper[4860]: I1211 09:19:18.871548 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-4gz5s_ba902791-09ec-4471-98b8-e79cb9f40013/registry-server/0.log" Dec 11 09:19:18 crc kubenswrapper[4860]: I1211 09:19:18.950183 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-nkvw8_74e80c39-0b0c-42f6-88b6-a800ec910606/extract-utilities/0.log" Dec 11 09:19:19 crc kubenswrapper[4860]: I1211 09:19:19.145001 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-nkvw8_74e80c39-0b0c-42f6-88b6-a800ec910606/extract-utilities/0.log" Dec 11 09:19:19 crc kubenswrapper[4860]: I1211 09:19:19.196831 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-nkvw8_74e80c39-0b0c-42f6-88b6-a800ec910606/extract-content/0.log" Dec 11 09:19:19 crc kubenswrapper[4860]: I1211 09:19:19.223716 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-nkvw8_74e80c39-0b0c-42f6-88b6-a800ec910606/extract-content/0.log" Dec 11 09:19:19 crc kubenswrapper[4860]: I1211 09:19:19.351241 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-nkvw8_74e80c39-0b0c-42f6-88b6-a800ec910606/extract-content/0.log" Dec 11 09:19:19 crc kubenswrapper[4860]: I1211 09:19:19.351881 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-nkvw8_74e80c39-0b0c-42f6-88b6-a800ec910606/extract-utilities/0.log" Dec 11 09:19:19 crc kubenswrapper[4860]: I1211 09:19:19.496711 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-7vj8m_7d8a48c0-183c-4699-86d1-f6d0ef767cdf/extract-utilities/0.log" Dec 11 09:19:19 crc kubenswrapper[4860]: I1211 09:19:19.503320 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-nkvw8_74e80c39-0b0c-42f6-88b6-a800ec910606/registry-server/0.log" Dec 11 09:19:19 crc kubenswrapper[4860]: I1211 09:19:19.636031 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-7vj8m_7d8a48c0-183c-4699-86d1-f6d0ef767cdf/extract-utilities/0.log" Dec 11 09:19:19 crc kubenswrapper[4860]: I1211 09:19:19.648392 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-7vj8m_7d8a48c0-183c-4699-86d1-f6d0ef767cdf/extract-content/0.log" Dec 11 09:19:19 crc kubenswrapper[4860]: I1211 09:19:19.688461 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-7vj8m_7d8a48c0-183c-4699-86d1-f6d0ef767cdf/extract-content/0.log" Dec 11 09:19:19 crc kubenswrapper[4860]: I1211 09:19:19.850496 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-7vj8m_7d8a48c0-183c-4699-86d1-f6d0ef767cdf/extract-content/0.log" Dec 11 09:19:19 crc kubenswrapper[4860]: I1211 09:19:19.855991 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-7vj8m_7d8a48c0-183c-4699-86d1-f6d0ef767cdf/extract-utilities/0.log" Dec 11 09:19:20 crc kubenswrapper[4860]: I1211 09:19:20.519685 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-7vj8m_7d8a48c0-183c-4699-86d1-f6d0ef767cdf/registry-server/0.log" Dec 11 09:19:21 crc kubenswrapper[4860]: I1211 09:19:21.133726 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-4gz5s" Dec 11 09:19:21 crc kubenswrapper[4860]: I1211 09:19:21.135097 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-4gz5s" Dec 11 09:19:21 crc kubenswrapper[4860]: I1211 09:19:21.183911 4860 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-4gz5s" Dec 11 09:19:22 crc kubenswrapper[4860]: I1211 09:19:22.884114 4860 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-4gz5s" Dec 11 09:19:22 crc kubenswrapper[4860]: I1211 09:19:22.936411 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4gz5s"] Dec 11 09:19:24 crc kubenswrapper[4860]: I1211 09:19:24.859621 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-4gz5s" podUID="ba902791-09ec-4471-98b8-e79cb9f40013" containerName="registry-server" containerID="cri-o://de21d0ef0bfe8e75da5d6898af8402709c4651605fcb667a43e0b9f3fee7541d" gracePeriod=2 Dec 11 09:19:25 crc kubenswrapper[4860]: I1211 09:19:25.329330 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4gz5s" Dec 11 09:19:25 crc kubenswrapper[4860]: I1211 09:19:25.445337 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba902791-09ec-4471-98b8-e79cb9f40013-utilities\") pod \"ba902791-09ec-4471-98b8-e79cb9f40013\" (UID: \"ba902791-09ec-4471-98b8-e79cb9f40013\") " Dec 11 09:19:25 crc kubenswrapper[4860]: I1211 09:19:25.445522 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2dbsh\" (UniqueName: \"kubernetes.io/projected/ba902791-09ec-4471-98b8-e79cb9f40013-kube-api-access-2dbsh\") pod \"ba902791-09ec-4471-98b8-e79cb9f40013\" (UID: \"ba902791-09ec-4471-98b8-e79cb9f40013\") " Dec 11 09:19:25 crc kubenswrapper[4860]: I1211 09:19:25.445572 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba902791-09ec-4471-98b8-e79cb9f40013-catalog-content\") pod \"ba902791-09ec-4471-98b8-e79cb9f40013\" (UID: \"ba902791-09ec-4471-98b8-e79cb9f40013\") " Dec 11 09:19:25 crc kubenswrapper[4860]: I1211 09:19:25.446311 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ba902791-09ec-4471-98b8-e79cb9f40013-utilities" (OuterVolumeSpecName: "utilities") pod "ba902791-09ec-4471-98b8-e79cb9f40013" (UID: "ba902791-09ec-4471-98b8-e79cb9f40013"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 09:19:25 crc kubenswrapper[4860]: I1211 09:19:25.451401 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ba902791-09ec-4471-98b8-e79cb9f40013-kube-api-access-2dbsh" (OuterVolumeSpecName: "kube-api-access-2dbsh") pod "ba902791-09ec-4471-98b8-e79cb9f40013" (UID: "ba902791-09ec-4471-98b8-e79cb9f40013"). InnerVolumeSpecName "kube-api-access-2dbsh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 09:19:25 crc kubenswrapper[4860]: I1211 09:19:25.473988 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ba902791-09ec-4471-98b8-e79cb9f40013-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ba902791-09ec-4471-98b8-e79cb9f40013" (UID: "ba902791-09ec-4471-98b8-e79cb9f40013"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 09:19:25 crc kubenswrapper[4860]: I1211 09:19:25.547457 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2dbsh\" (UniqueName: \"kubernetes.io/projected/ba902791-09ec-4471-98b8-e79cb9f40013-kube-api-access-2dbsh\") on node \"crc\" DevicePath \"\"" Dec 11 09:19:25 crc kubenswrapper[4860]: I1211 09:19:25.547492 4860 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba902791-09ec-4471-98b8-e79cb9f40013-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 09:19:25 crc kubenswrapper[4860]: I1211 09:19:25.547504 4860 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba902791-09ec-4471-98b8-e79cb9f40013-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 09:19:25 crc kubenswrapper[4860]: I1211 09:19:25.869326 4860 generic.go:334] "Generic (PLEG): container finished" podID="ba902791-09ec-4471-98b8-e79cb9f40013" containerID="de21d0ef0bfe8e75da5d6898af8402709c4651605fcb667a43e0b9f3fee7541d" exitCode=0 Dec 11 09:19:25 crc kubenswrapper[4860]: I1211 09:19:25.869391 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4gz5s" event={"ID":"ba902791-09ec-4471-98b8-e79cb9f40013","Type":"ContainerDied","Data":"de21d0ef0bfe8e75da5d6898af8402709c4651605fcb667a43e0b9f3fee7541d"} Dec 11 09:19:25 crc kubenswrapper[4860]: I1211 09:19:25.869432 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4gz5s" event={"ID":"ba902791-09ec-4471-98b8-e79cb9f40013","Type":"ContainerDied","Data":"f6edc62a2fbdabcb58e5af9b9f1f636f2dd69a2945c6ac1ff84972bf91fbdab9"} Dec 11 09:19:25 crc kubenswrapper[4860]: I1211 09:19:25.869444 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4gz5s" Dec 11 09:19:25 crc kubenswrapper[4860]: I1211 09:19:25.869460 4860 scope.go:117] "RemoveContainer" containerID="de21d0ef0bfe8e75da5d6898af8402709c4651605fcb667a43e0b9f3fee7541d" Dec 11 09:19:25 crc kubenswrapper[4860]: I1211 09:19:25.891309 4860 scope.go:117] "RemoveContainer" containerID="0d916c0c2d6fae93a30d00536cd82ff80165a66e7c7261dd36f1874145bb1f47" Dec 11 09:19:25 crc kubenswrapper[4860]: I1211 09:19:25.901279 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4gz5s"] Dec 11 09:19:25 crc kubenswrapper[4860]: I1211 09:19:25.910847 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-4gz5s"] Dec 11 09:19:26 crc kubenswrapper[4860]: I1211 09:19:26.293827 4860 scope.go:117] "RemoveContainer" containerID="4218b1c0ad730d76e3d50dc38793e1ec1b6aac5c50842277332c0064258cbb61" Dec 11 09:19:26 crc kubenswrapper[4860]: I1211 09:19:26.491093 4860 scope.go:117] "RemoveContainer" containerID="de21d0ef0bfe8e75da5d6898af8402709c4651605fcb667a43e0b9f3fee7541d" Dec 11 09:19:26 crc kubenswrapper[4860]: E1211 09:19:26.491545 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de21d0ef0bfe8e75da5d6898af8402709c4651605fcb667a43e0b9f3fee7541d\": container with ID starting with de21d0ef0bfe8e75da5d6898af8402709c4651605fcb667a43e0b9f3fee7541d not found: ID does not exist" containerID="de21d0ef0bfe8e75da5d6898af8402709c4651605fcb667a43e0b9f3fee7541d" Dec 11 09:19:26 crc kubenswrapper[4860]: I1211 09:19:26.491599 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de21d0ef0bfe8e75da5d6898af8402709c4651605fcb667a43e0b9f3fee7541d"} err="failed to get container status \"de21d0ef0bfe8e75da5d6898af8402709c4651605fcb667a43e0b9f3fee7541d\": rpc error: code = NotFound desc = could not find container \"de21d0ef0bfe8e75da5d6898af8402709c4651605fcb667a43e0b9f3fee7541d\": container with ID starting with de21d0ef0bfe8e75da5d6898af8402709c4651605fcb667a43e0b9f3fee7541d not found: ID does not exist" Dec 11 09:19:26 crc kubenswrapper[4860]: I1211 09:19:26.491630 4860 scope.go:117] "RemoveContainer" containerID="0d916c0c2d6fae93a30d00536cd82ff80165a66e7c7261dd36f1874145bb1f47" Dec 11 09:19:26 crc kubenswrapper[4860]: E1211 09:19:26.491959 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0d916c0c2d6fae93a30d00536cd82ff80165a66e7c7261dd36f1874145bb1f47\": container with ID starting with 0d916c0c2d6fae93a30d00536cd82ff80165a66e7c7261dd36f1874145bb1f47 not found: ID does not exist" containerID="0d916c0c2d6fae93a30d00536cd82ff80165a66e7c7261dd36f1874145bb1f47" Dec 11 09:19:26 crc kubenswrapper[4860]: I1211 09:19:26.492001 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d916c0c2d6fae93a30d00536cd82ff80165a66e7c7261dd36f1874145bb1f47"} err="failed to get container status \"0d916c0c2d6fae93a30d00536cd82ff80165a66e7c7261dd36f1874145bb1f47\": rpc error: code = NotFound desc = could not find container \"0d916c0c2d6fae93a30d00536cd82ff80165a66e7c7261dd36f1874145bb1f47\": container with ID starting with 0d916c0c2d6fae93a30d00536cd82ff80165a66e7c7261dd36f1874145bb1f47 not found: ID does not exist" Dec 11 09:19:26 crc kubenswrapper[4860]: I1211 09:19:26.492029 4860 scope.go:117] "RemoveContainer" containerID="4218b1c0ad730d76e3d50dc38793e1ec1b6aac5c50842277332c0064258cbb61" Dec 11 09:19:26 crc kubenswrapper[4860]: E1211 09:19:26.492509 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4218b1c0ad730d76e3d50dc38793e1ec1b6aac5c50842277332c0064258cbb61\": container with ID starting with 4218b1c0ad730d76e3d50dc38793e1ec1b6aac5c50842277332c0064258cbb61 not found: ID does not exist" containerID="4218b1c0ad730d76e3d50dc38793e1ec1b6aac5c50842277332c0064258cbb61" Dec 11 09:19:26 crc kubenswrapper[4860]: I1211 09:19:26.492541 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4218b1c0ad730d76e3d50dc38793e1ec1b6aac5c50842277332c0064258cbb61"} err="failed to get container status \"4218b1c0ad730d76e3d50dc38793e1ec1b6aac5c50842277332c0064258cbb61\": rpc error: code = NotFound desc = could not find container \"4218b1c0ad730d76e3d50dc38793e1ec1b6aac5c50842277332c0064258cbb61\": container with ID starting with 4218b1c0ad730d76e3d50dc38793e1ec1b6aac5c50842277332c0064258cbb61 not found: ID does not exist" Dec 11 09:19:27 crc kubenswrapper[4860]: I1211 09:19:27.592492 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ba902791-09ec-4471-98b8-e79cb9f40013" path="/var/lib/kubelet/pods/ba902791-09ec-4471-98b8-e79cb9f40013/volumes" Dec 11 09:19:29 crc kubenswrapper[4860]: I1211 09:19:29.579505 4860 scope.go:117] "RemoveContainer" containerID="6704d8c98ab428aac5737fa114668f6cf3119541ee0fe55ea326d429fcbe2d82" Dec 11 09:19:29 crc kubenswrapper[4860]: E1211 09:19:29.580055 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-99qgp_openshift-machine-config-operator(31c30642-6e60-41b4-a477-0d802424e0aa)\"" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" Dec 11 09:19:36 crc kubenswrapper[4860]: E1211 09:19:36.645137 4860 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.169:60946->38.102.83.169:39441: write tcp 38.102.83.169:60946->38.102.83.169:39441: write: broken pipe Dec 11 09:19:43 crc kubenswrapper[4860]: I1211 09:19:43.582352 4860 scope.go:117] "RemoveContainer" containerID="6704d8c98ab428aac5737fa114668f6cf3119541ee0fe55ea326d429fcbe2d82" Dec 11 09:19:43 crc kubenswrapper[4860]: E1211 09:19:43.589478 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-99qgp_openshift-machine-config-operator(31c30642-6e60-41b4-a477-0d802424e0aa)\"" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" Dec 11 09:19:55 crc kubenswrapper[4860]: I1211 09:19:55.579391 4860 scope.go:117] "RemoveContainer" containerID="6704d8c98ab428aac5737fa114668f6cf3119541ee0fe55ea326d429fcbe2d82" Dec 11 09:19:55 crc kubenswrapper[4860]: E1211 09:19:55.580519 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-99qgp_openshift-machine-config-operator(31c30642-6e60-41b4-a477-0d802424e0aa)\"" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" Dec 11 09:20:07 crc kubenswrapper[4860]: I1211 09:20:07.593372 4860 scope.go:117] "RemoveContainer" containerID="6704d8c98ab428aac5737fa114668f6cf3119541ee0fe55ea326d429fcbe2d82" Dec 11 09:20:07 crc kubenswrapper[4860]: E1211 09:20:07.594027 4860 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-99qgp_openshift-machine-config-operator(31c30642-6e60-41b4-a477-0d802424e0aa)\"" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" Dec 11 09:20:20 crc kubenswrapper[4860]: I1211 09:20:20.579128 4860 scope.go:117] "RemoveContainer" containerID="6704d8c98ab428aac5737fa114668f6cf3119541ee0fe55ea326d429fcbe2d82" Dec 11 09:20:21 crc kubenswrapper[4860]: I1211 09:20:21.395292 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" event={"ID":"31c30642-6e60-41b4-a477-0d802424e0aa","Type":"ContainerStarted","Data":"6d9a1bcc95cac554ee629ce5ab961a89f0f93f98188a4ffb21541c3f6755546c"} Dec 11 09:21:06 crc kubenswrapper[4860]: I1211 09:21:06.865870 4860 generic.go:334] "Generic (PLEG): container finished" podID="38bf9964-7add-4228-9abc-c1a5bd18d568" containerID="d2806d73fffa4fc0a3571875988c8f69f884ca94d9b9fca50b635eadd9b5a188" exitCode=0 Dec 11 09:21:06 crc kubenswrapper[4860]: I1211 09:21:06.865960 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-fjd59/must-gather-phzqz" event={"ID":"38bf9964-7add-4228-9abc-c1a5bd18d568","Type":"ContainerDied","Data":"d2806d73fffa4fc0a3571875988c8f69f884ca94d9b9fca50b635eadd9b5a188"} Dec 11 09:21:06 crc kubenswrapper[4860]: I1211 09:21:06.866890 4860 scope.go:117] "RemoveContainer" containerID="d2806d73fffa4fc0a3571875988c8f69f884ca94d9b9fca50b635eadd9b5a188" Dec 11 09:21:07 crc kubenswrapper[4860]: I1211 09:21:07.306072 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-fjd59_must-gather-phzqz_38bf9964-7add-4228-9abc-c1a5bd18d568/gather/0.log" Dec 11 09:21:18 crc kubenswrapper[4860]: I1211 09:21:18.947317 4860 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-fjd59/must-gather-phzqz"] Dec 11 09:21:18 crc kubenswrapper[4860]: I1211 09:21:18.948323 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-fjd59/must-gather-phzqz" podUID="38bf9964-7add-4228-9abc-c1a5bd18d568" containerName="copy" containerID="cri-o://9aad5d3923ab3c049951961f15f9728b5ebf276e736f9ca6270c47c2dcd36786" gracePeriod=2 Dec 11 09:21:18 crc kubenswrapper[4860]: I1211 09:21:18.962305 4860 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-fjd59/must-gather-phzqz"] Dec 11 09:21:19 crc kubenswrapper[4860]: I1211 09:21:19.442352 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-fjd59_must-gather-phzqz_38bf9964-7add-4228-9abc-c1a5bd18d568/copy/0.log" Dec 11 09:21:19 crc kubenswrapper[4860]: I1211 09:21:19.443041 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-fjd59/must-gather-phzqz" Dec 11 09:21:19 crc kubenswrapper[4860]: I1211 09:21:19.511191 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qqwtc\" (UniqueName: \"kubernetes.io/projected/38bf9964-7add-4228-9abc-c1a5bd18d568-kube-api-access-qqwtc\") pod \"38bf9964-7add-4228-9abc-c1a5bd18d568\" (UID: \"38bf9964-7add-4228-9abc-c1a5bd18d568\") " Dec 11 09:21:19 crc kubenswrapper[4860]: I1211 09:21:19.511227 4860 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/38bf9964-7add-4228-9abc-c1a5bd18d568-must-gather-output\") pod \"38bf9964-7add-4228-9abc-c1a5bd18d568\" (UID: \"38bf9964-7add-4228-9abc-c1a5bd18d568\") " Dec 11 09:21:19 crc kubenswrapper[4860]: I1211 09:21:19.517834 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/38bf9964-7add-4228-9abc-c1a5bd18d568-kube-api-access-qqwtc" (OuterVolumeSpecName: "kube-api-access-qqwtc") pod "38bf9964-7add-4228-9abc-c1a5bd18d568" (UID: "38bf9964-7add-4228-9abc-c1a5bd18d568"). InnerVolumeSpecName "kube-api-access-qqwtc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 09:21:19 crc kubenswrapper[4860]: I1211 09:21:19.614246 4860 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qqwtc\" (UniqueName: \"kubernetes.io/projected/38bf9964-7add-4228-9abc-c1a5bd18d568-kube-api-access-qqwtc\") on node \"crc\" DevicePath \"\"" Dec 11 09:21:19 crc kubenswrapper[4860]: I1211 09:21:19.699989 4860 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/38bf9964-7add-4228-9abc-c1a5bd18d568-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "38bf9964-7add-4228-9abc-c1a5bd18d568" (UID: "38bf9964-7add-4228-9abc-c1a5bd18d568"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 09:21:19 crc kubenswrapper[4860]: I1211 09:21:19.716975 4860 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/38bf9964-7add-4228-9abc-c1a5bd18d568-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 11 09:21:19 crc kubenswrapper[4860]: I1211 09:21:19.985233 4860 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-fjd59_must-gather-phzqz_38bf9964-7add-4228-9abc-c1a5bd18d568/copy/0.log" Dec 11 09:21:19 crc kubenswrapper[4860]: I1211 09:21:19.986172 4860 generic.go:334] "Generic (PLEG): container finished" podID="38bf9964-7add-4228-9abc-c1a5bd18d568" containerID="9aad5d3923ab3c049951961f15f9728b5ebf276e736f9ca6270c47c2dcd36786" exitCode=143 Dec 11 09:21:19 crc kubenswrapper[4860]: I1211 09:21:19.986229 4860 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-fjd59/must-gather-phzqz" Dec 11 09:21:19 crc kubenswrapper[4860]: I1211 09:21:19.986239 4860 scope.go:117] "RemoveContainer" containerID="9aad5d3923ab3c049951961f15f9728b5ebf276e736f9ca6270c47c2dcd36786" Dec 11 09:21:20 crc kubenswrapper[4860]: I1211 09:21:20.098976 4860 scope.go:117] "RemoveContainer" containerID="d2806d73fffa4fc0a3571875988c8f69f884ca94d9b9fca50b635eadd9b5a188" Dec 11 09:21:20 crc kubenswrapper[4860]: I1211 09:21:20.195220 4860 scope.go:117] "RemoveContainer" containerID="9aad5d3923ab3c049951961f15f9728b5ebf276e736f9ca6270c47c2dcd36786" Dec 11 09:21:20 crc kubenswrapper[4860]: E1211 09:21:20.195744 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9aad5d3923ab3c049951961f15f9728b5ebf276e736f9ca6270c47c2dcd36786\": container with ID starting with 9aad5d3923ab3c049951961f15f9728b5ebf276e736f9ca6270c47c2dcd36786 not found: ID does not exist" containerID="9aad5d3923ab3c049951961f15f9728b5ebf276e736f9ca6270c47c2dcd36786" Dec 11 09:21:20 crc kubenswrapper[4860]: I1211 09:21:20.195790 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9aad5d3923ab3c049951961f15f9728b5ebf276e736f9ca6270c47c2dcd36786"} err="failed to get container status \"9aad5d3923ab3c049951961f15f9728b5ebf276e736f9ca6270c47c2dcd36786\": rpc error: code = NotFound desc = could not find container \"9aad5d3923ab3c049951961f15f9728b5ebf276e736f9ca6270c47c2dcd36786\": container with ID starting with 9aad5d3923ab3c049951961f15f9728b5ebf276e736f9ca6270c47c2dcd36786 not found: ID does not exist" Dec 11 09:21:20 crc kubenswrapper[4860]: I1211 09:21:20.195821 4860 scope.go:117] "RemoveContainer" containerID="d2806d73fffa4fc0a3571875988c8f69f884ca94d9b9fca50b635eadd9b5a188" Dec 11 09:21:20 crc kubenswrapper[4860]: E1211 09:21:20.196401 4860 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d2806d73fffa4fc0a3571875988c8f69f884ca94d9b9fca50b635eadd9b5a188\": container with ID starting with d2806d73fffa4fc0a3571875988c8f69f884ca94d9b9fca50b635eadd9b5a188 not found: ID does not exist" containerID="d2806d73fffa4fc0a3571875988c8f69f884ca94d9b9fca50b635eadd9b5a188" Dec 11 09:21:20 crc kubenswrapper[4860]: I1211 09:21:20.196461 4860 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d2806d73fffa4fc0a3571875988c8f69f884ca94d9b9fca50b635eadd9b5a188"} err="failed to get container status \"d2806d73fffa4fc0a3571875988c8f69f884ca94d9b9fca50b635eadd9b5a188\": rpc error: code = NotFound desc = could not find container \"d2806d73fffa4fc0a3571875988c8f69f884ca94d9b9fca50b635eadd9b5a188\": container with ID starting with d2806d73fffa4fc0a3571875988c8f69f884ca94d9b9fca50b635eadd9b5a188 not found: ID does not exist" Dec 11 09:21:21 crc kubenswrapper[4860]: I1211 09:21:21.590937 4860 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="38bf9964-7add-4228-9abc-c1a5bd18d568" path="/var/lib/kubelet/pods/38bf9964-7add-4228-9abc-c1a5bd18d568/volumes" Dec 11 09:22:38 crc kubenswrapper[4860]: I1211 09:22:38.795938 4860 patch_prober.go:28] interesting pod/machine-config-daemon-99qgp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 09:22:38 crc kubenswrapper[4860]: I1211 09:22:38.796591 4860 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 09:23:08 crc kubenswrapper[4860]: I1211 09:23:08.795197 4860 patch_prober.go:28] interesting pod/machine-config-daemon-99qgp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 09:23:08 crc kubenswrapper[4860]: I1211 09:23:08.795735 4860 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 09:23:38 crc kubenswrapper[4860]: I1211 09:23:38.795552 4860 patch_prober.go:28] interesting pod/machine-config-daemon-99qgp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 09:23:38 crc kubenswrapper[4860]: I1211 09:23:38.797097 4860 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 09:23:38 crc kubenswrapper[4860]: I1211 09:23:38.797178 4860 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" Dec 11 09:23:38 crc kubenswrapper[4860]: I1211 09:23:38.798119 4860 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6d9a1bcc95cac554ee629ce5ab961a89f0f93f98188a4ffb21541c3f6755546c"} pod="openshift-machine-config-operator/machine-config-daemon-99qgp" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 11 09:23:38 crc kubenswrapper[4860]: I1211 09:23:38.798186 4860 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" podUID="31c30642-6e60-41b4-a477-0d802424e0aa" containerName="machine-config-daemon" containerID="cri-o://6d9a1bcc95cac554ee629ce5ab961a89f0f93f98188a4ffb21541c3f6755546c" gracePeriod=600 Dec 11 09:23:39 crc kubenswrapper[4860]: I1211 09:23:39.420156 4860 generic.go:334] "Generic (PLEG): container finished" podID="31c30642-6e60-41b4-a477-0d802424e0aa" containerID="6d9a1bcc95cac554ee629ce5ab961a89f0f93f98188a4ffb21541c3f6755546c" exitCode=0 Dec 11 09:23:39 crc kubenswrapper[4860]: I1211 09:23:39.420226 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" event={"ID":"31c30642-6e60-41b4-a477-0d802424e0aa","Type":"ContainerDied","Data":"6d9a1bcc95cac554ee629ce5ab961a89f0f93f98188a4ffb21541c3f6755546c"} Dec 11 09:23:39 crc kubenswrapper[4860]: I1211 09:23:39.420977 4860 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-99qgp" event={"ID":"31c30642-6e60-41b4-a477-0d802424e0aa","Type":"ContainerStarted","Data":"834e8653c95d75fb3f3d889f22f8b88becb7117c291acd0f8e2acc55818ab7d0"} Dec 11 09:23:39 crc kubenswrapper[4860]: I1211 09:23:39.421016 4860 scope.go:117] "RemoveContainer" containerID="6704d8c98ab428aac5737fa114668f6cf3119541ee0fe55ea326d429fcbe2d82" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515116506767024463 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015116506770017372 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015116475663016523 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015116475663015473 5ustar corecore